[Users] Successfully virt-v2v from CentOS 6.3 VM to Ovirt 3.2 nightly

Gianluca Cecchi gianluca.cecchi at gmail.com
Wed Jan 9 15:55:53 UTC 2013


Hello,
on my oVirt Host configured with F18 and all-in-one and ovirt-nightly as of
ovirt-engine-3.2.0-1.20130107.git1a60fea.fc18.noarch

I was able to import a CentOS 5.8 VM coming from a CentOS 6.3 host.

The oVirt node server is the same where I'm unable to run a newly created
WIndows 7 32bit vm...
See http://lists.ovirt.org/pipermail/users/2013-January/011390.html

In this thread I would like to report about successful import phases and
some doubts about:
1) no password requested during virt-v2v
2) no connectivity in guest imported.

On CentOS 6.3 host
# virt-v2v -o rhev -osd 10.4.4.59:/EXPORT --network ovirtmgmt c56cr
c56cr_001: 100%
[===================================================================================]D
0h02m17s
virt-v2v: c56cr configured with virtio drivers.

---> I would expect to be asked for the password of a privileged user in
oVirt infra, instead the export process started without any prompt.
Is this correct?
In my opinion in this case it could be a security concern....

during virt-v2v command, on oVirt node I see this inside NFS Export domain:
$ sudo ls -l
/EXPORT/b878ad09-602f-47da-87f5-2829d67d3321/v2v.pmbPOGM_/30df5806-6911-41b3-8fef-1fd8d755659f
total 10485764
-rw-r--r--. 1 vdsm kvm 10737418240 Jan  9 16:05
0d0e8e12-8b35-4034-89fc-8cbd4a7d7d81

At the end of the process:
$ sudo ls -l /EXPORT/b878ad09-602f-47da-87f5-2829d67d3321/images/
total 4
drwxr-xr-x. 2 vdsm kvm 4096 Jan  9 16:05
30df5806-6911-41b3-8fef-1fd8d755659f

$ sudo ls -lR /EXPORT/
/EXPORT/:
total 4
drwxr-xr-x. 5 vdsm kvm 4096 Jan  9 16:06
b878ad09-602f-47da-87f5-2829d67d3321

/EXPORT/b878ad09-602f-47da-87f5-2829d67d3321:
total 12
drwxr-xr-x. 2 vdsm kvm 4096 Jan  9 16:01 dom_md
drwxr-xr-x. 3 vdsm kvm 4096 Jan  9 16:06 images
drwxr-xr-x. 4 vdsm kvm 4096 Jan  9 16:02 master

/EXPORT/b878ad09-602f-47da-87f5-2829d67d3321/dom_md:
total 8
-rw-rw----. 1 vdsm kvm   0 Jan  9 16:01 ids
-rw-rw----. 1 vdsm kvm   0 Jan  9 16:01 inbox
-rw-rw----. 1 vdsm kvm 512 Jan  9 16:01 leases
-rw-r--r--. 1 vdsm kvm 350 Jan  9 16:01 metadata
-rw-rw----. 1 vdsm kvm   0 Jan  9 16:01 outbox

/EXPORT/b878ad09-602f-47da-87f5-2829d67d3321/images:
total 4
drwxr-xr-x. 2 vdsm kvm 4096 Jan  9 16:05
30df5806-6911-41b3-8fef-1fd8d755659f

/EXPORT/b878ad09-602f-47da-87f5-2829d67d3321/images/30df5806-6911-41b3-8fef-1fd8d755659f:
total 10485768
-rw-r--r--. 1 vdsm kvm 10737418240 Jan  9 16:06
0d0e8e12-8b35-4034-89fc-8cbd4a7d7d81
-rw-r--r--. 1 vdsm kvm         330 Jan  9 16:05
0d0e8e12-8b35-4034-89fc-8cbd4a7d7d81.meta

/EXPORT/b878ad09-602f-47da-87f5-2829d67d3321/master:
total 8
drwxr-xr-x. 2 vdsm kvm 4096 Jan  9 16:02 tasks
drwxr-xr-x. 3 vdsm kvm 4096 Jan  9 16:06 vms

/EXPORT/b878ad09-602f-47da-87f5-2829d67d3321/master/tasks:
total 0

/EXPORT/b878ad09-602f-47da-87f5-2829d67d3321/master/vms:
total 4
drwxr-xr-x. 2 vdsm kvm 4096 Jan  9 16:06
2398149c-32b9-4bae-b572-134d973a759c

/EXPORT/b878ad09-602f-47da-87f5-2829d67d3321/master/vms/2398149c-32b9-4bae-b572-134d973a759c:
total 8
-rw-r--r--. 1 vdsm kvm 4649 Jan  9 16:06
2398149c-32b9-4bae-b572-134d973a759c.ovf


Then I began the vm import in webadmin:

Import process has begun for VM(s): c56cr.
You can check import status in the 'Events' tab of the specific destination
storage domain, or in the main 'Events' tab

---> regarding the import status, the "specific destination storage domain"
would be my DATA domain, correct?
Because I see nothing in it and nothing in export domain.
Instead I correctly see in main events tab of the cluster these two messages

2013-Jan-09, 16:16 Starting to import Vm c56cr to Data Center Poli, Cluster
Poli1
2013-Jan-09, 16:18 Vm c56cr was imported successfully to Data Center Poli,
Cluster Poli1

SO probably the first option should go away....?

During the import, on the oVirt host
[g.cecchi at f18aio ~]$ vmstat 3
procs -----------memory---------- ---swap-- -----io---- -system--
----cpu----
 r  b   swpd   free   buff  cache   si   so    bi    bo   in   cs us sy id
wa
 1  1      0 1684556 121824 28660956    0    0     8    69   21   66  0  0
99  0
 1  1      0 1515192 121824 28830112    0    0     0 58749 4468 6068  0  3
85 11
 0  1      0 1330708 121828 29014320    0    0     0 59415 4135 5149  0  4
85 11

$ sudo iotop -d 3 -P -o -k
Total DISK READ:       0.33 K/s | Total DISK WRITE:   56564.47 K/s
  PID  PRIO  USER     DISK READ  DISK WRITE  SWAPIN     IO>    COMMAND

22501 idle vdsm     55451.24 K/s 56459.45 K/s  0.00 % 91.03 % dd
if=/rhev/data-center/~count=10240 oflag=direct
  831 be/4 root        0.00 K/s    0.00 K/s  0.00 %  3.56 % [flush-253:1]
  576 be/3 root        0.00 K/s   19.69 K/s  0.00 %  0.72 % [jbd2/dm-1-8]
23309 be/3 vdsm        0.33 K/s    0.00 K/s  0.00 %  0.00 % python
/usr/share/vdsm/st~moteFileHandler.pyc 49 47
17057 be/4 apache      0.00 K/s    2.63 K/s  0.00 %  0.00 % httpd
-DFOREGROUND
15524 be/4 root        0.00 K/s    1.31 K/s  0.00 %  0.00 % libvirtd
--listen

$ ps -wfp 22501
UID        PID  PPID  C STIME TTY          TIME CMD
vdsm     22501 16120  8 16:16 ?        00:00:14 /usr/bin/dd
if=/rhev/data-center/89d40d09-5109-4070-b9b0-86f1addce8af/b878ad09-602f-

I was then able to power on and connect via vnc to the console.
But I noticed it has no connectivity with its gateway

Host is on vlan 65
(em3 + em3.65 cofigured)

host has
3: em3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master
ovirtmgmt state UP qlen 1000
    link/ether 00:1c:c4:ab:3a:dd brd ff:ff:ff:ff:ff:ff
    inet6 fe80::21c:c4ff:feab:3add/64 scope link
       valid_lft forever preferred_lft forever
...
6: ovirtmgmt: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
state UP
    link/ether 00:1c:c4:ab:3a:dd brd ff:ff:ff:ff:ff:ff
    inet6 fe80::21c:c4ff:feab:3add/64 scope link
       valid_lft forever preferred_lft forever
7: em3.65 at em3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
state UP
    link/ether 00:1c:c4:ab:3a:dd brd ff:ff:ff:ff:ff:ff
    inet 10.4.4.59/24 brd 10.4.4.255 scope global em3.65
    inet6 fe80::21c:c4ff:feab:3add/64 scope link
       valid_lft forever preferred_lft forever
...
13: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
master ovirtmgmt state UNKNOWN qlen 500
    link/ether fe:54:00:d3:8f:a3 brd ff:ff:ff:ff:ff:ff
    inet6 fe80::fc54:ff:fed3:8fa3/64 scope link
       valid_lft forever preferred_lft forever

[g.cecchi at f18aio ~]$ ip route list
default via 10.4.4.250 dev em3.65
10.4.4.0/24 dev em3.65  proto kernel  scope link  src 10.4.4.59

ovirtmgmt is tagged in datacenter Poli1

guest is originally configured (and it maintained this) on bridged vlan65
on CentOS 63 host. Its parameters

eth0 with
ip 10.4.4.53 and gw 10.4.4.250

from webadmin pov it seems ok. see also this screenshot
https://docs.google.com/open?id=0BwoPbcrMv8mvbENvR242VFJ2M1k

any help will be appreciated.
do I have to enable some kind of routing not enabled by default..?

Thanks,
Gianluca
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20130109/c51b6851/attachment-0001.html>


More information about the Users mailing list