This is a multi-part message in MIME format.
--------------090401040101080908060106
Content-Type: text/plain; charset=ISO-8859-1; format=flowed
Content-Transfer-Encoding: 7bit
On 06/28/2012 09:55 AM, jose garcia wrote:
On 06/28/2012 02:43 AM, Itamar Heim wrote:
> On 06/27/2012 09:15 AM, jose garcia wrote:
>> Good evening to all,
>>
>> I have added two nodes of Fedora 17 with vdsm 4.10 installed to oVirt
>> 3.1 engine. I was having problems
>> with SSL, so I have disabled it. When I added the nodes there was no
>> attempt of installation as it was
>> the case with oVirt 3.0, but the nodes get activated, provided that
>> ovirtmgmt bridge is present.
>>
>> I have been told that there is a configuration in the database that
>> make
>> this happen. I just recreated the
>> database via the script /dbscripts/create_db_devel.sh and run
>> engine-setup, after removing all packages from oVirt 3.0 and jboss and
>> installing ovirt 3.1 basic packages.
>>
>> My question is: What would be the 'standard' procedure to get oVirt 3.1
>> running?
>
> the standard procedure would be to install the rpm and engine-setup
> and add the host from UI so they will be configured with certificates
> and ssl
Good morning,
I wanted to know if there will be an installation process from scratch
with a reboot for the hosts as there was in 3.0.
I have an issue that may also be related to authentication. I am
unable to start my newly created VM. The engine seems to connect to
libvirt in read-only mode, and boot of the virtual Machine fails.
There is something quaint in virsh (Fedora 17 in the hosts):
when trying to connect without specifying the uri (qemu:///system) it
gives segmentation fault. If it is given it asks for user and password.
Tried to disable sasl in libvirtd.conf and now I can connect with
virsh to the default uri without providing a password, but booting of
the vm keeps failing.
It is required to use sasl or to add a sasl user in the engine, in the
hosts or in both?
vdsm.log reads:
Thread-72::DEBUG::2012-06-27
18:57:25,511::task::978::TaskManager.Task::(_decref)
Task=`e92b65bc-b9fe-492f-b77c-397321dbb105`::ref 0 aborting False
Thread-70::DEBUG::2012-06-27
18:57:25,531::vm::580::vm.Vm::(_startUnderlyingVm)
vmId=`32339151-23ed-4cc3-ada4-0f540ab81a97`::_ongoingCreations released
Thread-70::ERROR::2012-06-27
18:57:25,532::vm::604::vm.Vm::(_startUnderlyingVm)
vmId=`32339151-23ed-4cc3-ada4-0f540ab81a97`::The vm start process failed
Traceback (most recent call last):
File "/usr/share/vdsm/vm.py", line 570, in _startUnderlyingVm
self._run()
File "/usr/share/vdsm/libvirtvm.py", line 1364, in _run
self._connection.createXML(domxml, flags),
File "/usr/lib/python2.7/site-packages/vdsm/libvirtconnection.py",
line 82, in wrapper
ret = f(*args, **kwargs)
File "/usr/lib64/python2.7/site-packages/libvirt.py", line 2420, in
createXML
if ret is None:raise libvirtError('virDomainCreateXML() failed',
conn=self)
*libvirtError: internal error unsupported configuration: Readonly
leases are not supported*
Thread-70::DEBUG::2012-06-27
18:57:25,542::vm::920::vm.Vm::(setDownStatus)
vmId=`32339151-23ed-4cc3-ada4-0f540ab81a97`::Changed state to Down:
internal error unsupported configuration: Readonly leases are not
supported
Thread-75::DEBUG::2012-06-27
18:57:27,588::BindingXMLRPC::859::vds::(wrapper) client
[10.10.30.101]::call vmGetStats with
('32339151-23ed-4cc3-ada4-0f540ab81a97',) {}
Thread-75::DEBUG::2012-06-27
18:57:27,588::BindingXMLRPC::865::vds::(wrapper) return vmGetStats
with {'status': {'message': 'Done', 'code': 0},
'statsList':
[{'status': 'Down', 'hash': '0', 'exitMessage':
'internal error
unsupported configuration: Readonly leases are not supported', 'vmId':
'32339151-23ed-4cc3-ada4-0f540ab81a97', 'timeOffset': '0',
'exitCode':
1}]}
Thread-76::DEBUG::2012-06-27
18:57:27,594::BindingXMLRPC::859::vds::(wrapper) client
[10.10.30.101]::call vmDestroy with
('32339151-23ed-4cc3-ada4-0f540ab81a97',) {}
Thread-76::INFO::2012-06-27 18:57:27,595::API::317::vds::(destroy)
vmContainerLock acquired by vm 32339151-23ed-4cc3-ada4-0f540ab81a97
Thread-76::DEBUG::2012-06-27
18:57:27,595::libvirtvm::2085::vm.Vm::(destroy)
vmId=`32339151-23ed-4cc3-ada4-0f540ab81a97`::destroy Called
Thread-76::INFO::2012-06-27
18:57:27,595::libvirtvm::2040::vm.Vm::(releaseVm)
vmId=`32339151-23ed-4cc3-ada4-0f540ab81a97`::Release VM resources
Thread-76::WARNING::2012-06-27
18:57:27,596::vm::328::vm.Vm::(_set_lastStatus)
vmId=`32339151-23ed-4cc3-ada4-0f540ab81a97`::trying to set state to
Powering down when already Down
Thread-76::DEBUG::2012-06-27
18:57:27,596::__init__::1249::Storage.Misc.excCmd::(_log)
'/usr/bin/sudo -n /usr/sbin/service ksmtuned retune'
And there is a log in /var/log/libvirt/qemu for the VM that says:
starting up
LC_ALL=C PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin
QEMU_AUDIO_DRV=spice /usr/bin/qemu-kvm -S -M pc-0.14 -cpu
kvm64,+lahf_lm,+ssse3,-cx16 -enable-kvm -m 1024 -smp
1,sockets=1,cores=1,threads=1 -name fedora-test -uuid
32339151-23ed-4cc3-ada4-0f540ab81a97 -smbios type=1,manufacturer=Red
Hat,product=RHEV
Hypervisor,version=17-1,serial=03000200-0400-0500-0006-000700080009_00:30:18:a8:a8:42,uuid=32339151-23ed-4cc3-ada4-0f540ab81a97
-nodefconfig -nodefaults -chardev
socket,id=charmonitor,path=/var/lib/libvirt/qemu/fedora-test.monitor,server,nowait
-mon chardev=charmonitor,id=monitor,mode=control -rtc
base=2012-06-27T17:57:28,driftfix=slew -no-shutdown -device
piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -device
virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x4 -drive
file=/rhev/data-center/ee37f596-bf78-11e1-94ba-7309589a8ec2/891fa5d3-ceff-4711-8538-9bccd018969c/images/11111111-1111-1111-1111-111111111111/Fedora-16-x86_64-Live-KDE.iso,if=none,media=cdrom,id=drive-ide0-1-0,readonly=on,format=raw,serial=
-device
ide-drive,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0,bootindex=1 -drive
file=/rhev/data-center/ee37f596-bf78-11e1-94ba-7309589a8ec2/cdd28fda-e2fa-4ce4-9805-566ad7d69df2/images/81d0b868-9e5c-4ed8-9ac0-db6982da5de1/80a356f1-0932-4d54-be81-e112720c60b0,if=none,id=drive-virtio-disk0,format=raw,serial=81d0b868-9e5c-4ed8-9ac0-db6982da5de1,cache=none,werror=stop,rerror=stop,aio=threads
-device
virtio-blk-pci,scsi=off,bus=pci.0,addr=0x5,drive=drive-virtio-disk0,id=virtio-disk0,bootindex=2
-netdev tap,fd=25,id=hostnet0,vhost=on,vhostfd=26 -device
virtio-net-pci,netdev=hostnet0,id=net0,mac=00:1a:4a:16:01:51,bus=pci.0,addr=0x3,bootindex=3
-chardev
socket,id=charchannel0,path=/var/lib/libvirt/qemu/channels/fedora-test.com.redhat.rhevm.vdsm,server,nowait
-device
virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel0,name=com.redhat.rhevm.vdsm
-chardev spicevmc,id=charchannel1,name=vdagent -device
virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel1,id=channel1,name=com.redhat.spice.0
-chardev pty,id=charconsole0 -device
virtconsole,chardev=charconsole0,id=console0 -spice port=5900,addr=0
-k en-us -vga qxl -global qxl-vga.vram_size=67108864 -device
virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x6
*libvir: Locking error : unsupported configuration: Readonly leases
are not supported*
2012-06-27 17:57:28.241+0000: shutting down
Sorry for that, after reboot of the engine and the nodes I can boot the
VM, but what would be required to use sasl, which is the the default?
--------------090401040101080908060106
Content-Type: text/html; charset=ISO-8859-1
Content-Transfer-Encoding: 7bit
<html>
<head>
<meta content="text/html; charset=ISO-8859-1"
http-equiv="Content-Type">
</head>
<body bgcolor="#FFFFFF" text="#000000">
<div class="moz-cite-prefix">On 06/28/2012 09:55 AM, jose garcia
wrote:<br>
</div>
<blockquote cite="mid:4FEC1BF1.1070005@gmail.com"
type="cite">
<meta content="text/html; charset=ISO-8859-1"
http-equiv="Content-Type">
<div class="moz-cite-prefix">On 06/28/2012 02:43 AM, Itamar Heim
wrote:<br>
</div>
<blockquote cite="mid:4FEBB6D9.4000802@redhat.com"
type="cite">On
06/27/2012 09:15 AM, jose garcia wrote: <br>
<blockquote type="cite">Good evening to all, <br>
<br>
I have added two nodes of Fedora 17 with vdsm 4.10 installed
to oVirt <br>
3.1 engine. I was having problems <br>
with SSL, so I have disabled it. When I added the nodes there
was no <br>
attempt of installation as it was <br>
the case with oVirt 3.0, but the nodes get activated, provided
that <br>
ovirtmgmt bridge is present. <br>
<br>
I have been told that there is a configuration in the database
that make <br>
this happen. I just recreated the <br>
database via the script /dbscripts/create_db_devel.sh and run
<br>
engine-setup, after removing all packages from oVirt 3.0 and
jboss and <br>
installing ovirt 3.1 basic packages. <br>
<br>
My question is: What would be the 'standard' procedure to get
oVirt 3.1 <br>
running? <br>
</blockquote>
<br>
the standard procedure would be to install the rpm and
engine-setup and add the host from UI so they will be configured
with certificates and ssl <br>
</blockquote>
<br>
Good morning,<br>
<br>
I wanted to know if there will be an installation process from
scratch with a reboot for the hosts as there was in 3.0.<br>
<br>
I have an issue that may also be related to authentication. I am
unable to start my newly created VM. The engine seems to connect
to libvirt in read-only mode, and boot of the virtual Machine
fails. There is something quaint in virsh (Fedora 17 in the
hosts): <br>
<br>
when trying to connect without specifying the uri (qemu:///system)
it gives segmentation fault. If it is given it asks for user and
password. <br>
<br>
Tried to disable sasl in libvirtd.conf and now I can connect with
virsh to the default uri without providing a password, but booting
of the vm keeps failing. <br>
<br>
It is required to use sasl or to add a sasl user in the engine, in
the hosts or in both?<br>
<br>
vdsm.log reads:<br>
<br>
Thread-72::DEBUG::2012-06-27
18:57:25,511::task::978::TaskManager.Task::(_decref)
Task=`e92b65bc-b9fe-492f-b77c-397321dbb105`::ref 0 aborting False<br>
Thread-70::DEBUG::2012-06-27
18:57:25,531::vm::580::vm.Vm::(_startUnderlyingVm)
vmId=`32339151-23ed-4cc3-ada4-0f540ab81a97`::_ongoingCreations
released<br>
Thread-70::ERROR::2012-06-27
18:57:25,532::vm::604::vm.Vm::(_startUnderlyingVm)
vmId=`32339151-23ed-4cc3-ada4-0f540ab81a97`::The vm start process
failed<br>
Traceback (most recent call last):<br>
File "/usr/share/vdsm/vm.py", line 570, in
_startUnderlyingVm<br>
self._run()<br>
File "/usr/share/vdsm/libvirtvm.py", line 1364, in
_run<br>
self._connection.createXML(domxml, flags),<br>
File
"/usr/lib/python2.7/site-packages/vdsm/libvirtconnection.py", line
82, in wrapper<br>
ret = f(*args, **kwargs)<br>
File "/usr/lib64/python2.7/site-packages/libvirt.py", line
2420,
in createXML<br>
if ret is None:raise
libvirtError('virDomainCreateXML()
failed', conn=self)<br>
<b>libvirtError: internal error unsupported configuration:
Readonly leases are not supported</b><br>
Thread-70::DEBUG::2012-06-27
18:57:25,542::vm::920::vm.Vm::(setDownStatus)
vmId=`32339151-23ed-4cc3-ada4-0f540ab81a97`::Changed state to
Down: internal error unsupported configuration: Readonly leases
are not supported<br>
Thread-75::DEBUG::2012-06-27
18:57:27,588::BindingXMLRPC::859::vds::(wrapper) client
[10.10.30.101]::call vmGetStats with
('32339151-23ed-4cc3-ada4-0f540ab81a97',) {}<br>
Thread-75::DEBUG::2012-06-27
18:57:27,588::BindingXMLRPC::865::vds::(wrapper) return vmGetStats
with {'status': {'message': 'Done', 'code': 0},
'statsList':
[{'status': 'Down', 'hash': '0',
'exitMessage': 'internal error
unsupported configuration: Readonly leases are not supported',
'vmId': '32339151-23ed-4cc3-ada4-0f540ab81a97',
'timeOffset': '0',
'exitCode': 1}]}<br>
Thread-76::DEBUG::2012-06-27
18:57:27,594::BindingXMLRPC::859::vds::(wrapper) client
[10.10.30.101]::call vmDestroy with
('32339151-23ed-4cc3-ada4-0f540ab81a97',) {}<br>
Thread-76::<a moz-do-not-send="true"
class="moz-txt-link-freetext"
href="INFO::2012-06-27">INFO::2012-06-27</a>
18:57:27,595::API::317::vds::(destroy) vmContainerLock acquired by
vm 32339151-23ed-4cc3-ada4-0f540ab81a97<br>
Thread-76::DEBUG::2012-06-27
18:57:27,595::libvirtvm::2085::vm.Vm::(destroy)
vmId=`32339151-23ed-4cc3-ada4-0f540ab81a97`::destroy Called<br>
Thread-76::<a moz-do-not-send="true"
class="moz-txt-link-freetext"
href="INFO::2012-06-27">INFO::2012-06-27</a>
18:57:27,595::libvirtvm::2040::vm.Vm::(releaseVm)
vmId=`32339151-23ed-4cc3-ada4-0f540ab81a97`::Release VM resources<br>
Thread-76::WARNING::2012-06-27
18:57:27,596::vm::328::vm.Vm::(_set_lastStatus)
vmId=`32339151-23ed-4cc3-ada4-0f540ab81a97`::trying to set state
to Powering down when already Down<br>
Thread-76::DEBUG::2012-06-27
18:57:27,596::__init__::1249::Storage.Misc.excCmd::(_log)
'/usr/bin/sudo -n /usr/sbin/service ksmtuned retune'<br>
<br>
And there is a log in /var/log/libvirt/qemu for the VM that says:<br>
<br>
starting up<br>
LC_ALL=C PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin
QEMU_AUDIO_DRV=spice /usr/bin/qemu-kvm -S -M pc-0.14 -cpu
kvm64,+lahf_lm,+ssse3,-cx16 -enable-kvm -m 1024 -smp
1,sockets=1,cores=1,threads=1 -name fedora-test -uuid
32339151-23ed-4cc3-ada4-0f540ab81a97 -smbios
type=1,manufacturer=Red Hat,product=RHEV
Hypervisor,version=17-1,serial=03000200-0400-0500-0006-000700080009_00:30:18:a8:a8:42,uuid=32339151-23ed-4cc3-ada4-0f540ab81a97
-nodefconfig -nodefaults -chardev
socket,id=charmonitor,path=/var/lib/libvirt/qemu/fedora-test.monitor,server,nowait
-mon chardev=charmonitor,id=monitor,mode=control -rtc
base=2012-06-27T17:57:28,driftfix=slew -no-shutdown -device
piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -device
virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x4 -drive
file=/rhev/data-center/ee37f596-bf78-11e1-94ba-7309589a8ec2/891fa5d3-ceff-4711-8538-9bccd018969c/images/11111111-1111-1111-1111-111111111111/Fedora-16-x86_64-Live-KDE.iso,if=none,media=cdrom,id=drive-ide0-1-0,readonly=on,format=raw,serial=
-device
ide-drive,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0,bootindex=1
-drive
file=/rhev/data-center/ee37f596-bf78-11e1-94ba-7309589a8ec2/cdd28fda-e2fa-4ce4-9805-566ad7d69df2/images/81d0b868-9e5c-4ed8-9ac0-db6982da5de1/80a356f1-0932-4d54-be81-e112720c60b0,if=none,id=drive-virtio-disk0,format=raw,serial=81d0b868-9e5c-4ed8-9ac0-db6982da5de1,cache=none,werror=stop,rerror=stop,aio=threads
-device
virtio-blk-pci,scsi=off,bus=pci.0,addr=0x5,drive=drive-virtio-disk0,id=virtio-disk0,bootindex=2
-netdev tap,fd=25,id=hostnet0,vhost=on,vhostfd=26 -device
virtio-net-pci,netdev=hostnet0,id=net0,mac=00:1a:4a:16:01:51,bus=pci.0,addr=0x3,bootindex=3
-chardev
socket,id=charchannel0,path=/var/lib/libvirt/qemu/channels/fedora-test.com.redhat.rhevm.vdsm,server,nowait
-device
virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel0,name=com.redhat.rhevm.vdsm
-chardev spicevmc,id=charchannel1,name=vdagent -device
virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel1,id=channel1,name=com.redhat.spice.0
-chardev pty,id=charconsole0 -device
virtconsole,chardev=charconsole0,id=console0 -spice
port=5900,addr=0 -k en-us -vga qxl -global
qxl-vga.vram_size=67108864 -device
virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x6<br>
<b>libvir: Locking error : unsupported configuration: Readonly
leases are not supported</b><br>
2012-06-27 17:57:28.241+0000: shutting down<br>
<br>
<br>
<br>
</blockquote>
Sorry for that, after reboot of the engine and the nodes I can boot
the VM, but what would be required to use sasl, which is the the
default?<br>
<br>
</body>
</html>
--------------090401040101080908060106--