On 07/17/2013 09:04 PM, Steve Dainard wrote:
*Web-UI displays:*
VM VM1 is down. Exit message: internal error process exited while
connecting to monitor: qemu-system-x86_64: -drive
file=gluster://ovirt001/vol1/a87a7ef6-2c74-4d8e-a6e0-a392d0f791cf/images/238cc6cf-070c-4483-b686-c0de7ddf0dfa/ff2bca2d-4ed1-46c6-93c8-22a39bb1626a,if=none,id=drive-virtio-disk0,format=raw,serial=238cc6cf-070c-4483-b686-c0de7ddf0dfa,cache=none,werror=stop,rerror=stop,aio=threads:
could not open disk image
gluster://ovirt001/vol1/a87a7ef6-2c74-4d8e-a6e0-a392d0f791cf/images/238cc6cf-070c-4483-b686-c0de7ddf0dfa/ff2bca2d-4ed1-46c6-93c8-22a39bb1626a:
No such file or directory .
VM VM1 was started by admin@internal (Host: ovirt001).
The disk VM1_Disk1 was successfully added to VM VM1.
*I can see the image on the gluster machine, and it looks to have the
correct permissions:*
[root@ovirt001 238cc6cf-070c-4483-b686-c0de7ddf0dfa]# pwd
/mnt/storage1/vol1/a87a7ef6-2c74-4d8e-a6e0-a392d0f791cf/images/238cc6cf-070c-4483-b686-c0de7ddf0dfa
[root@ovirt001 238cc6cf-070c-4483-b686-c0de7ddf0dfa]# ll
total 1028
-rw-rw----. 2 vdsm kvm 32212254720 Jul 17 11:11
ff2bca2d-4ed1-46c6-93c8-22a39bb1626a
-rw-rw----. 2 vdsm kvm 1048576 Jul 17 11:11
ff2bca2d-4ed1-46c6-93c8-22a39bb1626a.lease
-rw-r--r--. 2 vdsm kvm 268 Jul 17 11:11
ff2bca2d-4ed1-46c6-93c8-22a39bb1626a.meta
Can you please try after doing these changes:
1) gluster volume set <volname> server.allow-insecure on
2) Edit /etc/glusterfs/glusterd.vol to contain this line:
option rpc-auth-allow-insecure on
Post 2), restarting glusterd would be necessary.
Thanks,
Vijay