
On 07/17/2013 10:20 PM, Steve Dainard wrote:
Completed changes:
*gluster> volume info vol1* Volume Name: vol1 Type: Replicate Volume ID: 97c3b2a7-0391-4fae-b541-cf04ce6bde0f Status: Started Number of Bricks: 1 x 2 = 2 Transport-type: tcp Bricks: Brick1: ovirt001.miovision.corp:/mnt/storage1/vol1 Brick2: ovirt002.miovision.corp:/mnt/storage1/vol1 Options Reconfigured: network.remote-dio: on cluster.eager-lock: enable performance.stat-prefetch: off performance.io-cache: off performance.read-ahead: off performance.quick-read: off storage.owner-gid: 36 storage.owner-uid: 36 auth.allow: * user.cifs: on nfs.disable: off server.allow-insecure: on
*Same error on VM run:* VM VM1 is down. Exit message: internal error process exited while connecting to monitor: qemu-system-x86_64: -drive file=gluster://ovirt001/vol1/a87a7ef6-2c74-4d8e-a6e0-a392d0f791cf/images/238cc6cf-070c-4483-b686-c0de7ddf0dfa/ff2bca2d-4ed1-46c6-93c8-22a39bb1626a,if=none,id=drive-virtio-disk0,format=raw,serial=238cc6cf-070c-4483-b686-c0de7ddf0dfa,cache=none,werror=stop,rerror=stop,aio=threads: could not open disk image gluster://ovirt001/vol1/a87a7ef6-2c74-4d8e-a6e0-a392d0f791cf/images/238cc6cf-070c-4483-b686-c0de7ddf0dfa/ff2bca2d-4ed1-46c6-93c8-22a39bb1626a: No such file or directory . VM VM1 was started by admin@internal (Host: ovirt001).
Do you see any errors in glusterd log while trying to run the VM? Log file can be found at (/var/log/glusterfs/...) on ovirt001. Thanks, Vijay