On VM start

ovirt001# tail -f /var/log/glusterfs/*

==> cli.log <==
[2013-07-17 20:23:09.187585] W [rpc-transport.c:175:rpc_transport_load] 0-rpc-transport: missing 'option transport-type'. defaulting to "socket"
[2013-07-17 20:23:09.189638] I [socket.c:3480:socket_init] 0-glusterfs: SSL support is NOT enabled
[2013-07-17 20:23:09.189660] I [socket.c:3495:socket_init] 0-glusterfs: using system polling thread

==> etc-glusterfs-glusterd.vol.log <==
[2013-07-17 20:23:09.252219] I [glusterd-handler.c:1007:__glusterd_handle_cli_get_volume] 0-glusterd: Received get vol req

==> cli.log <==
[2013-07-17 20:23:09.252506] I [cli-rpc-ops.c:538:gf_cli_get_volume_cbk] 0-cli: Received resp to get vol: 0
[2013-07-17 20:23:09.252961] E [cli-xml-output.c:2572:cli_xml_output_vol_info_end] 0-cli: Returning 0
[2013-07-17 20:23:09.252998] I [cli-rpc-ops.c:771:gf_cli_get_volume_cbk] 0-cli: Returning: 0
[2013-07-17 20:23:09.253107] I [input.c:36:cli_batch] 0-: Exiting with: 0

ovirt001# tail -f /var/log/vdsm/*
(see attachment)

Steve Dainard 
Infrastructure Manager
Miovision | Rethink Traffic
519-513-2407 ex.250
877-646-8476 (toll-free)

Blog  |  LinkedIn  |  Twitter  |  Facebook

Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON, Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If you are not the intended recipient, please delete the e-mail and any attachments and notify us immediately.


On Wed, Jul 17, 2013 at 2:40 PM, Vijay Bellur <vbellur@redhat.com> wrote:
On 07/17/2013 10:20 PM, Steve Dainard wrote:
Completed changes:

*gluster> volume info vol1*

Volume Name: vol1
Type: Replicate
Volume ID: 97c3b2a7-0391-4fae-b541-cf04ce6bde0f
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: ovirt001.miovision.corp:/mnt/storage1/vol1
Brick2: ovirt002.miovision.corp:/mnt/storage1/vol1
Options Reconfigured:
network.remote-dio: on
cluster.eager-lock: enable
performance.stat-prefetch: off
performance.io-cache: off
performance.read-ahead: off
performance.quick-read: off
storage.owner-gid: 36
storage.owner-uid: 36
auth.allow: *
user.cifs: on
nfs.disable: off
server.allow-insecure: on


*Same error on VM run:*

VM VM1 is down. Exit message: internal error process exited while
connecting to monitor: qemu-system-x86_64: -drive
file=gluster://ovirt001/vol1/a87a7ef6-2c74-4d8e-a6e0-a392d0f791cf/images/238cc6cf-070c-4483-b686-c0de7ddf0dfa/ff2bca2d-4ed1-46c6-93c8-22a39bb1626a,if=none,id=drive-virtio-disk0,format=raw,serial=238cc6cf-070c-4483-b686-c0de7ddf0dfa,cache=none,werror=stop,rerror=stop,aio=threads:
could not open disk image
gluster://ovirt001/vol1/a87a7ef6-2c74-4d8e-a6e0-a392d0f791cf/images/238cc6cf-070c-4483-b686-c0de7ddf0dfa/ff2bca2d-4ed1-46c6-93c8-22a39bb1626a:
No such file or directory .
VM VM1 was started by admin@internal (Host: ovirt001).

Do you see any errors in glusterd log while trying to run the VM? Log file can be found at (/var/log/glusterfs/...) on ovirt001.

Thanks,
Vijay