[Users] Ovirt 3.3 nightly, Gluster 3.4 stable, cannot launch VM with gluster storage domain backed disk

Steve Dainard sdainard at miovision.com
Wed Jul 17 20:30:06 UTC 2013


On VM start

*ovirt001# tail -f /var/log/glusterfs/**

==> cli.log <==
[2013-07-17 20:23:09.187585] W [rpc-transport.c:175:rpc_transport_load]
0-rpc-transport: missing 'option transport-type'. defaulting to "socket"
[2013-07-17 20:23:09.189638] I [socket.c:3480:socket_init] 0-glusterfs: SSL
support is NOT enabled
[2013-07-17 20:23:09.189660] I [socket.c:3495:socket_init] 0-glusterfs:
using system polling thread

==> etc-glusterfs-glusterd.vol.log <==
[2013-07-17 20:23:09.252219] I
[glusterd-handler.c:1007:__glusterd_handle_cli_get_volume] 0-glusterd:
Received get vol req

==> cli.log <==
[2013-07-17 20:23:09.252506] I [cli-rpc-ops.c:538:gf_cli_get_volume_cbk]
0-cli: Received resp to get vol: 0
[2013-07-17 20:23:09.252961] E
[cli-xml-output.c:2572:cli_xml_output_vol_info_end] 0-cli: Returning 0
[2013-07-17 20:23:09.252998] I [cli-rpc-ops.c:771:gf_cli_get_volume_cbk]
0-cli: Returning: 0
[2013-07-17 20:23:09.253107] I [input.c:36:cli_batch] 0-: Exiting with: 0

*ovirt001# tail -f /var/log/vdsm/**
(see attachment)

*Steve Dainard *
Infrastructure Manager
Miovision <http://miovision.com/> | *Rethink Traffic*
519-513-2407 ex.250
877-646-8476 (toll-free)

*Blog <http://miovision.com/blog>  |
**LinkedIn<https://www.linkedin.com/company/miovision-technologies>  |
 Twitter <https://twitter.com/miovision>  |
Facebook<https://www.facebook.com/miovision>
*
------------------------------
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If
you are not the intended recipient, please delete the e-mail and any
attachments and notify us immediately.


On Wed, Jul 17, 2013 at 2:40 PM, Vijay Bellur <vbellur at redhat.com> wrote:

> On 07/17/2013 10:20 PM, Steve Dainard wrote:
>
>> Completed changes:
>>
>> *gluster> volume info vol1*
>>
>> Volume Name: vol1
>> Type: Replicate
>> Volume ID: 97c3b2a7-0391-4fae-b541-**cf04ce6bde0f
>> Status: Started
>> Number of Bricks: 1 x 2 = 2
>> Transport-type: tcp
>> Bricks:
>> Brick1: ovirt001.miovision.corp:/mnt/**storage1/vol1
>> Brick2: ovirt002.miovision.corp:/mnt/**storage1/vol1
>> Options Reconfigured:
>> network.remote-dio: on
>> cluster.eager-lock: enable
>> performance.stat-prefetch: off
>> performance.io-cache: off
>> performance.read-ahead: off
>> performance.quick-read: off
>> storage.owner-gid: 36
>> storage.owner-uid: 36
>> auth.allow: *
>> user.cifs: on
>> nfs.disable: off
>> server.allow-insecure: on
>>
>>
>> *Same error on VM run:*
>>
>> VM VM1 is down. Exit message: internal error process exited while
>> connecting to monitor: qemu-system-x86_64: -drive
>> file=gluster://ovirt001/vol1/**a87a7ef6-2c74-4d8e-a6e0-**
>> a392d0f791cf/images/238cc6cf-**070c-4483-b686-c0de7ddf0dfa/**
>> ff2bca2d-4ed1-46c6-93c8-**22a39bb1626a,if=none,id=drive-**
>> virtio-disk0,format=raw,**serial=238cc6cf-070c-4483-**
>> b686-c0de7ddf0dfa,cache=none,**werror=stop,rerror=stop,aio=**threads:
>> could not open disk image
>> gluster://ovirt001/vol1/**a87a7ef6-2c74-4d8e-a6e0-**
>> a392d0f791cf/images/238cc6cf-**070c-4483-b686-c0de7ddf0dfa/**
>> ff2bca2d-4ed1-46c6-93c8-**22a39bb1626a:
>> No such file or directory .
>> VM VM1 was started by admin at internal (Host: ovirt001).
>>
>
> Do you see any errors in glusterd log while trying to run the VM? Log file
> can be found at (/var/log/glusterfs/...) on ovirt001.
>
> Thanks,
> Vijay
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20130717/ce1a6e0f/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: vdsm-logs
Type: application/octet-stream
Size: 39663 bytes
Desc: not available
URL: <http://lists.ovirt.org/pipermail/users/attachments/20130717/ce1a6e0f/attachment-0001.obj>


More information about the Users mailing list