On 05/17/2012 11:05 PM, Itamar Heim wrote:
On 05/17/2012 06:55 PM, Bharata B Rao wrote:
> On Wed, May 16, 2012 at 3:29 PM, Itamar Heim<iheim(a)redhat.com> wrote:
>> On 05/15/2012 07:35 PM, Andrei Vakhnin wrote:
>>>
>>> Yair
>>>
>>> Thanks for an update. Can I have KVM hypervisors also function as
>>> storage
>>> nodes for glusterfs? What is a release date for glusterfs support?
>>> We're
>>> looking for a production deployment in June. Thanks
>>
>>
>> current status is
>> 1. patches for provisioning gluster clusters and volumes via ovirt
>> are in
>> review, trying to cover this feature set [1].
>> I'm not sure if all of them will make the ovirt 3.1 version which is
>> slated
>> to branch for stabilization June 1st, but i think "enough" is there.
>> so i'd start trying current upstream version to help find issues
>> blocking
>> you, and following on them during june as we stabilize ovirt 3.1 for
>> release
>> (planned for end of june).
>>
>> 2. you should be able to use same hosts for both gluster and virt,
>> but there
>> is no special logic/handling for this yet (i.e., trying and providing
>> feedback would help improve this mode).
>> I would suggest start from separate clusters though first, and only
>> later
>> trying the joint mode.
>>
>> 3. creating a storage domain on top of gluster:
>> - expose NFS on top of it, and consume as a normal nfs storage domain
>> - use posixfs storage domain with gluster mount semantics
>> - future: probably native gluster storage domain, up to native
>> integration with qemu
>
> I am looking at GlusterFS integration with QEMU which involves adding
> GlusterFS as block backend in QEMU. This will involve QEMU talking to
> gluster directly via libglusterfs bypassing FUSE. I could specify a
> volume file and the VM image directly on QEMU command line to boot
> from the VM image that resides on a gluster volume.
>
> Eg: qemu -drive file=client.vol:/Fedora.img,format=gluster
>
> In this example, Fedora.img is being served by gluster and client.vol
> would have client-side translators specified.
>
> I am not sure if this use case would be served if GlusterFS is
> integrated as posixfs storage domain in VDSM. Posixfs would involve
> normal FUSE mount and QEMU would be required to work with images from
> FUSE mount path ?
>
> With QEMU supporting GlusterFS backend natively, further optimizations
> are possible in case of gluster volume being local to the host node.
> In this case, one could provide QEMU with a simple volume file that
> would not contain client or server xlators, but instead just the posix
> xlator. This would lead to most optimal IO path that bypasses RPC
> calls.
>
> So do you think, this use case (QEMU supporting GlusterFS backend
> natively and using volume file to specify the needed translators)
> warrants a specialized storage domain type for GlusterFS in VDSM ?
I'm not sure if a special storage domain, or a PosixFS based domain
with enhanced capabilities.
Ayal?
Related Question:
With QEMU using GlusterFS backend natively (as described above), it
also means that
it needs addnl options/parameters as part of qemu command line (as given
above).
How does VDSM today support generating a custom qemu cmdline. I know
VDSM talks to libvirt,
so is there a framework in VDSM to edit/modify the domxml based on some
pre-conditions,
and how / where one should hook up to do that modification ? I know of
libvirt hooks
framework in VDSM, but that was more for temporary/experimental needs,
or am i completely
wrong here ?
Irrespective of whether GlusterFS integrates into VDSM as PosixFS or
special storage domain
it won't address the need to generate a custom qemu cmdline if a
file/image was served by
GlusterFS. Whats the way to address this issue in VDSM ?
I am assuming here that special storage domain (aka repo engine) is only
to manage image
repository, and image related operations, won't help in modifying qemu
cmd line being generated.
[Ccing vdsm-devel also]
thanx,
deepak