On 10/15/2020 12:07 PM, Michael Thomas wrote:
On 10/15/20 10:19 AM, Jeff Bailey wrote:

On 10/15/2020 10:01 AM, Michael Thomas wrote:
Getting closer...

I recreated the storage domain and added rbd_default_features=3 to ceph.conf.  Now I see the new disk being created with (what I think is) the correct set of features:

# rbd info rbd.ovirt.data/volume-f4ac68c6-e5f7-4b01-aed0-36a55b901
fbf
rbd image 'volume-f4ac68c6-e5f7-4b01-aed0-36a55b901fbf':
        size 100 GiB in 25600 objects
        order 22 (4 MiB objects)
        snapshot_count: 0
        id: 70aab541cb331
        block_name_prefix: rbd_data.70aab541cb331
        format: 2
        features: layering
        op_features:
        flags:
        create_timestamp: Thu Oct 15 06:53:23 2020
        access_timestamp: Thu Oct 15 06:53:23 2020
        modify_timestamp: Thu Oct 15 06:53:23 2020

However, I'm still unable to attach the disk to a VM.  This time it's a permissions issue on the ovirt node where the VM is running.  It looks like it can't read the temporary ceph config file that is sent over from the engine:


Are you using octopus?  If so, the config file that's generated is missing the "[global]" at the top and octopus doesn't like that.  It's been patched upstream.

Yes, I am using Octopus (15.2.4).  Do you have a pointer to the upstream patch or issue so that I can watch for a release with the fix?


https://bugs.launchpad.net/cinder/+bug/1865754


It's a simple fix.  I just changed line 100 of /usr/lib/python3.6/site-packages/os_brick/initiator/connectors/rbd.py to:

conf_file.writelines(["[global]", "\n", mon_hosts, "\n", keyring, "\n"])



--Mike