Can you enable vdsm debug level logs[1] so we can see the exact command?
[1] -
https://www.ovirt.org/develop/developer-guide/vdsm/log-files.html
On Sat, Apr 20, 2019 at 1:42 PM Callum Smith <callum(a)well.ox.ac.uk> wrote:
>
> Dear Benny,
>
> This is via VDSM, but a manual one fails too.
>
> Background: cloning of an image fails in any context, whether creating from template,
using Clone VM, etc etc. Tested on any 4.3+ on multiple NFS servers etc. Even tried
allowing root access and not, seems to not fix.
>
> gdb is blank and has no relevant output
>
> /var/log/messages
> Appear to be getting this regularly ~5mins:
> Apr 20 10:15:42 virthyp03 nfsidmap[57709]: nss_getpwnam: name 'nobody' does
not
> map into domain 'virt.in.bmrc.ox.ac.uk'
> Apr 20 10:15:42 virthyp03 nfsidmap[57713]: nss_name_to_gid: name 'nobody'
does n
> ot map into domain ‘virt.in.bmrc.ox.ac.uk'
>
> Some relevant area of /var/log/vdsm/vdsm.log
> 2019-04-20 10:17:19,644+0000 INFO (jsonrpc/5) [vdsm.api] START
copyImage(sdUUID=u'0e01f014-530b-4067-aa1d-4e9378626a9d',
spUUID=u'df02d494-56b9-11e9-a05b-00163e4d2a92', vmUUID='',
srcImgUUID=u'6597eede-9fa0-4451-84fc-9f9c070cb5f3',
srcVolUUID=u'765fa48b-2e77-4637-b4ca-e1affcd71e48',
dstImgUUID=u'e1c182ae-f25d-464c-b557-93088e894452',
dstVolUUID=u'a1157ad0-44a8-4073-a20c-468978973f4f', description=u'',
dstSdUUID=u'0e01f014-530b-4067-aa1d-4e9378626a9d', volType=8, volFormat=5,
preallocate=2, postZero=u'false', force=u'false', discard=False)
from=::ffff:10.141.31.240,44298, flow_id=82e5a31b-5010-4e50-a333-038b16738ecb,
task_id=d6bf22cf-1ea3-419d-8248-6c8842537a2e (api:48)
> 2019-04-20 10:17:19,649+0000 INFO (jsonrpc/5) [storage.Image] image
6597eede-9fa0-4451-84fc-9f9c070cb5f3 in domain 0e01f014-530b-4067-aa1d-4e9378626a9d has
vollist [u'765fa48b-2e77-4637-b4ca-e1affcd71e48'] (image:313)
> 2019-04-20 10:17:19,657+0000 INFO (jsonrpc/5) [storage.Image] Current
chain=765fa48b-2e77-4637-b4ca-e1affcd71e48 (top) (image:702)
> 2019-04-20 10:17:19,658+0000 INFO (jsonrpc/5) [IOProcessClient] (Global) Starting
client (__init__:308)
> 2019-04-20 10:17:19,668+0000 INFO (ioprocess/57971) [IOProcess] (Global) Starting
ioprocess (__init__:434)
> 2019-04-20 10:17:19,679+0000 INFO (jsonrpc/5) [vdsm.api] FINISH copyImage
return=None from=::ffff:10.141.31.240,44298, flow_id=82e5a31b-5010-4e50-a333-038b16738ecb,
task_id=d6bf22cf-1ea3-419d-8248-6c8842537a2e (api:54)
> 2019-04-20 10:17:19,696+0000 INFO (jsonrpc/5) [jsonrpc.JsonRpcServer] RPC call
Volume.copy succeeded in 0.05 seconds (__init__:312)
> 2019-04-20 10:17:19,697+0000 INFO (tasks/7) [storage.ThreadPool.WorkerThread] START
task d6bf22cf-1ea3-419d-8248-6c8842537a2e (cmd=<bound method Task.commit of
<vdsm.storage.task.Task instance at 0x7f9f40033c68>>, args=None)
(threadPool:208)
> 2019-04-20 10:17:19,713+0000 INFO (tasks/7) [storage.Image]
sdUUID=0e01f014-530b-4067-aa1d-4e9378626a9d vmUUID=
srcImgUUID=6597eede-9fa0-4451-84fc-9f9c070cb5f3
srcVolUUID=765fa48b-2e77-4637-b4ca-e1affcd71e48
dstImgUUID=e1c182ae-f25d-464c-b557-93088e894452
dstVolUUID=a1157ad0-44a8-4073-a20c-468978973f4f
dstSdUUID=0e01f014-530b-4067-aa1d-4e9378626a9d volType=8 volFormat=RAW preallocate=SPARSE
force=False postZero=False discard=False (image:724)
> 2019-04-20 10:17:19,719+0000 INFO (tasks/7) [storage.VolumeManifest] Volume:
preparing volume 0e01f014-530b-4067-aa1d-4e9378626a9d/765fa48b-2e77-4637-b4ca-e1affcd71e48
(volume:567)
> 2019-04-20 10:17:19,722+0000 INFO (tasks/7) [storage.Image] copy source
0e01f014-530b-4067-aa1d-4e9378626a9d:6597eede-9fa0-4451-84fc-9f9c070cb5f3:765fa48b-2e77-4637-b4ca-e1affcd71e48
size 104857600 blocks destination
0e01f014-530b-4067-aa1d-4e9378626a9d:e1c182ae-f25d-464c-b557-93088e894452:a1157ad0-44a8-4073-a20c-468978973f4f
allocating 104857600 blocks (image:767)
> 2019-04-20 10:17:19,722+0000 INFO (tasks/7) [storage.Image] image
e1c182ae-f25d-464c-b557-93088e894452 in domain 0e01f014-530b-4067-aa1d-4e9378626a9d has
vollist [] (image:313)
> 2019-04-20 10:17:19,723+0000 INFO (tasks/7) [storage.StorageDomain] Create
placeholder
/rhev/data-center/mnt/10.141.15.248:_export_instruct_vm__storage/0e01f014-530b-4067-aa1d-4e9378626a9d/images/e1c182ae-f25d-464c-b557-93088e894452
for image's volumes (sd:1288)
> 2019-04-20 10:17:19,731+0000 INFO (tasks/7) [storage.Volume] Creating volume
a1157ad0-44a8-4073-a20c-468978973f4f (volume:1183)
> 2019-04-20 10:17:19,754+0000 INFO (tasks/7) [storage.Volume] Request to create RAW
volume
/rhev/data-center/mnt/10.141.15.248:_export_instruct_vm__storage/0e01f014-530b-4067-aa1d-4e9378626a9d/images/e1c182ae-f25d-464c-b557-93088e894452/a1157ad0-44a8-4073-a20c-468978973f4f
with size = 20480 blocks (fileVolume:463)
> 2019-04-20 10:17:19,754+0000 INFO (tasks/7) [storage.Volume] Changing volume
u'/rhev/data-center/mnt/10.141.15.248:_export_instruct_vm__storage/0e01f014-530b-4067-aa1d-4e9378626a9d/images/e1c182ae-f25d-464c-b557-93088e894452/a1157ad0-44a8-4073-a20c-468978973f4f'
permission to 0660 (fileVolume:480)
> 2019-04-20 10:17:19,800+0000 INFO (tasks/7) [storage.VolumeManifest] Volume:
preparing volume 0e01f014-530b-4067-aa1d-4e9378626a9d/a1157ad0-44a8-4073-a20c-468978973f4f
(volume:567)
>
> I tried to filter the usual noise out of VDSM.log so hopefully this is the relevant
bit you need - let me know if the full thing would help.
>
> Regards,
> Callum
>
> --
>
> Callum Smith
> Research Computing Core
> Wellcome Trust Centre for Human Genetics
> University of Oxford
> e. callum(a)well.ox.ac.uk
>
> On 20 Apr 2019, at 10:54, Benny Zlotnik <bzlotnik(a)redhat.com> wrote:
>
> Sorry, I kind of lost track of what the problem is
>
> The "KeyError: 'appsList'" issue is a known bug[1]
>
> If a manual (not via vdsm) run of qemu-img is actually stuck, then
> let's involve the qemu-discuss list, with the version of the relevant
> packages (qemu, qemu-img, kernel, your distro) and the output of gdb
> commands
>
> [1] -
https://bugzilla.redhat.com/show_bug.cgi?id=1690301
>
>
>
> On Sat, Apr 20, 2019 at 1:36 AM Callum Smith <callum(a)well.ox.ac.uk> wrote:
>
>
> Dear Benny and others,
>
> So it seems I wasn’t being patient with GDB and it does show me some output. This
error of qemu-img convert even is failing and preventing updating ovirt-node version from
4.3.2 to 4.3.3.1. I get a feeling this is an unrelated error, but I thought I’d be
complete:
>
> Excuse any typos, im having to type this manually from a remote session, but the
error:
>
> [733272.427922] hid-generic 0003:0624:0249.0001: usb_submit_urb(ctrl) failed: -19
>
> If this bug is preventing even a local yum updatei can’t see how it’s any issue other
than somehow involved with the hardware of the hypervisor, our network and storage
configuration must be irrelevant to this fact at this stage?
>
> Regards,
> Callum
>
> --
>
> Callum Smith
> Research Computing Core
> Wellcome Trust Centre for Human Genetics
> University of Oxford
> e. callum(a)well.ox.ac.uk
>
> On 11 Apr 2019, at 12:00, Callum Smith <callum(a)well.ox.ac.uk> wrote:
>
> Without the sudo and running in a dir where the root has access to, gdb has zero
output:
>
> <Screenshot 2019-04-11 at 12.00.27.png>
> Regards,
> Callum
>
> --
>
> Callum Smith
> Research Computing Core
> Wellcome Trust Centre for Human Genetics
> University of Oxford
> e. callum(a)well.ox.ac.uk
>
> On 11 Apr 2019, at 11:54, Callum Smith <callum(a)well.ox.ac.uk> wrote:
>
> Some more information:
>
> running qemu-img convert manually having captured the failed attempt from the
previous:
>
> sudo -u vdsm /usr/bin/qemu-img convert -p -t none -T none -f raw
/rhev/data-center/mnt/10.141.15.248:_export_instruct_vm__storage/0e01f014-530b-4067-aa1d-4e9378626a9d/images/6597eede-9fa0-4451-84fc-9f9c070cb5f3/765fa48b-2e77-4637-b4ca-e1affcd71e48
-O raw
/rhev/data-center/mnt/10.141.15.248:_export_instruct_vm__storage/0e01f014-530b-4067-aa1d-4e9378626a9d/images/9cc99110-70a2-477f-b3ef-1031a912d12b/c2776107-4579-43a6-9d60-93a5ea9c64c5
-W
>
> Added the -W flag just to see what would happen:
>
> gdb -p 79913 -batch -ex "t a a bt"
> [Thread debugging using libthread_db enabled]
> Using host libthread_db library "/lib64/libthread_db.so.1".
> 0x00007f528d7661f0 in __poll_nocancel () from /lib64/libc.so.6
>
> Thread 1 (Thread 0x7f528e6bb840 (LWP 79913)):
> #0 0x00007f528d7661f0 in __poll_nocancel () from /lib64/libc.so.6
> #1 0x00007f528dc510fb in sudo_ev_scan_impl () from
/usr/libexec/sudo/libsudo_util.so.0
> #2 0x00007f528dc49b44 in sudo_ev_loop_v1 () from
/usr/libexec/sudo/libsudo_util.so.0
> #3 0x000055e94aa0e271 in exec_nopty ()
> #4 0x000055e94aa0afda in sudo_execute ()
> #5 0x000055e94aa18a12 in run_command ()
> #6 0x000055e94aa0969e in main ()
>
>
> And without -W:
>
> gdb -p 85235 -batch -ex "t a a bt"
> [Thread debugging using libthread_db enabled]
> Using host libthread_db library "/lib64/libthread_db.so.1".
> 0x00007fc0cc69b1f0 in __poll_nocancel () from /lib64/libc.so.6
>
> Thread 1 (Thread 0x7fc0cd5f0840 (LWP 85235)):
> #0 0x00007fc0cc69b1f0 in __poll_nocancel () from /lib64/libc.so.6
> #1 0x00007fc0ccb860fb in sudo_ev_scan_impl () from
/usr/libexec/sudo/libsudo_util.so.0
> #2 0x00007fc0ccb7eb44 in sudo_ev_loop_v1 () from
/usr/libexec/sudo/libsudo_util.so.0
> #3 0x00005610f4397271 in exec_nopty ()
> #4 0x00005610f4393fda in sudo_execute ()
> #5 0x00005610f43a1a12 in run_command ()
> #6 0x00005610f439269e in main ()
>
> Regards,
> Callum
>
> --
>
> Callum Smith
> Research Computing Core
> Wellcome Trust Centre for Human Genetics
> University of Oxford
> e. callum(a)well.ox.ac.uk
>
> On 11 Apr 2019, at 09:57, Callum Smith <callum(a)well.ox.ac.uk> wrote:
>
> Dear Benny,
>
> It would seem that even cloning a VM is failing, creating a VM works on the same
storage. This is the only error i could find:
>
> ERROR Internal server error
> Traceback (most recent
call last):
> File
"/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py", line 345, in
_handle_request
> res =
method(**params)
> File
"/usr/lib/python2.7/site-packages/vdsm/rpc/Bridge.py", line 194, in
_dynamicMethod
> result =
fn(*methodArgs)
> File
"<string>", line 2, in getAllVmStats
> File
"/usr/lib/python2.7/site-packages/vdsm/common/api.py", line 50, in method
> ret = func(*args,
**kwargs)
> File
"/usr/lib/python2.7/site-packages/vdsm/API.py", line 1388, in getAllVmStats
> statsList =
self._cif.getAllVmStats()
> File
"/usr/lib/python2.7/site-packages/vdsm/clientIF.py", line 567, in getAllVmStats
> return [v.getStats()
for v in self.vmContainer.values()]
> File
"/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 1766, in getStats
> oga_stats =
self._getGuestStats()
> File
"/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 1967, in
_getGuestStats
> stats =
self.guestAgent.getGuestInfo()
> File
"/usr/lib/python2.7/site-packages/vdsm/virt/guestagent.py", line 505, in
getGuestInfo
> del
qga['appsList']
> KeyError:
'appsList'
>
>
> It's the qemu-img convert for sure that's just failing to do anything, this
is the command from the clone:
> /usr/bin/qemu-img convert -p -t none -T none -f raw
/rhev/data-center/mnt/10.141.15.248:_export_instruct_vm__storage/0e01f014-530b-4067-aa1d-4e9378626a9d/images/6597eede-9fa0-4451-84fc-9f9c070cb5f3/765fa48b-2e77-4637-b4ca-e1affcd71e48
-O raw
/rhev/data-center/mnt/10.141.15.248:_export_instruct_vm__storage/0e01f014-530b-4067-aa1d-4e9378626a9d/images/f0700631-e60b-4c2a-a6f5-a6c818ae7651/d4fb05ec-7c78-4d89-9a66-614c093c6e16
>
> gdb has a blank output for this though. This means 4.3.2 is fairly unusable for us,
so two questions, can I downgrade to 4.2, and is there a fix coming in 4.3.3 for this?
>
> Regards,
> Callum
>
> --
>
> Callum Smith
> Research Computing Core
> Wellcome Trust Centre for Human Genetics
> University of Oxford
> e. callum(a)well.ox.ac.uk
>
> On 10 Apr 2019, at 11:22, Callum Smith <callum(a)well.ox.ac.uk> wrote:
>
> Creating a disk on the target share works fine. - This seems to specifically be an
issue to do with moving a disk to/from a share.
>
> Regards,
> Callum
>
> --
>
> Callum Smith
> Research Computing Core
> Wellcome Trust Centre for Human Genetics
> University of Oxford
> e. callum(a)well.ox.ac.uk
>
> On 10 Apr 2019, at 09:53, Callum Smith <callum(a)well.ox.ac.uk> wrote:
>
> gdb -p $(pidof qemu-img convert) -batch -ex "t a a bt"
> 289444: No such file or directory.
>
> Regards,
> Callum
>
> --
>
> Callum Smith
> Research Computing Core
> Wellcome Trust Centre for Human Genetics
> University of Oxford
> e. callum(a)well.ox.ac.uk
>
> On 10 Apr 2019, at 09:36, Benny Zlotnik <bzlotnik(a)redhat.com> wrote:
>
> Can you run:
> $ gdb -p $(pidof qemu-img convert) -batch -ex "t a a bt"
>
>
> On Wed, Apr 10, 2019 at 11:26 AM Callum Smith <callum(a)well.ox.ac.uk> wrote:
>
>
> Dear All,
>
> Further to this, I can't migrate a disk to different storage using the GUI. Both
disks are configured identically and on the same physical NFS provider.
>
> Regards,
> Callum
>
> --
>
> Callum Smith
> Research Computing Core
> Wellcome Trust Centre for Human Genetics
> University of Oxford
> e. callum(a)well.ox.ac.uk
>
> On 9 Apr 2019, at 12:12, Callum Smith <callum(a)well.ox.ac.uk> wrote:
>
> Dear All,
>
> It would seem this is a bug in 4.3.? - As upgrading the old oVirt HE to 4.3 (from
4.2.latest) now means that the export of VMs to export domain no longer works.
>
> Again qemu-img convert is using some cpu, but no network. Progress is 0.
>
> Regards,
> Callum
>
> --
>
> Callum Smith
> Research Computing Core
> Wellcome Trust Centre for Human Genetics
> University of Oxford
> e. callum(a)well.ox.ac.uk
>
> On 8 Apr 2019, at 15:42, Callum Smith <callum(a)well.ox.ac.uk> wrote:
>
> Dear All,
>
> We've exported some VMs from our old oVirt infrastructure and want to import them
into the new one, but qemu-img appears to be failing. We have mounted an export domain
populated from the old oVirt in the new hosted engine and are using the GUI to import the
VM. Manually running the command sits at 16% CPU, 0% network usage and no progress. It
appears to lock the NFS mount and ls and lsof both hang.
>
> sudo -u vdsm /usr/bin/qemu-img convert -p -t none -T none -f raw <path to
source> -O raw <path to dest>
>
> Conversely a simple cp will work (ruling out file permissions errors):
> sudo -u vddm cp <path to source> <path to test>
>
> What might we be doing wrong?
>
> Regards,
> Callum
>
> --
>
> Callum Smith
> Research Computing Core
> Wellcome Trust Centre for Human Genetics
> University of Oxford
> e. callum(a)well.ox.ac.uk
>
> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org
> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZUBKWQIGGSJ...
>
>
> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org
> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/6EE6X43GTAJ...
>
>
> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org
> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/GD2XCKEIPNC...
>
>
> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org
> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/SV3V34DQJZ7...
>
>
> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org
> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/MICXCNEDJL5...
>
>
> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org
> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/KRV6NPWUCOA...
>
>
>
> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org
> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/R53H5T4ZHFG...
>
>
>