
Most probably a gluster bug. Best Regards, Strahil Nikolov В неделя, 15 ноември 2020 г., 22:31:24 Гринуич+2, suporte@logicworks.pt <suporte@logicworks.pt> написа: So, you think it's really a bug? ________________________________ De: "Nir Soffer" <nsoffer@redhat.com> Para: suporte@logicworks.pt Cc: "users" <users@ovirt.org>, "Sahina Bose" <sabose@redhat.com>, "Krutika Dhananjay" <kdhananj@redhat.com>, "Nisan, Tal" <tnisan@redhat.com> Enviadas: Domingo, 15 De Novembro de 2020 15:03:21 Assunto: Re: [ovirt-users] Cannot copy or move disks On Sat, Nov 14, 2020 at 4:45 PM <suporte@logicworks.pt> wrote:
Hello,
I just update to Version 4.4.3.11-1.el8. Engine and host
and now I cannot copy or move disks.
Storage domains are glusterfs
# gluster --version glusterfs 7.8
Here is what I found on vdsm.log
2020-11-14 14:08:16,917+0000 INFO (tasks/5) [storage.SANLock] Releasing Lease(name='01178644-2ad6-4d37-8657-f33f547bee6b', path='/rhev/data-center/mnt/glusterSD/node1-teste.aclou d.pt:_data2/83f8bbfd-cfa3-46d9-a823-c36054826d13/images/97977cbf-eecc-4476-a11f-7798425d40c4/01178644-2ad6-4d37-8657-f33f547bee6b.lease', offset=0) (clusterlock:530) 2020-11-14 14:08:17,015+0000 INFO (tasks/5) [storage.SANLock] Successfully released Lease(name='01178644-2ad6-4d37-8657-f33f547bee6b', path='/rhev/data-center/mnt/glusterSD/node1 -teste.acloud.pt:_data2/83f8bbfd-cfa3-46d9-a823-c36054826d13/images/97977cbf-eecc-4476-a11f-7798425d40c4/01178644-2ad6-4d37-8657-f33f547bee6b.lease', offset=0) (clusterlock:540) 2020-11-14 14:08:17,016+0000 ERROR (tasks/5) [root] Job '8cd732fc-d69b-4c32-8b35-e4a8e47396fb' failed (jobs:223) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/jobs.py", line 159, in run self._run() File "/usr/lib/python3.6/site-packages/vdsm/storage/sdm/api/copy_data.py", line 110, in _run self._operation.run() File "/usr/lib/python3.6/site-packages/vdsm/storage/qemuimg.py", line 374, in run for data in self._operation.watch(): File "/usr/lib/python3.6/site-packages/vdsm/storage/operation.py", line 106, in watch self._finalize(b"", err) File "/usr/lib/python3.6/site-packages/vdsm/storage/operation.py", line 179, in _finalize raise cmdutils.Error(self._cmd, rc, out, err) vdsm.common.cmdutils.Error: Command ['/usr/bin/qemu-img', 'convert', '-p', '-t', 'none', '-T', 'none', '-f', 'raw', '-O', 'raw', '/rhev/data-center/mnt/glusterSD/node1-teste.aclou d.pt:_data2/83f8bbfd-cfa3-46d9-a823-c36054826d13/images/789f6e50-b954-4dda-a6d5-077fdfb357d2/d95a3e83-74d2-40a6-9f8f-e6ae68794051', '/rhev/data-center/mnt/glusterSD/node1-teste.ac loud.pt:_data2/83f8bbfd-cfa3-46d9-a823-c36054826d13/images/97977cbf-eecc-4476-a11f-7798425d40c4/01178644-2ad6-4d37-8657-f33f547bee6b'] failed with rc=1 out=b'' err=bytearray(b'qem u-img: error while reading sector 260177858: No such file or directory\n')
This is an impossible error for read(), preadv() etc.
2020-11-14 14:08:17,017+0000 INFO (tasks/5) [root] Job '8cd732fc-d69b-4c32-8b35-e4a8e47396fb' will be deleted in 3600 seconds (jobs:251) 2020-11-14 14:08:17,017+0000 INFO (tasks/5) [storage.ThreadPool.WorkerThread] FINISH task 6cb1d496-d1ca-40b5-a488-a72982738bab (threadPool:151) 2020-11-14 14:08:17,316+0000 INFO (jsonrpc/2) [api.host] START getJobs(job_type='storage', job_ids=['8cd732fc-d69b-4c32-8b35-e4a8e47396fb']) from=::ffff:192.168.5.165,36616, flow _id=49320e0a-14fb-4cbb-bdfd-b2546c260bf7 (api:48)
This was reported here a long time ago with various versions of gluster. I don't think we got any response from gluster folks about it yet. Can you file an oVirt bug about this? Nir _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/V3F3OONLMQXAFY...