I rebooted it with no luck, them I used the auto-gen snapshot , same luck.attaching the logs in gdrivethanks in advance2018-05-11 12:50 GMT-03:00 Benny Zlotnik <bzlotnik@redhat.com>:I see here a failed attempt:2018-05-09 16:00:20,129-03 ERROR [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLo gDirector] (EE-ManagedThreadFactory-engin eScheduled-Thread-67) [bd8eeb1d-f49a-4f91-a521-e0f31 b4a7cbd] EVENT_ID: USER_MOVED_DISK_FINISHED_FAILU RE(2,011), User admin@internal-authz have failed to move disk mail02-int_Disk1 to domain 2penLA. Then another:2018-05-09 16:15:06,998-03 ERROR [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLo gDirector] (EE-ManagedThreadFactory-engin eScheduled-Thread-34) [] EVENT_ID: USER_MOVED_DISK_FINISHED_FAILU RE(2,011), User admin@internal-authz have failed to move disk mail02-int_Disk1 to domain 2penLA. Here I see a successful attempt:2018-05-09 21:58:42,628-03 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLo gDirector] (default task-50) [940b051c-8c63-4711-baf9-f3520 bb2b825] EVENT_ID: USER_MOVED_DISK(2,008), User admin@internal-authz moving disk mail02-int_Disk1 to domain 2penLA. Then, in the last attempt I see the attempt was successful but live merge failed:2018-05-11 03:37:59,509-03 ERROR [org.ovirt.engine.core.bll.MergeStatusCommand] (EE-ManagedThreadFactory-comma ndCoordinator-Thread-2) [d5b7fdf5-9c37-4c1f-8543-a7bc7 5c993a5] Failed to live merge, still in volume chain: [5d9d2958-96bc-49fa-9100-2f33a 3ba737f, 52532d05-970e-4643-9774-96c317 96062c] 2018-05-11 03:38:01,495-03 INFO [org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallb ack] (EE-ManagedThreadFactory-engin eScheduled-Thread-51) [d5b7fdf5-9c37-4c1f-8543-a7bc7 5c993a5] Command 'LiveMigrateDisk' (id: '115fc375-6018-4d59-b9f2-51ee0 5ca49f8') waiting on child command id: '26bc52a4-4509-4577-b342-44a67 9bc628f' type:'RemoveSnapshot' to complete 2018-05-11 03:38:01,501-03 ERROR [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDis kLiveCommand] (EE-ManagedThreadFactory-engin eScheduled-Thread-51) [d5b7fdf5-9c37-4c1f-8543-a7bc7 5c993a5] Command id: '4936d196-a891-4484-9cf5-fceaa fbf3364 failed child command status for step 'MERGE_STATUS' 2018-05-11 03:38:01,501-03 INFO [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDis kLiveCommandCallback] (EE-ManagedThreadFactory-engin eScheduled-Thread-51) [d5b7fdf5-9c37-4c1f-8543-a7bc7 5c993a5] Command 'RemoveSnapshotSingleDiskLive' id: '4936d196-a891-4484-9cf5-fceaa fbf3364' child commands '[8da5f261-7edd-4930-8d9d-d34f 232d84b3, 1c320f4b-7296-43c4-a3e6-8a868e 23fc35, a0e9e70c-cd65-4dfb-bd00-076c4e 99556a]' executions were completed, status 'FAILED' 2018-05-11 03:38:02,513-03 ERROR [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDis kLiveCommand] (EE-ManagedThreadFactory-engin eScheduled-Thread-2) [d5b7fdf5-9c37-4c1f-8543-a7bc7 5c993a5] Merging of snapshot '319e8bbb-9efe-4de4-a9a6-862e3 deb891f' images '52532d05-970e-4643-9774-96c31 796062c'..'5d9d2958-96bc-49fa- 9100-2f33a3ba737f' failed. Images have been marked illegal and can no longer be previewed or reverted to. Please retry Live Merge on the snapshot to complete the operation. 2018-05-11 03:38:02,519-03 ERROR [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDis kLiveCommand] (EE-ManagedThreadFactory-engin eScheduled-Thread-2) [d5b7fdf5-9c37-4c1f-8543-a7bc7 5c993a5] Ending command 'org.ovirt.engine.core.bll.sna pshots.RemoveSnapshotSingleDis kLiveCommand' with failure. 2018-05-11 03:38:03,530-03 INFO [org.ovirt.engine.core.bll.ConcurrentChildCommandsExecutionC allback] (EE-ManagedThreadFactory-engin eScheduled-Thread-37) [d5b7fdf5-9c37-4c1f-8543-a7bc7 5c993a5] Command 'RemoveSnapshot' id: '26bc52a4-4509-4577-b342-44a67 9bc628f' child commands '[4936d196-a891-4484-9cf5-fcea afbf3364]' executions were completed, status 'FAILED' 2018-05-11 03:38:04,548-03 ERROR [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand] (EE-ManagedThreadFactory-engin eScheduled-Thread-66) [d5b7fdf5-9c37-4c1f-8543-a7bc7 5c993a5] Ending command 'org.ovirt.engine.core.bll.sna pshots.RemoveSnapshotCommand' with failure. 2018-05-11 03:38:04,557-03 INFO [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand] (EE-ManagedThreadFactory-engin eScheduled-Thread-66) [d5b7fdf5-9c37-4c1f-8543-a7bc7 5c993a5] Lock freed to object 'EngineLock:{exclusiveLocks='[ 4808bb70-c9cc-4286-aa39-16b579 8213ac=LIVE_STORAGE_MIGRATION] ', sharedLocks=''}' I do not see the merge attempt in the vdsm.log, so please send vdsm logs for node02.phy.eze.ampgn.com.ar from that time.Also, did you use the auto-generated snapshot to start the vm?On Fri, May 11, 2018 at 6:11 PM, Juan Pablo <pablo.localhost@gmail.com> wrote:after the xfs_repair, it says: sorry I could not find valid secondary superblock2018-05-11 12:09 GMT-03:00 Juan Pablo <pablo.localhost@gmail.com>:hi,Alias:mail02-int_Disk1Description:ID:65ec515e-0aae-4fe6-a561-387929c7fb4d Alignment:UnknownDisk Profile:Wipe After Delete:Nothat one2018-05-11 11:12 GMT-03:00 Benny Zlotnik <bzlotnik@redhat.com>:I looked at the logs and I see some disks have moved successfully and some failed. Which disk is causing the problems?On Fri, May 11, 2018 at 5:02 PM, Juan Pablo <pablo.localhost@gmail.com> wrote:Hi, just sent you via drive the files. attaching some extra info, thanks thanks and thanks :from inside the migrated vm I had the following attached dmesg output before rebootingregards and thanks again for the help,2018-05-11 10:45 GMT-03:00 Benny Zlotnik <bzlotnik@redhat.com>:Dropbox or google drive I guess. Also, can you attach engine.log?On Fri, May 11, 2018 at 4:43 PM, Juan Pablo <pablo.localhost@gmail.com> wrote:
vdsm is too big for gmail ...any other way I can share it with you?---------- Forwrded message ----------
From: Juan Pablo <pablo.localhost@gmail.com>
Date: 2018-05-11 10:40 GMT-03:00
Subject: Re: [ovirt-users] strange issue: vm lost info on disk
To: Benny Zlotnik <bzlotnik@redhat.com>
Cc: users <Users@ovirt.org>Benny, thanks for your reply! it was a Live migration. sorry, it was from nfs to iscsi, not otherwise. I have reboot the vm for rescue and it does not detect any partitions with fdisk, Im running a xfs_repair with -n and found some corrupted primary superblock., its still running... ( so... there's info in the disk maybe?)attaching logs, let me know if those are the ones.thanks again!2018-05-11 9:45 GMT-03:00 Benny Zlotnik <bzlotnik@redhat.com>:Can you provide the logs? engine and vdsm.Did you perform a live migration (the VM is running) or cold?On Fri, May 11, 2018 at 2:49 PM, Juan Pablo <pablo.localhost@gmail.com> wrote:______________________________Hi! , Im strugled about an ongoing problem:
after migrating a vm's disk from an iscsi domain to a nfs and ovirt reporting the migration was successful, I see there's no data 'inside' the vm's disk. we never had this issues with ovirt so Im stranged about the root cause and if theres a chance of recovering the information.can you please help me out troubleshooting this one? I would really appreciate it =)running ovirt 4.2.1 here!thanks in advance,JP_________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org