I removed the auto-snapshot and still no lucky. no bootable disk found. =(
ideas?
2018-05-13 12:26 GMT-03:00 Juan Pablo <pablo.localhost(a)gmail.com>:
benny, thanks for your reply:
ok, so the steps are : removing the snapshot on the first place. then what
do you suggest?
2018-05-12 15:23 GMT-03:00 Nir Soffer <nsoffer(a)redhat.com>:
> On Sat, 12 May 2018, 11:32 Benny Zlotnik, <bzlotnik(a)redhat.com> wrote:
>
>> Using the auto-generated snapshot is generally a bad idea as it's
>> inconsistent,
>>
>
> What do you mean by inconsistant?
>
>
> you should remove it before moving further
>>
>> On Fri, May 11, 2018 at 7:25 PM, Juan Pablo <pablo.localhost(a)gmail.com>
>> wrote:
>>
>>> I rebooted it with no luck, them I used the auto-gen snapshot , same
>>> luck.
>>> attaching the logs in gdrive
>>>
>>> thanks in advance
>>>
>>> 2018-05-11 12:50 GMT-03:00 Benny Zlotnik <bzlotnik(a)redhat.com>:
>>>
>>>> I see here a failed attempt:
>>>> 2018-05-09 16:00:20,129-03 ERROR [org.ovirt.engine.core.dal.dbb
>>>> roker.auditloghandling.AuditLogDirector]
>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-67)
>>>> [bd8eeb1d-f49a-4f91-a521-e0f31b4a7cbd] EVENT_ID:
>>>> USER_MOVED_DISK_FINISHED_FAILURE(2,011), User admin@internal-authz
>>>> have failed to move disk mail02-int_Disk1 to domain 2penLA.
>>>>
>>>> Then another:
>>>> 2018-05-09 16:15:06,998-03 ERROR [org.ovirt.engine.core.dal.dbb
>>>> roker.auditloghandling.AuditLogDirector]
>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-34) [] EVENT_ID:
>>>> USER_MOVED_DISK_FINISHED_FAILURE(2,011), User admin@internal-authz
>>>> have failed to move disk mail02-int_Disk1 to domain 2penLA.
>>>>
>>>> Here I see a successful attempt:
>>>> 2018-05-09 21:58:42,628-03 INFO [org.ovirt.engine.core.dal.dbb
>>>> roker.auditloghandling.AuditLogDirector] (default task-50)
>>>> [940b051c-8c63-4711-baf9-f3520bb2b825] EVENT_ID:
>>>> USER_MOVED_DISK(2,008), User admin@internal-authz moving disk
>>>> mail02-int_Disk1 to domain 2penLA.
>>>>
>>>>
>>>> Then, in the last attempt I see the attempt was successful but live
>>>> merge failed:
>>>> 2018-05-11 03:37:59,509-03 ERROR
[org.ovirt.engine.core.bll.MergeStatusCommand]
>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2)
>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Failed to live merge, still in
>>>> volume chain: [5d9d2958-96bc-49fa-9100-2f33a3ba737f,
>>>> 52532d05-970e-4643-9774-96c31796062c]
>>>> 2018-05-11 03:38:01,495-03 INFO [org.ovirt.engine.core.bll.Ser
>>>> ialChildCommandsExecutionCallback]
(EE-ManagedThreadFactory-engineScheduled-Thread-51)
>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command 'LiveMigrateDisk'
(id:
>>>> '115fc375-6018-4d59-b9f2-51ee05ca49f8') waiting on child command
id:
>>>> '26bc52a4-4509-4577-b342-44a679bc628f'
type:'RemoveSnapshot' to
>>>> complete
>>>> 2018-05-11 03:38:01,501-03 ERROR [org.ovirt.engine.core.bll.sna
>>>> pshots.RemoveSnapshotSingleDiskLiveCommand]
>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-51)
>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command id:
>>>> '4936d196-a891-4484-9cf5-fceaafbf3364 failed child command status
for
>>>> step 'MERGE_STATUS'
>>>> 2018-05-11 03:38:01,501-03 INFO [org.ovirt.engine.core.bll.sna
>>>> pshots.RemoveSnapshotSingleDiskLiveCommandCallback]
>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-51)
>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command
>>>> 'RemoveSnapshotSingleDiskLive' id:
'4936d196-a891-4484-9cf5-fceaafbf3364'
>>>> child commands '[8da5f261-7edd-4930-8d9d-d34f232d84b3,
>>>> 1c320f4b-7296-43c4-a3e6-8a868e23fc35,
a0e9e70c-cd65-4dfb-bd00-076c4e99556a]'
>>>> executions were completed, status 'FAILED'
>>>> 2018-05-11 03:38:02,513-03 ERROR [org.ovirt.engine.core.bll.sna
>>>> pshots.RemoveSnapshotSingleDiskLiveCommand]
>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-2)
>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Merging of snapshot
>>>> '319e8bbb-9efe-4de4-a9a6-862e3deb891f' images
>>>>
'52532d05-970e-4643-9774-96c31796062c'..'5d9d2958-96bc-49fa-9100-2f33a3ba737f'
>>>> failed. Images have been marked illegal and can no longer be previewed
or
>>>> reverted to. Please retry Live Merge on the snapshot to complete the
>>>> operation.
>>>> 2018-05-11 03:38:02,519-03 ERROR [org.ovirt.engine.core.bll.sna
>>>> pshots.RemoveSnapshotSingleDiskLiveCommand]
>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-2)
>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command
>>>>
'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand'
>>>> with failure.
>>>> 2018-05-11 03:38:03,530-03 INFO [org.ovirt.engine.core.bll.Con
>>>> currentChildCommandsExecutionCallback]
(EE-ManagedThreadFactory-engineScheduled-Thread-37)
>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command 'RemoveSnapshot'
id:
>>>> '26bc52a4-4509-4577-b342-44a679bc628f' child commands
>>>> '[4936d196-a891-4484-9cf5-fceaafbf3364]' executions were
completed,
>>>> status 'FAILED'
>>>> 2018-05-11 03:38:04,548-03 ERROR
[org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand]
>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-66)
>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command
>>>> 'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand' with
>>>> failure.
>>>> 2018-05-11 03:38:04,557-03 INFO
[org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand]
>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-66)
>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Lock freed to object
>>>> 'EngineLock:{exclusiveLocks='[4808bb70-c9cc-4286-aa39-16b579
>>>> 8213ac=LIVE_STORAGE_MIGRATION]', sharedLocks=''}'
>>>>
>>>> I do not see the merge attempt in the vdsm.log, so please send vdsm
>>>> logs for node02.phy.eze.ampgn.com.ar from that time.
>>>>
>>>> Also, did you use the auto-generated snapshot to start the vm?
>>>>
>>>>
>>>> On Fri, May 11, 2018 at 6:11 PM, Juan Pablo
<pablo.localhost(a)gmail.com
>>>> > wrote:
>>>>
>>>>> after the xfs_repair, it says: sorry I could not find valid
secondary
>>>>> superblock
>>>>>
>>>>> 2018-05-11 12:09 GMT-03:00 Juan Pablo
<pablo.localhost(a)gmail.com>:
>>>>>
>>>>>> hi,
>>>>>> Alias:
>>>>>> mail02-int_Disk1
>>>>>> Description:
>>>>>> ID:
>>>>>> 65ec515e-0aae-4fe6-a561-387929c7fb4d
>>>>>> Alignment:
>>>>>> Unknown
>>>>>> Disk Profile:
>>>>>> Wipe After Delete:
>>>>>> No
>>>>>>
>>>>>> that one
>>>>>>
>>>>>> 2018-05-11 11:12 GMT-03:00 Benny Zlotnik
<bzlotnik(a)redhat.com>:
>>>>>>
>>>>>>> I looked at the logs and I see some disks have moved
successfully
>>>>>>> and some failed. Which disk is causing the problems?
>>>>>>>
>>>>>>> On Fri, May 11, 2018 at 5:02 PM, Juan Pablo <
>>>>>>> pablo.localhost(a)gmail.com> wrote:
>>>>>>>
>>>>>>>> Hi, just sent you via drive the files. attaching some
extra info,
>>>>>>>> thanks thanks and thanks :
>>>>>>>>
>>>>>>>> from inside the migrated vm I had the following attached
dmesg
>>>>>>>> output before rebooting
>>>>>>>>
>>>>>>>> regards and thanks again for the help,
>>>>>>>>
>>>>>>>> 2018-05-11 10:45 GMT-03:00 Benny Zlotnik
<bzlotnik(a)redhat.com>:
>>>>>>>>
>>>>>>>>> Dropbox or google drive I guess. Also, can you attach
engine.log?
>>>>>>>>>
>>>>>>>>> On Fri, May 11, 2018 at 4:43 PM, Juan Pablo <
>>>>>>>>> pablo.localhost(a)gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> vdsm is too big for gmail ...any other way I can
share it with
>>>>>>>>>> you?
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> ---------- Forwrded message ----------
>>>>>>>>>> From: Juan Pablo
<pablo.localhost(a)gmail.com>
>>>>>>>>>> Date: 2018-05-11 10:40 GMT-03:00
>>>>>>>>>> Subject: Re: [ovirt-users] strange issue: vm lost
info on disk
>>>>>>>>>> To: Benny Zlotnik <bzlotnik(a)redhat.com>
>>>>>>>>>> Cc: users <Users(a)ovirt.org>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Benny, thanks for your reply! it was a Live
migration. sorry, it
>>>>>>>>>> was from nfs to iscsi, not otherwise. I have
reboot the vm for rescue and
>>>>>>>>>> it does not detect any partitions with fdisk, Im
running a xfs_repair with
>>>>>>>>>> -n and found some corrupted primary superblock.,
its still running... (
>>>>>>>>>> so... there's info in the disk maybe?)
>>>>>>>>>>
>>>>>>>>>> attaching logs, let me know if those are the
ones.
>>>>>>>>>> thanks again!
>>>>>>>>>>
>>>>>>>>>> 2018-05-11 9:45 GMT-03:00 Benny Zlotnik
<bzlotnik(a)redhat.com>:
>>>>>>>>>>
>>>>>>>>>>> Can you provide the logs? engine and vdsm.
>>>>>>>>>>> Did you perform a live migration (the VM is
running) or cold?
>>>>>>>>>>>
>>>>>>>>>>> On Fri, May 11, 2018 at 2:49 PM, Juan Pablo
<
>>>>>>>>>>> pablo.localhost(a)gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Hi! , Im strugled about an ongoing
problem:
>>>>>>>>>>>> after migrating a vm's disk from an
iscsi domain to a nfs and
>>>>>>>>>>>> ovirt reporting the migration was
successful, I see there's no data
>>>>>>>>>>>> 'inside' the vm's disk. we
never had this issues with ovirt so Im stranged
>>>>>>>>>>>> about the root cause and if theres a
chance of recovering the information.
>>>>>>>>>>>>
>>>>>>>>>>>> can you please help me out
troubleshooting this one? I would
>>>>>>>>>>>> really appreciate it =)
>>>>>>>>>>>> running ovirt 4.2.1 here!
>>>>>>>>>>>>
>>>>>>>>>>>> thanks in advance,
>>>>>>>>>>>> JP
>>>>>>>>>>>>
>>>>>>>>>>>>
_______________________________________________
>>>>>>>>>>>> Users mailing list -- users(a)ovirt.org
>>>>>>>>>>>> To unsubscribe send an email to
users-leave(a)ovirt.org
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>> _______________________________________________
>> Users mailing list -- users(a)ovirt.org
>> To unsubscribe send an email to users-leave(a)ovirt.org
>>
>