On Sat, 12 May 2018, 11:32 Benny Zlotnik, <bzlotnik(a)redhat.com> wrote:
Using the auto-generated snapshot is generally a bad idea as
it's
inconsistent,
What do you mean by inconsistant?
you should remove it before moving further
On Fri, May 11, 2018 at 7:25 PM, Juan Pablo <pablo.localhost(a)gmail.com>
wrote:
> I rebooted it with no luck, them I used the auto-gen snapshot , same luck.
> attaching the logs in gdrive
>
> thanks in advance
>
> 2018-05-11 12:50 GMT-03:00 Benny Zlotnik <bzlotnik(a)redhat.com>:
>
>> I see here a failed attempt:
>> 2018-05-09 16:00:20,129-03 ERROR
>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>> (EE-ManagedThreadFactory-engineScheduled-Thread-67)
>> [bd8eeb1d-f49a-4f91-a521-e0f31b4a7cbd] EVENT_ID:
>> USER_MOVED_DISK_FINISHED_FAILURE(2,011), User admin@internal-authz have
>> failed to move disk mail02-int_Disk1 to domain 2penLA.
>>
>> Then another:
>> 2018-05-09 16:15:06,998-03 ERROR
>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>> (EE-ManagedThreadFactory-engineScheduled-Thread-34) [] EVENT_ID:
>> USER_MOVED_DISK_FINISHED_FAILURE(2,011), User admin@internal-authz have
>> failed to move disk mail02-int_Disk1 to domain 2penLA.
>>
>> Here I see a successful attempt:
>> 2018-05-09 21:58:42,628-03 INFO
>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>> (default task-50) [940b051c-8c63-4711-baf9-f3520bb2b825] EVENT_ID:
>> USER_MOVED_DISK(2,008), User admin@internal-authz moving disk
>> mail02-int_Disk1 to domain 2penLA.
>>
>>
>> Then, in the last attempt I see the attempt was successful but live
>> merge failed:
>> 2018-05-11 03:37:59,509-03 ERROR
>> [org.ovirt.engine.core.bll.MergeStatusCommand]
>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2)
>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Failed to live merge, still in
>> volume chain: [5d9d2958-96bc-49fa-9100-2f33a3ba737f,
>> 52532d05-970e-4643-9774-96c31796062c]
>> 2018-05-11 03:38:01,495-03 INFO
>> [org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback]
>> (EE-ManagedThreadFactory-engineScheduled-Thread-51)
>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command 'LiveMigrateDisk' (id:
>> '115fc375-6018-4d59-b9f2-51ee05ca49f8') waiting on child command id:
>> '26bc52a4-4509-4577-b342-44a679bc628f' type:'RemoveSnapshot' to
complete
>> 2018-05-11 03:38:01,501-03 ERROR
>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand]
>> (EE-ManagedThreadFactory-engineScheduled-Thread-51)
>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command id:
>> '4936d196-a891-4484-9cf5-fceaafbf3364 failed child command status for step
>> 'MERGE_STATUS'
>> 2018-05-11 03:38:01,501-03 INFO
>>
[org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommandCallback]
>> (EE-ManagedThreadFactory-engineScheduled-Thread-51)
>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command
>> 'RemoveSnapshotSingleDiskLive' id:
'4936d196-a891-4484-9cf5-fceaafbf3364'
>> child commands '[8da5f261-7edd-4930-8d9d-d34f232d84b3,
>> 1c320f4b-7296-43c4-a3e6-8a868e23fc35,
>> a0e9e70c-cd65-4dfb-bd00-076c4e99556a]' executions were completed, status
>> 'FAILED'
>> 2018-05-11 03:38:02,513-03 ERROR
>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand]
>> (EE-ManagedThreadFactory-engineScheduled-Thread-2)
>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Merging of snapshot
>> '319e8bbb-9efe-4de4-a9a6-862e3deb891f' images
>>
'52532d05-970e-4643-9774-96c31796062c'..'5d9d2958-96bc-49fa-9100-2f33a3ba737f'
>> failed. Images have been marked illegal and can no longer be previewed or
>> reverted to. Please retry Live Merge on the snapshot to complete the
>> operation.
>> 2018-05-11 03:38:02,519-03 ERROR
>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand]
>> (EE-ManagedThreadFactory-engineScheduled-Thread-2)
>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command
>>
'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand'
>> with failure.
>> 2018-05-11 03:38:03,530-03 INFO
>> [org.ovirt.engine.core.bll.ConcurrentChildCommandsExecutionCallback]
>> (EE-ManagedThreadFactory-engineScheduled-Thread-37)
>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command 'RemoveSnapshot' id:
>> '26bc52a4-4509-4577-b342-44a679bc628f' child commands
>> '[4936d196-a891-4484-9cf5-fceaafbf3364]' executions were completed,
status
>> 'FAILED'
>> 2018-05-11 03:38:04,548-03 ERROR
>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand]
>> (EE-ManagedThreadFactory-engineScheduled-Thread-66)
>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command
>> 'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand' with
failure.
>> 2018-05-11 03:38:04,557-03 INFO
>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand]
>> (EE-ManagedThreadFactory-engineScheduled-Thread-66)
>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Lock freed to object
>>
'EngineLock:{exclusiveLocks='[4808bb70-c9cc-4286-aa39-16b5798213ac=LIVE_STORAGE_MIGRATION]',
>> sharedLocks=''}'
>>
>> I do not see the merge attempt in the vdsm.log, so please send vdsm logs
>> for node02.phy.eze.ampgn.com.ar from that time.
>>
>> Also, did you use the auto-generated snapshot to start the vm?
>>
>>
>> On Fri, May 11, 2018 at 6:11 PM, Juan Pablo <pablo.localhost(a)gmail.com>
>> wrote:
>>
>>> after the xfs_repair, it says: sorry I could not find valid secondary
>>> superblock
>>>
>>> 2018-05-11 12:09 GMT-03:00 Juan Pablo <pablo.localhost(a)gmail.com>:
>>>
>>>> hi,
>>>> Alias:
>>>> mail02-int_Disk1
>>>> Description:
>>>> ID:
>>>> 65ec515e-0aae-4fe6-a561-387929c7fb4d
>>>> Alignment:
>>>> Unknown
>>>> Disk Profile:
>>>> Wipe After Delete:
>>>> No
>>>>
>>>> that one
>>>>
>>>> 2018-05-11 11:12 GMT-03:00 Benny Zlotnik <bzlotnik(a)redhat.com>:
>>>>
>>>>> I looked at the logs and I see some disks have moved successfully
and
>>>>> some failed. Which disk is causing the problems?
>>>>>
>>>>> On Fri, May 11, 2018 at 5:02 PM, Juan Pablo <
>>>>> pablo.localhost(a)gmail.com> wrote:
>>>>>
>>>>>> Hi, just sent you via drive the files. attaching some extra
info,
>>>>>> thanks thanks and thanks :
>>>>>>
>>>>>> from inside the migrated vm I had the following attached dmesg
>>>>>> output before rebooting
>>>>>>
>>>>>> regards and thanks again for the help,
>>>>>>
>>>>>> 2018-05-11 10:45 GMT-03:00 Benny Zlotnik
<bzlotnik(a)redhat.com>:
>>>>>>
>>>>>>> Dropbox or google drive I guess. Also, can you attach
engine.log?
>>>>>>>
>>>>>>> On Fri, May 11, 2018 at 4:43 PM, Juan Pablo <
>>>>>>> pablo.localhost(a)gmail.com> wrote:
>>>>>>>
>>>>>>>>
>>>>>>>> vdsm is too big for gmail ...any other way I can share it
with you?
>>>>>>>>
>>>>>>>>
>>>>>>>> ---------- Forwrded message ----------
>>>>>>>> From: Juan Pablo <pablo.localhost(a)gmail.com>
>>>>>>>> Date: 2018-05-11 10:40 GMT-03:00
>>>>>>>> Subject: Re: [ovirt-users] strange issue: vm lost info on
disk
>>>>>>>> To: Benny Zlotnik <bzlotnik(a)redhat.com>
>>>>>>>> Cc: users <Users(a)ovirt.org>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> Benny, thanks for your reply! it was a Live migration.
sorry, it
>>>>>>>> was from nfs to iscsi, not otherwise. I have reboot the
vm for rescue and
>>>>>>>> it does not detect any partitions with fdisk, Im running
a xfs_repair with
>>>>>>>> -n and found some corrupted primary superblock., its
still running... (
>>>>>>>> so... there's info in the disk maybe?)
>>>>>>>>
>>>>>>>> attaching logs, let me know if those are the ones.
>>>>>>>> thanks again!
>>>>>>>>
>>>>>>>> 2018-05-11 9:45 GMT-03:00 Benny Zlotnik
<bzlotnik(a)redhat.com>:
>>>>>>>>
>>>>>>>>> Can you provide the logs? engine and vdsm.
>>>>>>>>> Did you perform a live migration (the VM is running)
or cold?
>>>>>>>>>
>>>>>>>>> On Fri, May 11, 2018 at 2:49 PM, Juan Pablo <
>>>>>>>>> pablo.localhost(a)gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Hi! , Im strugled about an ongoing problem:
>>>>>>>>>> after migrating a vm's disk from an iscsi
domain to a nfs and
>>>>>>>>>> ovirt reporting the migration was successful, I
see there's no data
>>>>>>>>>> 'inside' the vm's disk. we never had
this issues with ovirt so Im stranged
>>>>>>>>>> about the root cause and if theres a chance of
recovering the information.
>>>>>>>>>>
>>>>>>>>>> can you please help me out troubleshooting this
one? I would
>>>>>>>>>> really appreciate it =)
>>>>>>>>>> running ovirt 4.2.1 here!
>>>>>>>>>>
>>>>>>>>>> thanks in advance,
>>>>>>>>>> JP
>>>>>>>>>>
>>>>>>>>>> _______________________________________________
>>>>>>>>>> Users mailing list -- users(a)ovirt.org
>>>>>>>>>> To unsubscribe send an email to
users-leave(a)ovirt.org
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>
_______________________________________________
Users mailing list -- users(a)ovirt.org
To unsubscribe send an email to users-leave(a)ovirt.org