Am 03.07.2018 um 22:25 schrieb Yuval Turgeman <yturgema@redhat.com>:Hi Oliver,I would try the following, but please notice it is *very* dangerous, so a backup is probably a good idea (man vgcfgrestore)... 1. vgcfgrestore --list onn_ovn-monster2. search for a .vg file that was created before deleting those 2 lvs (ovirt-node-ng-4.2.3-0.20180524.0 and ovirt-node-ng-4.2.3.1-0. 20180530.0) 4. lvremove onn_ovn-monster/ovirt-node-ng-4.2.4-0.20180626.0 5. lvremove onn_ovn-monster/ovirt-node-ng-4.2.4-0.20180626.0+1 6. lvremove the lvs from the thinpool that are not mounted/used (var_crash?)7. nodectl info to make sure everything is ok8. reinstall the image-update rpmThanks,Yuval.On Tue, Jul 3, 2018 at 10:57 PM, Yuval Turgeman <yturgema@redhat.com> wrote:Hi Oliver,The KeyError happens because there are no bases for the layers. For each LV that ends with a +1, there should be a base read-only LV without +1. So for 3 ovirt-node-ng images, you're supposed to have 6 layers. This is the reason nodectl info fails, and the upgrade will fail also. In your original email it looks OK - I have never seen this happen, was this a manual lvremove ? I need to reproduce this and check what can be done.You can find me on #ovirt (irc.oftc.net) also :)On Tue, Jul 3, 2018 at 10:41 PM, Oliver Riesener <Oliver.Riesener@hs-bremen.de> wrote:Yuval, here comes the lvs output.The IO Errors are because Node is in maintenance.The LV root is from previous installed centos 7.5.The i have installed node-ng 4.2.1 and got this MIX.The LV turbo is a SSD in it’s own VG named ovirt.I have removed LV ovirt-node-ng-4.2.1-0.20180223.0 and (+1) removed because nodectl info error:KeyError: <NVR ovirt-node-ng-4.2.1.1-0.20180223.0 Now i get the error @4.2.3:[root@ovn-monster ~]# nodectl infoTraceback (most recent call last):File "/usr/lib64/python2.7/runpy.py", line 162, in _run_module_as_main "__main__", fname, loader, pkg_name)File "/usr/lib64/python2.7/runpy.py", line 72, in _run_code exec code in run_globalsFile "/usr/lib/python2.7/site-packages/nodectl/__main__.py", line 42, in <module> CliApplication()File "/usr/lib/python2.7/site-packages/nodectl/__init__.py", line 200, in CliApplication return cmdmap.command(args)File "/usr/lib/python2.7/site-packages/nodectl/__init__.py", line 118, in command return self.commands[command](**kwargs) File "/usr/lib/python2.7/site-packages/nodectl/__init__.py", line 76, in info Info(self.imgbased, self.machine).write()File "/usr/lib/python2.7/site-packages/nodectl/info.py", line 45, in __init__ self._fetch_information()File "/usr/lib/python2.7/site-packages/nodectl/info.py", line 49, in _fetch_information self._get_layout()File "/usr/lib/python2.7/site-packages/nodectl/info.py", line 66, in _get_layout layout = LayoutParser(self.app.imgbase.layout()).parse() File "/usr/lib/python2.7/site-packages/imgbased/imgbase.py", line 155, in layout return self.naming.layout()File "/usr/lib/python2.7/site-packages/imgbased/naming.py", line 109, in layout tree = self.tree(lvs)File "/usr/lib/python2.7/site-packages/imgbased/naming.py", line 224, in tree bases[img.base.nvr].layers.append(img) KeyError: <NVR ovirt-node-ng-4.2.3-0.20180524.0 /> lvs -a[root@ovn-monster ~]# lvs -a/dev/mapper/36090a02860ee716bee5e05b11dc52616: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/mapper/36090a02860ee716bee5e05b11dc52616: read failed after 0 of 4096 at 5497568559104: Eingabe-/Ausgabefehler /dev/mapper/36090a02860ee716bee5e05b11dc52616: read failed after 0 of 4096 at 5497568616448: Eingabe-/Ausgabefehler /dev/mapper/36090a02860ee716bee5e05b11dc52616: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler /dev/mapper/36090a02860ee9137c5ae35cd4bc5f6b8: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/mapper/36090a02860ee9137c5ae35cd4bc5f6b8: read failed after 0 of 4096 at 1099526242304: Eingabe-/Ausgabefehler /dev/mapper/36090a02860ee9137c5ae35cd4bc5f6b8: read failed after 0 of 4096 at 1099526299648: Eingabe-/Ausgabefehler /dev/mapper/36090a02860ee9137c5ae35cd4bc5f6b8: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler /dev/mapper/36090a02860eea13dc5aed55e4cc57698: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/mapper/36090a02860eea13dc5aed55e4cc57698: read failed after 0 of 4096 at 1099526242304: Eingabe-/Ausgabefehler /dev/mapper/36090a02860eea13dc5aed55e4cc57698: read failed after 0 of 4096 at 1099526299648: Eingabe-/Ausgabefehler /dev/mapper/36090a02860eea13dc5aed55e4cc57698: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/metadata: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/metadata: read failed after 0 of 4096 at 536805376: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/metadata: read failed after 0 of 4096 at 536862720: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/metadata: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/ids: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/ids: read failed after 0 of 4096 at 134152192: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/ids: read failed after 0 of 4096 at 134209536: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/ids: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/leases: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/leases: read failed after 0 of 4096 at 2147418112: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/leases: read failed after 0 of 4096 at 2147475456: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/leases: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/outbox: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/outbox: read failed after 0 of 4096 at 134152192: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/outbox: read failed after 0 of 4096 at 134209536: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/outbox: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/xleases: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/xleases: read failed after 0 of 4096 at 1073676288: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/xleases: read failed after 0 of 4096 at 1073733632: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/xleases: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/inbox: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/inbox: read failed after 0 of 4096 at 134152192: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/inbox: read failed after 0 of 4096 at 134209536: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/inbox: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/master: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/master: read failed after 0 of 4096 at 1073676288: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/master: read failed after 0 of 4096 at 1073733632: Eingabe-/Ausgabefehler /dev/675cb45d-3746-4f3b-b9ee-516612da50e5/master: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/metadata: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/metadata: read failed after 0 of 4096 at 536805376: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/metadata: read failed after 0 of 4096 at 536862720: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/metadata: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/metadata: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/metadata: read failed after 0 of 4096 at 536805376: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/metadata: read failed after 0 of 4096 at 536862720: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/metadata: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/ids: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/ids: read failed after 0 of 4096 at 134152192: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/ids: read failed after 0 of 4096 at 134209536: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/ids: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/ids: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/ids: read failed after 0 of 4096 at 134152192: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/ids: read failed after 0 of 4096 at 134209536: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/ids: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/leases: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/leases: read failed after 0 of 4096 at 2147418112: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/leases: read failed after 0 of 4096 at 2147475456: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/leases: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/leases: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/leases: read failed after 0 of 4096 at 2147418112: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/leases: read failed after 0 of 4096 at 2147475456: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/leases: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/outbox: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/outbox: read failed after 0 of 4096 at 134152192: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/outbox: read failed after 0 of 4096 at 134209536: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/outbox: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/xleases: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/xleases: read failed after 0 of 4096 at 1073676288: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/xleases: read failed after 0 of 4096 at 1073733632: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/xleases: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/inbox: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/inbox: read failed after 0 of 4096 at 134152192: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/inbox: read failed after 0 of 4096 at 134209536: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/inbox: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/master: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/master: read failed after 0 of 4096 at 1073676288: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/master: read failed after 0 of 4096 at 1073733632: Eingabe-/Ausgabefehler /dev/bcdbb66e-6196-4366-be25-a3e9ab948839/master: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/outbox: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/outbox: read failed after 0 of 4096 at 134152192: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/outbox: read failed after 0 of 4096 at 134209536: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/outbox: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/inbox: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/inbox: read failed after 0 of 4096 at 134152192: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/inbox: read failed after 0 of 4096 at 134209536: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/inbox: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/master: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/master: read failed after 0 of 4096 at 1073676288: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/master: read failed after 0 of 4096 at 1073733632: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/master: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/xleases: read failed after 0 of 4096 at 0: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/xleases: read failed after 0 of 4096 at 1073676288: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/xleases: read failed after 0 of 4096 at 1073733632: Eingabe-/Ausgabefehler /dev/c91974bf-fd64-4630-8005-e785b73acbef/xleases: read failed after 0 of 4096 at 4096: Eingabe-/Ausgabefehler LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Converthome onn_ovn-monster Vwi-aotz-- 1,00g pool00 4,79[lvol0_pmspare] onn_ovn-monster ewi------- 144,00movirt-node-ng-4.2.3-0.20180524.0+1 onn_ovn-monster Vwi-aotz-- <252,38g pool00 2,88 ovirt-node-ng-4.2.3.1-0.20180530.0+1 onn_ovn-monster Vwi-a-tz-- <252,38g pool00 0,86 ovirt-node-ng-4.2.4-0.20180626.0 onn_ovn-monster Vri-a-tz-k <252,38g pool00 0,85 ovirt-node-ng-4.2.4-0.20180626.0+1 onn_ovn-monster Vwi-a-tz-- <252,38g pool00 ovirt-node-ng-4.2.4-0.20180626 .0 0,85 pool00 onn_ovn-monster twi-aotz-- <279,38g 6,76 1,01[pool00_tdata] onn_ovn-monster Twi-ao---- <279,38g[pool00_tmeta] onn_ovn-monster ewi-ao---- 1,00groot onn_ovn-monster Vwi-a-tz-- <252,38g pool00 1,24swap onn_ovn-monster -wi-ao---- 4,00gtmp onn_ovn-monster Vwi-aotz-- 1,00g pool00 5,01var onn_ovn-monster Vwi-aotz-- 15,00g pool00 3,56var_crash onn_ovn-monster Vwi-aotz-- 10,00g pool00 2,86var_log onn_ovn-monster Vwi-aotz-- 8,00g pool00 38,48var_log_audit onn_ovn-monster Vwi-aotz-- 2,00g pool00 6,77turbo ovirt -wi-ao---- 894,25gAm 03.07.2018 um 12:58 schrieb Yuval Turgeman <yturgema@redhat.com>:Oliver, can you share the output from lvs ?______________________________On Tue, Jul 3, 2018 at 12:06 AM, Oliver Riesener <Oliver.Riesener@hs-bremen.de> wrote:Hi Yuval,* reinstallation failed, because LV already exists.ovirt-node-ng-4.2.4-0.20180626.0 onn_ovn-monster Vri-a-tz-k <252,38g pool00 0,85 ovirt-node-ng-4.2.4-0.20180626.0+1 onn_ovn-monster Vwi-a-tz-- <252,38g pool00 ovirt-node-ng-4.2.4-0.20180626 .0 0,85 See attachment imgbased.reinstall.log* I removed them and re-reinstall without luck.I got KeyError: <NVR ovirt-node-ng-4.2.1.1-0.20180223.0 /> See attachment imgbased.rereinstall.logAlso a new problem with nodectl info[root@ovn-monster tmp]# nodectl infoTraceback (most recent call last):File "/usr/lib64/python2.7/runpy.py", line 162, in _run_module_as_main "__main__", fname, loader, pkg_name)File "/usr/lib64/python2.7/runpy.py", line 72, in _run_code exec code in run_globalsFile "/usr/lib/python2.7/site-packages/nodectl/__main__.py", line 42, in <module> CliApplication()File "/usr/lib/python2.7/site-packages/nodectl/__init__.py", line 200, in CliApplication return cmdmap.command(args)File "/usr/lib/python2.7/site-packages/nodectl/__init__.py", line 118, in command return self.commands[command](**kwargs) File "/usr/lib/python2.7/site-packages/nodectl/__init__.py", line 76, in info Info(self.imgbased, self.machine).write()File "/usr/lib/python2.7/site-packages/nodectl/info.py", line 45, in __init__ self._fetch_information()File "/usr/lib/python2.7/site-packages/nodectl/info.py", line 49, in _fetch_information self._get_layout()File "/usr/lib/python2.7/site-packages/nodectl/info.py", line 66, in _get_layout layout = LayoutParser(self.app.imgbase.layout()).parse() File "/usr/lib/python2.7/site-packages/imgbased/imgbase.py", line 155, in layout return self.naming.layout()File "/usr/lib/python2.7/site-packages/imgbased/naming.py", line 109, in layout tree = self.tree(lvs)File "/usr/lib/python2.7/site-packages/imgbased/naming.py", line 224, in tree bases[img.base.nvr].layers.append(img) KeyError: <NVR ovirt-node-ng-4.2.3-0.20180524.0 /> Am 02.07.2018 um 22:22 schrieb Oliver Riesener <Oliver.Riesener@hs-bremen.de>: ______________________________Hi Yuval,yes you are right, there was a unused and deactivated var_crash LV.* I activated and mount it to /var/crash via /etc/fstab.* /var/crash was empty, and LV has already ext4 fs.var_crash onn_ovn-monster Vwi-aotz-- 10,00g pool00 2,86* Now i will try to upgrade again.* yum reinstall ovirt-node-ng-image-update.noarch BTW, no more imgbased.log files found.Am 02.07.2018 um 20:57 schrieb Yuval Turgeman <yturgema@redhat.com>:List Archives: https://lists.ovirt.From your log:AssertionError: Path is already a volume: /var/crashBasically, it means that you already have an LV for /var/crash but it's not mounted for some reason, so either mount it (if the data good) or remove it and then reinstall the image-update rpm. Before that, check that you dont have any other LVs in that same state - or you can post the output for lvs... btw, do you have any more imgbased.log files laying around ?You can find more details about this here:______________________________On Mon, Jul 2, 2018 at 8:12 PM, Oliver Riesener <Oliver.Riesener@hs-bremen.de > wrote:Hi,i attached my /tmp/imgbased.logSheersOliverAm 02.07.2018 um 13:58 schrieb Yuval Turgeman <yuvalt@redhat.com>:Looks like the upgrade script failed - can you please attach /var/log/imgbased.log or /tmp/imgbased.log ?Thanks,Yuval.______________________________On Mon, Jul 2, 2018 at 2:54 PM, Sandro Bonazzola <sbonazzo@redhat.com> wrote:Yuval, can you please have a look?List Archives: https://lists.ovirt.Yes, here is the same.It seams the bootloader isn’t configured right ?I did the Upgrade and reboot to 4.2.4 from UI and got:[root@ovn-monster ~]# nodectl infolayers:ovirt-node-ng-4.2.4-0.20180626.0: ovirt-node-ng-4.2.4-0.20180626.0+1 ovirt-node-ng-4.2.3.1-0.20180530.0: ovirt-node-ng-4.2.3.1-0.20180530.0+1 ovirt-node-ng-4.2.3-0.20180524.0: ovirt-node-ng-4.2.3-0.20180524.0+1 ovirt-node-ng-4.2.1.1-0.20180223.0: ovirt-node-ng-4.2.1.1-0.20180223.0+1 bootloader:default: ovirt-node-ng-4.2.3-0.20180524.0+1 entries:ovirt-node-ng-4.2.3-0.20180524.0+1: index: 0title: ovirt-node-ng-4.2.3-0.20180524.0 kernel: /boot/ovirt-node-ng-4.2.3-0.20180524.0+1/vmlinuz-3.10.0-862. 3.2.el7.x86_64 args: "ro crashkernel=auto rd.lvm.lv=onn_ovn-monster/ovirt-node-ng-4.2 .3-0.20180524.0+1 rd.lvm.lv=on n_ovn-monster/swap rd.md.uuid=c6c3013b:027a9346:6 7dfd181:89635587 rhgb quiet LANG=de_DE.UTF-8 img.bootid=ovirt-node-ng-4.2.3 -0.20180524.0+1" initrd: /boot/ovirt-node-ng-4.2.3-0.20180524.0+1/initramfs-3.10.0-86 2.3.2.el7.x86_64.img root: /dev/onn_ovn-monster/ovirt-node-ng-4.2.3-0.20180524.0+1 ovirt-node-ng-4.2.1.1-0.20180223.0+1: index: 1title: ovirt-node-ng-4.2.1.1-0.20180223.0 kernel: /boot/ovirt-node-ng-4.2.1.1-0.20180223.0+1/vmlinuz-3.10.0-69 3.17.1.el7.x86_64 args: "ro crashkernel=auto rd.lvm.lv=onn_ovn-monster/ovirt-node-ng-4.2 .1.1-0.20180223.0+1 rd.lvm.lv= onn_ovn-monster/swap rd.md.uuid=c6c3013b:027a9346:6 7dfd181:89635587 rhgb quiet LANG=de_DE.UTF-8 img.bootid=ovirt-node-ng-4.2.1 .1-0.20180223.0+1" initrd: /boot/ovirt-node-ng-4.2.1.1-0.20180223.0+1/initramfs-3.10.0- 693.17.1.el7.x86_64.img root: /dev/onn_ovn-monster/ovirt-node-ng-4.2.1.1-0.20180223.0+1 current_layer: ovirt-node-ng-4.2.3-0.20180524.0+1 [root@ovn-monster ~]# uptime07:35:27 up 2 days, 15:42, 1 user, load average: 1,07, 1,00, 0,95Am 29.06.2018 um 23:53 schrieb Matt Simonsen <matt@khoza.com>:Hello,
I did yum updates on 2 of my oVirt 4.2.3 nodes running the prebuilt node platform and it doesn't appear the updates worked.
[root@node6-g8-h4 ~]# yum update
Loaded plugins: enabled_repos_upload, fastestmirror, imgbased-persist,
: package_upload, product-id, search-disabled-repos, subscription-
: manager
This system is not registered with an entitlement server. You can use subscription-manager to register.
Loading mirror speeds from cached hostfile
* ovirt-4.2-epel: linux.mirrors.es.net
Resolving Dependencies
--> Running transaction check
---> Package ovirt-node-ng-image-update.noarch 0:4.2.3.1-1.el7 will be updated
---> Package ovirt-node-ng-image-update.noarch 0:4.2.4-1.el7 will be obsoleting
---> Package ovirt-node-ng-image-update-placeholder.noarch 0:4.2.3.1-1.el7 will be obsoleted
--> Finished Dependency Resolution
Dependencies Resolved
============================================================ ============================== ============================== =
PackageArch Version Repository Size
============================================================ ============================== ============================== =
Installing:
ovirt-node-ng-image-updatenoarch 4.2.4-1.el7 ovirt-4.2 647 M
replacing ovirt-node-ng-image-update-placeholder.noarch 4.2.3.1-1.el7
Transaction Summary
============================================================ ============================== ============================== =
Install 1 Package
Total download size: 647 M
Is this ok [y/d/N]: y
Downloading packages:
warning: /var/cache/yum/x86_64/7/ovirt-4.2/packages/ovirt-node-ng-ima ge-update-4.2.4-1.el7.noarch.r pm: Header V4 RSA/SHA1 Signature, key ID fe590cb7: NOKEY
Public key for ovirt-node-ng-image-update-4.2.4-1.el7.noarch.rpm is not installed
ovirt-node-ng-image-update-4.2.4-1.el7.noarch.rpm | 647 MB 00:02:07
Retrieving key from file:///etc/pki/rpm-gpg/RPM-GPG-ovirt-4.2
Importing GPG key 0xFE590CB7:
Userid : "oVirt <infra@ovirt.org>"
Fingerprint: 31a5 d783 7fad 7cb2 86cd 3469 ab8c 4f9d fe59 0cb7
Package : ovirt-release42-4.2.3.1-1.el7.noarch (installed)
From : /etc/pki/rpm-gpg/RPM-GPG-ovirt-4.2
Is this ok [y/N]: y
Running transaction check
Running transaction test
Transaction test succeeded
Running transaction
Installing : ovirt-node-ng-image-update-4.2.4-1.el7.noarch 1/3
warning: %post(ovirt-node-ng-image-update-4.2.4-1.el7.noarch) scriptlet failed, exit status 1
Non-fatal POSTIN scriptlet failure in rpm package ovirt-node-ng-image-update-4.2.4-1.el7.noarch
Erasing : ovirt-node-ng-image-update-placeholder-4.2.3.1-1.el7.noarch 2/3
Cleanup : ovirt-node-ng-image-update-4.2.3.1-1.el7.noarch 3/3
warning: file /usr/share/ovirt-node-ng/image/ovirt-node-ng-4.2.0-0.2018053 0.0.el7.squashfs.img: remove failed: No such file or directory
Uploading Package Profile
Unable to upload Package Profile
Verifying : ovirt-node-ng-image-update-4.2.4-1.el7.noarch 1/3
Verifying : ovirt-node-ng-image-update-4.2.3.1-1.el7.noarch 2/3
Verifying : ovirt-node-ng-image-update-placeholder-4.2.3.1-1.el7.noarch 3/3
Installed:
ovirt-node-ng-image-update.noarch 0:4.2.4-1.el7
Replaced:
ovirt-node-ng-image-update-placeholder.noarch 0:4.2.3.1-1.el7
Complete!
Uploading Enabled Repositories Report
Loaded plugins: fastestmirror, product-id, subscription-manager
This system is not registered with an entitlement server. You can use subscription-manager to register.
Cannot upload enabled repos report, is this client registered?
My engine shows the nodes as having no updates, however the major components including the kernel version and port 9090 admin GUI show 4.2.3
Is there anything I can provide to help diagnose the issue?
[root@node6-g8-h4 ~]# rpm -qa | grep ovirt
ovirt-imageio-common-1.3.1.2-0.el7.centos.noarch
ovirt-host-deploy-1.7.3-1.el7.centos.noarch
ovirt-vmconsole-host-1.0.5-4.el7.centos.noarch
ovirt-provider-ovn-driver-1.2.10-1.el7.centos.noarch
ovirt-engine-sdk-python-3.6.9.1-1.el7.noarch
ovirt-setup-lib-1.1.4-1.el7.centos.noarch
ovirt-release42-4.2.3.1-1.el7.noarch
ovirt-imageio-daemon-1.3.1.2-0.el7.centos.noarch
ovirt-hosted-engine-setup-2.2.20-1.el7.centos.noarch
ovirt-host-dependencies-4.2.2-2.el7.centos.x86_64
ovirt-hosted-engine-ha-2.2.11-1.el7.centos.noarch
ovirt-host-4.2.2-2.el7.centos.x86_64
ovirt-node-ng-image-update-4.2.4-1.el7.noarch
ovirt-vmconsole-1.0.5-4.el7.centos.noarch
ovirt-release-host-node-4.2.3.1-1.el7.noarch
cockpit-ovirt-dashboard-0.11.24-1.el7.centos.noarch
ovirt-node-ng-nodectl-4.2.0-0.20180524.0.el7.noarch
python-ovirt-engine-sdk4-4.2.6-2.el7.centos.x86_64
[root@node6-g8-h4 ~]# yum update
Loaded plugins: enabled_repos_upload, fastestmirror, imgbased-persist, package_upload, product-id, search-disabled-repos, subscription-manager
This system is not registered with an entitlement server. You can use subscription-manager to register.
Loading mirror speeds from cached hostfile
* ovirt-4.2-epel: linux.mirrors.es.net
No packages marked for update
Uploading Enabled Repositories Report
Loaded plugins: fastestmirror, product-id, subscription-manager
This system is not registered with an entitlement server. You can use subscription-manager to register.
Cannot upload enabled repos report, is this client registered?
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-gui delines/
List Archives: https://lists.ovirt.org/archives/list/users@ovirt. org/message/UHQMGULUHL4GBBHUBN GOAICJEM6W3RVW/
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-gui delines/ org/archives/list/users@ovirt. org/message/RTOXFNAXQ3NJBWX7RX OYK5H5RZBHX2OK/ --_________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-gui delines/
List Archives: https://lists.ovirt.org/archives/list/users@ovirt. org/message/ACKMTWOUNXB7NWOCRZ IOWJI7U5HIQAUV/ _________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ org/archives/list/users@ovirt. org/message/E5MFPG7WW5BIQ5BOC7 LG5C23YZ7B7UJ2/ _________________ List Archives: https://lists.ovirt.org/archiv
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ es/list/users@ovirt.org/messag e/IPKGIHH7SFTW7K7YOKL4QXBYXR5K MKNB/ _________________ List Archives: https://lists.ovirt.org/archiv
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ es/list/users@ovirt.org/messag e/6HKLX746IND2NQVRRGKACMSQCA3G K6GA/