New oVirt Node install on oVirt Cluster 4.0.5 - How can I install oVirt Node with same 4.0.5 version ???

Hi oVirt Troopers, I have two segregated oVirt Clusters running 4.0.5 ( DEV and PROD enviroments). Now i need to install a new oVirt Node Server (Dell PowerEdge M620), but i see that does not exist an 4.0.5 iso on http://resources.ovirt.org/pub/ovirt-4.0/iso/ovirt-node-ng-installer/ , only 4.0.3 and 4.0.6. How can i install this new server and go to the same version of all 20 others ??? There is a way to install 4.0.3 and update to 4.0.5 only ? Thanks in advance.

I can think in two ways. Please let me know if have a change to go ok. First, download ovirt-node-ng-image-update and ovirt-node-ng-image from 4.0.5 version and run yum localinstall. Second, create a rpm list from other 4.0.5 node, do an diff agains my 4.0.3 node and use the diff to download packages from 4.0.5 version and run yum localintall. My concern with this are that i can not upgrade ovirt engine every time i need to install a new ovirt node. Thanks. Em sex, 5 de mai de 2017 às 11:26, Rogério Ceni Coelho < rogeriocenicoelho@gmail.com> escreveu:
Hi oVirt Troopers,
I have two segregated oVirt Clusters running 4.0.5 ( DEV and PROD enviroments).
Now i need to install a new oVirt Node Server (Dell PowerEdge M620), but i see that does not exist an 4.0.5 iso on http://resources.ovirt.org/pub/ovirt-4.0/iso/ovirt-node-ng-installer/ , only 4.0.3 and 4.0.6.
How can i install this new server and go to the same version of all 20 others ???
There is a way to install 4.0.3 and update to 4.0.5 only ?
Thanks in advance.

I take it updating everything to 4.0.6 is not an option ? Thanks, Yuval On May 5, 2017 6:16 PM, "Rogério Ceni Coelho" <rogeriocenicoelho@gmail.com> wrote:
I can think in two ways. Please let me know if have a change to go ok.
First, download ovirt-node-ng-image-update and ovirt-node-ng-image from 4.0.5 version and run yum localinstall.
Second, create a rpm list from other 4.0.5 node, do an diff agains my 4.0.3 node and use the diff to download packages from 4.0.5 version and run yum localintall.
My concern with this are that i can not upgrade ovirt engine every time i need to install a new ovirt node.
Thanks.
Em sex, 5 de mai de 2017 às 11:26, Rogério Ceni Coelho < rogeriocenicoelho@gmail.com> escreveu:
Hi oVirt Troopers,
I have two segregated oVirt Clusters running 4.0.5 ( DEV and PROD enviroments).
Now i need to install a new oVirt Node Server (Dell PowerEdge M620), but i see that does not exist an 4.0.5 iso on http://resources.ovirt.org/ pub/ovirt-4.0/iso/ovirt-node-ng-installer/ , only 4.0.3 and 4.0.6.
How can i install this new server and go to the same version of all 20 others ???
There is a way to install 4.0.3 and update to 4.0.5 only ?
Thanks in advance.
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users

Hi Yuval, Yes, it is, but every time I need to install a new oVirt node i will need to upgrade ovirt engine at least ? And i will need to upgrade ovirt nodes that already exist ? I have 20 ovirt nodes .... So, this means a lot of work. My enviroment are stable with 4.0.5 and i am happy for now. oVirt is an excellent product. Thanks for that. For example, this morning i try to put an export import storage domain in manteinance and an error occur only with my new ovirt node running 4.0.6.1 and yesterday a lost 2 days debbuging another problem with network start with many vlans on 4.0.6.1 ... :-( StorageDomainDoesNotExist: Storage domain does not exist: (u'f9e051a9-6660-4e49-a3f1-354583501610',) Thread-12::DEBUG::2017-05-05 10:39:07,473::check::296::storage.check::(_start_process) START check '/dev/c58ce4b0-7145-4cd0-900e-eeb99177a7de/metadata' cmd=['/usr/bin/taskset', '--cpu-list', '0-31', '/usr/bin/dd', 'if=/dev/c58ce4b0-7145-4cd0-900e-eeb99177a7de/metadata', 'of=/dev/null', 'bs=4096', 'count=1', 'iflag=direct'] delay=0.00 Thread-12::DEBUG::2017-05-05 10:39:07,524::asyncevent::564::storage.asyncevent::(reap) Process <cpopen.CPopen object at 0x401c790> terminated (count=1) Thread-12::DEBUG::2017-05-05 10:39:07,525::check::327::storage.check::(_check_completed) FINISH check '/dev/c58ce4b0-7145-4cd0-900e-eeb99177a7de/metadata' rc=0 err=bytearray(b'1+0 records in\n1+0 records out\n4096 bytes (4.1 kB) copied, 0.000292117 s, 14.0 MB/s\n') elapsed=0.06 Thread-12::DEBUG::2017-05-05 10:39:07,886::check::296::storage.check::(_start_process) START check u'/rhev/data-center/mnt/corot.rbs.com.br:_u00_oVirt_PRD_ISO__DOMAIN/7b8c9293-f103-401a-93ac-550981837224/dom_md/metadata' cmd=['/usr/bin/taskset', '--cpu-list', '0-31', '/usr/bin/dd', u'if=/rhev/data-center/mnt/corot.rbs.com.br:_u00_oVirt_PRD_ISO__DOMAIN/7b8c9293-f103-401a-93ac-550981837224/dom_md/metadata', 'of=/dev/null', 'bs=4096', 'count=1', 'iflag=direct'] delay=0.00 Thread-12::DEBUG::2017-05-05 10:39:07,898::check::296::storage.check::(_start_process) START check u'/rhev/data-center/mnt/dd6701.bkp.srvr.rbs.net:_data_col1_ovirt__prd/db89d5df-00ac-4e58-a7e5-e31272f9ea92/dom_md/metadata' cmd=['/usr/bin/taskset', '--cpu-list', '0-31', '/usr/bin/dd', u'if=/rhev/data-center/mnt/dd6701.bkp.srvr.rbs.net:_data_col1_ovirt__prd/db89d5df-00ac-4e58-a7e5-e31272f9ea92/dom_md/metadata', 'of=/dev/null', 'bs=4096', 'count=1', 'iflag=direct'] delay=0.00 Thread-12::DEBUG::2017-05-05 10:39:07,906::check::327::storage.check::(_check_completed) FINISH check u'/rhev/data-center/mnt/corot.rbs.com.br:_u00_oVirt_PRD_ISO__DOMAIN/7b8c9293-f103-401a-93ac-550981837224/dom_md/metadata' rc=0 err=bytearray(b'0+1 records in\n0+1 records out\n386 bytes (386 B) copied, 0.00038325 s, 1.0 MB/s\n') elapsed=0.02 Thread-12::DEBUG::2017-05-05 10:39:07,916::check::296::storage.check::(_start_process) START check u'/rhev/data-center/mnt/vnx01.srv.srvr.rbs.net:_fs__ovirt_prd__data__domain/fdcf130d-53b8-4978-8f97-82f364639b4a/dom_md/metadata' cmd=['/usr/bin/taskset', '--cpu-list', '0-31', '/usr/bin/dd', u'if=/rhev/data-center/mnt/vnx01.srv.srvr.rbs.net:_fs__ovirt_prd__data__domain/fdcf130d-53b8-4978-8f97-82f364639b4a/dom_md/metadata', 'of=/dev/null', 'bs=4096', 'count=1', 'iflag=direct'] delay=0.00 Thread-12::DEBUG::2017-05-05 10:39:07,930::check::327::storage.check::(_check_completed) FINISH check u'/rhev/data-center/mnt/dd6701.bkp.srvr.rbs.net:_data_col1_ovirt__prd/db89d5df-00ac-4e58-a7e5-e31272f9ea92/dom_md/metadata' rc=0 err=bytearray(b'0+1 records in\n0+1 records out\n360 bytes (360 B) copied, 0.000363751 s, 990 kB/s\n') elapsed=0.03 Thread-12::DEBUG::2017-05-05 10:39:07,964::asyncevent::564::storage.asyncevent::(reap) Process <cpopen.CPopen object at 0x3a48950> terminated (count=1) Thread-12::DEBUG::2017-05-05 10:39:07,964::check::327::storage.check::(_check_completed) FINISH check u'/rhev/data-center/mnt/vnx01.srv.srvr.rbs.net:_fs__ovirt_prd__data__domain/fdcf130d-53b8-4978-8f97-82f364639b4a/dom_md/metadata' rc=0 err=bytearray(b'0+1 records in\n0+1 records out\n369 bytes (369 B) copied, 0.000319659 s, 1.2 MB/s\n') elapsed=0.05 Thread-12::DEBUG::2017-05-05 10:39:09,035::check::296::storage.check::(_start_process) START check '/dev/1804d02e-7865-4acd-a04f-8200ac2d2b84/metadata' cmd=['/usr/bin/taskset', '--cpu-list', '0-31', '/usr/bin/dd', 'if=/dev/1804d02e-7865-4acd-a04f-8200ac2d2b84/metadata', 'of=/dev/null', 'bs=4096', 'count=1', 'iflag=direct'] delay=0.00 Thread-12::DEBUG::2017-05-05 10:39:09,084::asyncevent::564::storage.asyncevent::(reap) Process <cpopen.CPopen object at 0x383d690> terminated (count=1) Thread-12::DEBUG::2017-05-05 10:39:09,085::check::327::storage.check::(_check_completed) FINISH check '/dev/1804d02e-7865-4acd-a04f-8200ac2d2b84/metadata' rc=0 err=bytearray(b'1+0 records in\n1+0 records out\n4096 bytes (4.1 kB) copied, 0.000190859 s, 21.5 MB/s\n') elapsed=0.05 Thread-37::DEBUG::2017-05-05 10:39:12,431::monitor::365::Storage.Monitor::(_produceDomain) Producing domain f9e051a9-6660-4e49-a3f1-354583501610 Thread-37::ERROR::2017-05-05 10:39:12,431::sdc::140::Storage.StorageDomainCache::(_findDomain) looking for unfetched domain f9e051a9-6660-4e49-a3f1-354583501610 Thread-37::ERROR::2017-05-05 10:39:12,431::sdc::157::Storage.StorageDomainCache::(_findUnfetchedDomain) looking for domain f9e051a9-6660-4e49-a3f1-354583501610 Thread-37::DEBUG::2017-05-05 10:39:12,432::lvm::288::Storage.Misc.excCmd::(cmd) /usr/bin/taskset --cpu-list 0-31 /usr/bin/sudo -n /usr/sbin/lvm vgs --config ' devices { preferred_names = ["^/dev/mapper/"] ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3 filter = [ '\''a|/dev/mapper/36000144000000010206a1c589de93966|/dev/mapper/36000144000000010206a1c589de939b1|/dev/mapper/36000144000000010206a1c589de93a4d|/dev/mapper/3600601600f7025006828c2f47c7be611|'\'', '\''r|.*|'\'' ] } global { locking_type=1 prioritise_write_locks=1 wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days = 0 } ' --noheadings --units b --nosuffix --separator '|' --ignoreskippedcluster -o uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name f9e051a9-6660-4e49-a3f1-354583501610 (cwd None) Thread-37::DEBUG::2017-05-05 10:39:12,591::lvm::288::Storage.Misc.excCmd::(cmd) FAILED: <err> = ' WARNING: Not using lvmetad because config setting use_lvmetad=0.\n WARNING: To avoid corruption, rescan devices to make changes visible (pvscan --cache).\n Volume group "f9e051a9-6660-4e49-a3f1-354583501610" not found\n Cannot process volume group f9e051a9-6660-4e49-a3f1-354583501610\n'; <rc> = 5 Thread-37::WARNING::2017-05-05 10:39:12,597::lvm::376::Storage.LVM::(_reloadvgs) lvm vgs failed: 5 [] [' WARNING: Not using lvmetad because config setting use_lvmetad=0.', ' WARNING: To avoid corruption, rescan devices to make changes visible (pvscan --cache).', ' Volume group "f9e051a9-6660-4e49-a3f1-354583501610" not found', ' Cannot process volume group f9e051a9-6660-4e49-a3f1-354583501610'] *Thread-37::ERROR::2017-05-05 10:39:12,602::sdc::146::Storage.StorageDomainCache::(_findDomain) domain f9e051a9-6660-4e49-a3f1-354583501610 not found* Traceback (most recent call last): File "/usr/share/vdsm/storage/sdc.py", line 144, in _findDomain dom = findMethod(sdUUID) File "/usr/share/vdsm/storage/sdc.py", line 174, in _findUnfetchedDomain raise se.StorageDomainDoesNotExist(sdUUID) StorageDomainDoesNotExist: Storage domain does not exist: (u'f9e051a9-6660-4e49-a3f1-354583501610',) Thread-37::ERROR::2017-05-05 10:39:12,602::monitor::328::Storage.Monitor::(_setupLoop) Setting up monitor for f9e051a9-6660-4e49-a3f1-354583501610 failed Traceback (most recent call last): File "/usr/share/vdsm/storage/monitor.py", line 325, in _setupLoop self._setupMonitor() File "/usr/share/vdsm/storage/monitor.py", line 348, in _setupMonitor self._produceDomain() File "/usr/lib/python2.7/site-packages/vdsm/utils.py", line 405, in wrapper value = meth(self, *a, **kw) File "/usr/share/vdsm/storage/monitor.py", line 366, in _produceDomain self.domain = sdCache.produce(self.sdUUID) File "/usr/share/vdsm/storage/sdc.py", line 101, in produce domain.getRealDomain() File "/usr/share/vdsm/storage/sdc.py", line 53, in getRealDomain return self._cache._realProduce(self._sdUUID) File "/usr/share/vdsm/storage/sdc.py", line 125, in _realProduce domain = self._findDomain(sdUUID) File "/usr/share/vdsm/storage/sdc.py", line 144, in _findDomain dom = findMethod(sdUUID) File "/usr/share/vdsm/storage/sdc.py", line 174, in _findUnfetchedDomain raise se.StorageDomainDoesNotExist(sdUUID) StorageDomainDoesNotExist: Storage domain does not exist: (u'f9e051a9-6660-4e49-a3f1-354583501610',) jsonrpc.Executor/3::DEBUG::2017-05-05 10:39:13,057::__init__::530::jsonrpc.JsonRpcServer::(_handle_request) Calling 'Host.getAllVmStats' in bridge with {} jsonrpc.Executor/3::DEBUG::2017-05-05 10:39:13,081::__init__::555::jsonrpc.JsonRpcServer::(_handle_request) Return 'Host.getAllVmStats' in bridge with (suppressed) jsonrpc.Executor/3::INFO::2017-05-05 10:39:13,088::__init__::513::jsonrpc.JsonRpcServer::(_serveRequest) RPC call Host.getAllVmStats succeeded in 0.03 seconds Em sex, 5 de mai de 2017 às 14:40, Yuval Turgeman <yuvalt@redhat.com> escreveu:
I take it updating everything to 4.0.6 is not an option ?
Thanks, Yuval
On May 5, 2017 6:16 PM, "Rogério Ceni Coelho" <rogeriocenicoelho@gmail.com> wrote:
I can think in two ways. Please let me know if have a change to go ok.
First, download ovirt-node-ng-image-update and ovirt-node-ng-image from 4.0.5 version and run yum localinstall.
Second, create a rpm list from other 4.0.5 node, do an diff agains my 4.0.3 node and use the diff to download packages from 4.0.5 version and run yum localintall.
My concern with this are that i can not upgrade ovirt engine every time i need to install a new ovirt node.
Thanks.
Em sex, 5 de mai de 2017 às 11:26, Rogério Ceni Coelho < rogeriocenicoelho@gmail.com> escreveu:
Hi oVirt Troopers,
I have two segregated oVirt Clusters running 4.0.5 ( DEV and PROD enviroments).
Now i need to install a new oVirt Node Server (Dell PowerEdge M620), but i see that does not exist an 4.0.5 iso on http://resources.ovirt.org/pub/ovirt-4.0/iso/ovirt-node-ng-installer/ , only 4.0.3 and 4.0.6.
How can i install this new server and go to the same version of all 20 others ???
There is a way to install 4.0.3 and update to 4.0.5 only ?
Thanks in advance.
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users

On Fri, May 5, 2017 at 9:03 PM, Rogério Ceni Coelho < rogeriocenicoelho@gmail.com> wrote:
Hi Yuval,
Yes, it is, but every time I need to install a new oVirt node i will need to upgrade ovirt engine at least ?
Of course not. You can upgrade the Engine independently from the hosts.
And i will need to upgrade ovirt nodes that already exist ? I have 20 ovirt nodes .... So, this means a lot of work.
No, you can upgrade each host independently.
My enviroment are stable with 4.0.5 and i am happy for now. oVirt is an excellent product. Thanks for that.
For example, this morning i try to put an export import storage domain in manteinance and an error occur only with my new ovirt node running 4.0.6.1 and yesterday a lost 2 days debbuging another problem with network start with many vlans on 4.0.6.1 ... :-(
It'd be great if you can start a different thread about it, or file a bug, with all the details (VDSM and Engine logs attached). Y.
StorageDomainDoesNotExist: Storage domain does not exist: (u'f9e051a9-6660-4e49-a3f1-354583501610',) Thread-12::DEBUG::2017-05-05 10:39:07,473::check::296:: storage.check::(_start_process) START check '/dev/c58ce4b0-7145-4cd0-900e-eeb99177a7de/metadata' cmd=['/usr/bin/taskset', '--cpu-list', '0-31', '/usr/bin/dd', 'if=/dev/c58ce4b0-7145-4cd0-900e-eeb99177a7de/metadata', 'of=/dev/null', 'bs=4096', 'count=1', 'iflag=direct'] delay=0.00 Thread-12::DEBUG::2017-05-05 10:39:07,524::asyncevent::564::storage.asyncevent::(reap) Process <cpopen.CPopen object at 0x401c790> terminated (count=1) Thread-12::DEBUG::2017-05-05 10:39:07,525::check::327:: storage.check::(_check_completed) FINISH check '/dev/c58ce4b0-7145-4cd0-900e-eeb99177a7de/metadata' rc=0 err=bytearray(b'1+0 records in\n1+0 records out\n4096 bytes (4.1 kB) copied, 0.000292117 s, 14.0 MB/s\n') elapsed=0.06 Thread-12::DEBUG::2017-05-05 10:39:07,886::check::296:: storage.check::(_start_process) START check u'/rhev/data-center/mnt/corot. rbs.com.br:_u00_oVirt_PRD_ISO__DOMAIN/7b8c9293-f103-401a- 93ac-550981837224/dom_md/metadata' cmd=['/usr/bin/taskset', '--cpu-list', '0-31', '/usr/bin/dd', u'if=/rhev/data-center/mnt/corot.rbs.com.br: _u00_oVirt_PRD_ISO__DOMAIN/7b8c9293-f103-401a-93ac-550981837224/dom_md/metadata', 'of=/dev/null', 'bs=4096', 'count=1', 'iflag=direct'] delay=0.00 Thread-12::DEBUG::2017-05-05 10:39:07,898::check::296:: storage.check::(_start_process) START check u'/rhev/data-center/mnt/ dd6701.bkp.srvr.rbs.net:_data_col1_ovirt__prd/db89d5df-00ac- 4e58-a7e5-e31272f9ea92/dom_md/metadata' cmd=['/usr/bin/taskset', '--cpu-list', '0-31', '/usr/bin/dd', u'if=/rhev/data-center/mnt/ dd6701.bkp.srvr.rbs.net:_data_col1_ovirt__prd/db89d5df-00ac- 4e58-a7e5-e31272f9ea92/dom_md/metadata', 'of=/dev/null', 'bs=4096', 'count=1', 'iflag=direct'] delay=0.00 Thread-12::DEBUG::2017-05-05 10:39:07,906::check::327:: storage.check::(_check_completed) FINISH check u'/rhev/data-center/mnt/corot.rbs.com.br:_u00_oVirt_PRD_ISO_ _DOMAIN/7b8c9293-f103-401a-93ac-550981837224/dom_md/metadata' rc=0 err=bytearray(b'0+1 records in\n0+1 records out\n386 bytes (386 B) copied, 0.00038325 s, 1.0 MB/s\n') elapsed=0.02 Thread-12::DEBUG::2017-05-05 10:39:07,916::check::296:: storage.check::(_start_process) START check u'/rhev/data-center/mnt/vnx01. srv.srvr.rbs.net:_fs__ovirt_prd__data__domain/fdcf130d- 53b8-4978-8f97-82f364639b4a/dom_md/metadata' cmd=['/usr/bin/taskset', '--cpu-list', '0-31', '/usr/bin/dd', u'if=/rhev/data-center/mnt/ vnx01.srv.srvr.rbs.net:_fs__ovirt_prd__data__domain/ fdcf130d-53b8-4978-8f97-82f364639b4a/dom_md/metadata', 'of=/dev/null', 'bs=4096', 'count=1', 'iflag=direct'] delay=0.00 Thread-12::DEBUG::2017-05-05 10:39:07,930::check::327:: storage.check::(_check_completed) FINISH check u'/rhev/data-center/mnt/ dd6701.bkp.srvr.rbs.net:_data_col1_ovirt__prd/db89d5df-00ac- 4e58-a7e5-e31272f9ea92/dom_md/metadata' rc=0 err=bytearray(b'0+1 records in\n0+1 records out\n360 bytes (360 B) copied, 0.000363751 s, 990 kB/s\n') elapsed=0.03 Thread-12::DEBUG::2017-05-05 10:39:07,964::asyncevent::564::storage.asyncevent::(reap) Process <cpopen.CPopen object at 0x3a48950> terminated (count=1) Thread-12::DEBUG::2017-05-05 10:39:07,964::check::327:: storage.check::(_check_completed) FINISH check u'/rhev/data-center/mnt/vnx01.srv.srvr.rbs.net:_fs__ovirt_ prd__data__domain/fdcf130d-53b8-4978-8f97-82f364639b4a/dom_md/metadata' rc=0 err=bytearray(b'0+1 records in\n0+1 records out\n369 bytes (369 B) copied, 0.000319659 s, 1.2 MB/s\n') elapsed=0.05 Thread-12::DEBUG::2017-05-05 10:39:09,035::check::296:: storage.check::(_start_process) START check '/dev/1804d02e-7865-4acd-a04f-8200ac2d2b84/metadata' cmd=['/usr/bin/taskset', '--cpu-list', '0-31', '/usr/bin/dd', 'if=/dev/1804d02e-7865-4acd-a04f-8200ac2d2b84/metadata', 'of=/dev/null', 'bs=4096', 'count=1', 'iflag=direct'] delay=0.00 Thread-12::DEBUG::2017-05-05 10:39:09,084::asyncevent::564::storage.asyncevent::(reap) Process <cpopen.CPopen object at 0x383d690> terminated (count=1) Thread-12::DEBUG::2017-05-05 10:39:09,085::check::327:: storage.check::(_check_completed) FINISH check '/dev/1804d02e-7865-4acd-a04f-8200ac2d2b84/metadata' rc=0 err=bytearray(b'1+0 records in\n1+0 records out\n4096 bytes (4.1 kB) copied, 0.000190859 s, 21.5 MB/s\n') elapsed=0.05 Thread-37::DEBUG::2017-05-05 10:39:12,431::monitor::365:: Storage.Monitor::(_produceDomain) Producing domain f9e051a9-6660-4e49-a3f1-354583501610 Thread-37::ERROR::2017-05-05 10:39:12,431::sdc::140:: Storage.StorageDomainCache::(_findDomain) looking for unfetched domain f9e051a9-6660-4e49-a3f1-354583501610 Thread-37::ERROR::2017-05-05 10:39:12,431::sdc::157:: Storage.StorageDomainCache::(_findUnfetchedDomain) looking for domain f9e051a9-6660-4e49-a3f1-354583501610 Thread-37::DEBUG::2017-05-05 10:39:12,432::lvm::288::Storage.Misc.excCmd::(cmd) /usr/bin/taskset --cpu-list 0-31 /usr/bin/sudo -n /usr/sbin/lvm vgs --config ' devices { preferred_names = ["^/dev/mapper/"] ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3 filter = [ '\''a|/dev/mapper/36000144000000010206a1c589de93 966|/dev/mapper/36000144000000010206a1c589de939b1|/dev/mapper/ 36000144000000010206a1c589de93a4d|/dev/mapper/ 3600601600f7025006828c2f47c7be611|'\'', '\''r|.*|'\'' ] } global { locking_type=1 prioritise_write_locks=1 wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days = 0 } ' --noheadings --units b --nosuffix --separator '|' --ignoreskippedcluster -o uuid,name,attr,size,free,extent_size,extent_count,free_ count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name f9e051a9-6660-4e49-a3f1-354583501610 (cwd None) Thread-37::DEBUG::2017-05-05 10:39:12,591::lvm::288::Storage.Misc.excCmd::(cmd) FAILED: <err> = ' WARNING: Not using lvmetad because config setting use_lvmetad=0.\n WARNING: To avoid corruption, rescan devices to make changes visible (pvscan --cache).\n Volume group "f9e051a9-6660-4e49-a3f1-354583501610" not found\n Cannot process volume group f9e051a9-6660-4e49-a3f1-354583501610\n'; <rc> = 5 Thread-37::WARNING::2017-05-05 10:39:12,597::lvm::376::Storage.LVM::(_reloadvgs) lvm vgs failed: 5 [] [' WARNING: Not using lvmetad because config setting use_lvmetad=0.', ' WARNING: To avoid corruption, rescan devices to make changes visible (pvscan --cache).', ' Volume group "f9e051a9-6660-4e49-a3f1-354583501610" not found', ' Cannot process volume group f9e051a9-6660-4e49-a3f1-354583501610'] *Thread-37::ERROR::2017-05-05 10:39:12,602::sdc::146::Storage.StorageDomainCache::(_findDomain) domain f9e051a9-6660-4e49-a3f1-354583501610 not found* Traceback (most recent call last): File "/usr/share/vdsm/storage/sdc.py", line 144, in _findDomain dom = findMethod(sdUUID) File "/usr/share/vdsm/storage/sdc.py", line 174, in _findUnfetchedDomain raise se.StorageDomainDoesNotExist(sdUUID) StorageDomainDoesNotExist: Storage domain does not exist: (u'f9e051a9-6660-4e49-a3f1-354583501610',) Thread-37::ERROR::2017-05-05 10:39:12,602::monitor::328::Storage.Monitor::(_setupLoop) Setting up monitor for f9e051a9-6660-4e49-a3f1-354583501610 failed Traceback (most recent call last): File "/usr/share/vdsm/storage/monitor.py", line 325, in _setupLoop self._setupMonitor() File "/usr/share/vdsm/storage/monitor.py", line 348, in _setupMonitor self._produceDomain() File "/usr/lib/python2.7/site-packages/vdsm/utils.py", line 405, in wrapper value = meth(self, *a, **kw) File "/usr/share/vdsm/storage/monitor.py", line 366, in _produceDomain self.domain = sdCache.produce(self.sdUUID) File "/usr/share/vdsm/storage/sdc.py", line 101, in produce domain.getRealDomain() File "/usr/share/vdsm/storage/sdc.py", line 53, in getRealDomain return self._cache._realProduce(self._sdUUID) File "/usr/share/vdsm/storage/sdc.py", line 125, in _realProduce domain = self._findDomain(sdUUID) File "/usr/share/vdsm/storage/sdc.py", line 144, in _findDomain dom = findMethod(sdUUID) File "/usr/share/vdsm/storage/sdc.py", line 174, in _findUnfetchedDomain raise se.StorageDomainDoesNotExist(sdUUID) StorageDomainDoesNotExist: Storage domain does not exist: (u'f9e051a9-6660-4e49-a3f1-354583501610',) jsonrpc.Executor/3::DEBUG::2017-05-05 10:39:13,057::__init__::530:: jsonrpc.JsonRpcServer::(_handle_request) Calling 'Host.getAllVmStats' in bridge with {} jsonrpc.Executor/3::DEBUG::2017-05-05 10:39:13,081::__init__::555:: jsonrpc.JsonRpcServer::(_handle_request) Return 'Host.getAllVmStats' in bridge with (suppressed) jsonrpc.Executor/3::INFO::2017-05-05 10:39:13,088::__init__::513:: jsonrpc.JsonRpcServer::(_serveRequest) RPC call Host.getAllVmStats succeeded in 0.03 seconds
Em sex, 5 de mai de 2017 às 14:40, Yuval Turgeman <yuvalt@redhat.com> escreveu:
I take it updating everything to 4.0.6 is not an option ?
Thanks, Yuval
On May 5, 2017 6:16 PM, "Rogério Ceni Coelho" < rogeriocenicoelho@gmail.com> wrote:
I can think in two ways. Please let me know if have a change to go ok.
First, download ovirt-node-ng-image-update and ovirt-node-ng-image from 4.0.5 version and run yum localinstall.
Second, create a rpm list from other 4.0.5 node, do an diff agains my 4.0.3 node and use the diff to download packages from 4.0.5 version and run yum localintall.
My concern with this are that i can not upgrade ovirt engine every time i need to install a new ovirt node.
Thanks.
Em sex, 5 de mai de 2017 às 11:26, Rogério Ceni Coelho < rogeriocenicoelho@gmail.com> escreveu:
Hi oVirt Troopers,
I have two segregated oVirt Clusters running 4.0.5 ( DEV and PROD enviroments).
Now i need to install a new oVirt Node Server (Dell PowerEdge M620), but i see that does not exist an 4.0.5 iso on http://resources.ovirt.org/pub/ovirt-4.0/iso/ovirt-node-ng-installer/ , only 4.0.3 and 4.0.6.
How can i install this new server and go to the same version of all 20 others ???
There is a way to install 4.0.3 and update to 4.0.5 only ?
Thanks in advance.
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
participants (3)
-
Rogério Ceni Coelho
-
Yaniv Kaul
-
Yuval Turgeman