[Users] Booting oVirt node image 2.3.0, no install option

Joey Boggs jboggs at redhat.com
Tue Apr 17 15:11:00 UTC 2012


On 04/17/2012 10:51 AM, Adam vonNieda wrote:
>     Thanks for the reply Joey. I saw that too, and thought maybe my USB thumb drive was set to read only, but it's not. This box doesn't have a DVD drive, I'll try a different USB drive, and if that doesn't work, I'll dig up an external DVD drive.
>
>     Thanks again,
>
>        -Adam
>
> Adam vonNieda
> Adam at vonNieda.org
>
> On Apr 17, 2012, at 9:07, Joey Boggs<jboggs at redhat.com>  wrote:
>
>> On 04/17/2012 09:45 AM, Adam vonNieda wrote:
>>>     Hi folks,
>>>
>>>     Still hoping someone can give me a hand with this. I can't install
>>> overt-node 2.3.0 on a on a Dell C2100 server because it won't start the
>>> graphical interface. I booted up a standard F16 image this morning, and
>>> the graphical installer does start during that process. Logs are below.
>>>
>>>     Thanks very much,
>>>
>>>        -Adam
>>>
>>>
>>>> /tmp/ovirt.log
>>>> ==============
>>>>
>>>> /sbin/restorecon set context
>>>> /var/cache/yum->unconfined_u:object_r:rpm_var_cache_t:s0 failed:'Read-only
>>>> file system'
>>>> /sbin/restorecon reset /var/cache/yum context
>>>> unconfined_u:object_r:file_t:s0->unconfined_u:object_r:rpm_var_cache_t:s0
>>>> /sbin/restorecon reset /etc/sysctl.conf context
>>>> system_u:object_r:etc_runtime_t:s0->system_u:object_r:system_conf_t:s0
>>>> /sbin/restorecon reset /boot-kdump context
>>>> system_u:object_r:boot_t:s0->system_u:object_r:default_t:s0
>>>> 2012-04-16 09:36:26,827 - INFO - ovirt-config-installer - ::::live
>>>> device::::
>>>> /dev/sdb
>>>> 2012-04-16 09:36:26,836 - DEBUG - ovirtfunctions - cat /proc/mounts|grep
>>>> -q "none /live"
>>>> 2012-04-16 09:36:26,836 - DEBUG - ovirtfunctions -
>>>> 2012-04-16 09:36:26,915 - DEBUG - ovirtfunctions - umount /live
>>>> 2012-04-16 09:36:26,915 - DEBUG - ovirtfunctions -
>>>> 2012-04-16 09:36:27,455 - ERROR - ovirtfunctions - Failed to mount_live()
>>>>
>>>> /var/log/ovirt.log
>>>> ==================
>>>>
>>>> Apr 16 09:35:53 Starting ovirt-early
>>>> oVirt Node Hypervisor release 2.3.0 (1.0.fc16)
>>>> Apr 16 09:35:53 Updating /etc/default/ovirt
>>>> Apr 16 09:35:54 Updating OVIRT_BOOTIF to ''
>>>> Apr 16 09:35:54 Updating OVIRT_INIT to ''
>>>> Apr 16 09:35:54 Updating OVIRT_UPGRADE to ''
>>>> Apr 16 09:35:54 Updating OVIRT_STANDALONE to '1'
>>>> Apr 16 09:35:54 Updating OVIRT_BOOTPARAMS to 'nomodeset
>>>> crashkernel=512M-2G:64M,2G-:128M elevator=deadline quiet rd_NO_LVM rhgb
>>>> rd.luks=0 rd.md=0 rd.dm=0'
>>>> Apr 16 09:35:54 Updating OVIRT_RHN_TYPE to 'classic'
>>>> Apr 16 09:35:54 Updating OVIRT_INSTALL to '1'
>>>> Apr 16 09:35:54 Updating OVIRT_ISCSI_INSTALL to '1'
>>>> Apr 16 09:36:08 Setting temporary admin password: F8Ax67kfRPSAw
>>>> Apr 16 09:36:09 Setting temporary root password: F8Ax67kfRPSAw
>>>> Apr 16 09:36:09 Skip runtime mode configuration.
>>>> Apr 16 09:36:09 Completed ovirt-early
>>>> Apr 16 09:36:09 Starting ovirt-awake.
>>>> Apr 16 09:36:09 Node is operating in unmanaged mode.
>>>> Apr 16 09:36:09 Completed ovirt-awake: RETVAL=0
>>>> Apr 16 09:36:09 Starting ovirt
>>>> Apr 16 09:36:09 Completed ovirt
>>>> Apr 16 09:36:10 Starting ovirt-post
>>>> Apr 16 09:36:20 Hardware virtualization detected
>>>>   Volume group "HostVG" not found
>>>>   Skipping volume group HostVG
>>>> Restarting network (via systemctl):  [  OK  ]
>>>> Apr 16 09:36:20 Starting ovirt-post
>>>> Apr 16 09:36:21 Hardware virtualization detected
>>>>   Volume group "HostVG" not found
>>>>   Skipping volume group HostVG
>>>> Restarting network (via systemctl):  [  OK  ]
>>>> Apr 16 09:36:22 Starting ovirt-cim
>>>> Apr 16 09:36:22 Completed ovirt-cim
>>>> WARNING: persistent config storage not available
>>>>
>>>> /var/log/vdsm/vdsm.log
>>>> =======================
>>>>
>>>> MainThread::INFO::2012-04-16 09:36:21,828::vdsm::71::vds::(run) I am the
>>>> actual vdsm 4.9-0
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:23,873::resourceManager::376::ResourceManager::(registerNamespace)
>>>> Registering namespace 'Storage'
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:23,874::threadPool::45::Misc.ThreadPool::(__init__) Enter -
>>>> numThreads: 10.0, waitTimeout: 3, maxTasks: 500.0
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:23,918::multipath::85::Storage.Misc.excCmd::(isEnabled)
>>>> '/usr/bin/sudo -n /bin/cat /etc/multipath.conf' (cwd None)
>>>> MainThread::INFO::2012-04-16 09:36:25,000::vdsm::71::vds::(run) I am the
>>>> actual vdsm 4.9-0
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:25,199::resourceManager::376::ResourceManager::(registerNamespace)
>>>> Registering namespace 'Storage'
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:25,200::threadPool::45::Misc.ThreadPool::(__init__) Enter -
>>>> numThreads: 10.0, waitTimeout: 3, maxTasks: 500.0
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:25,231::multipath::85::Storage.Misc.excCmd::(isEnabled)
>>>> '/usr/bin/sudo -n /bin/cat /etc/multipath.conf' (cwd None)
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:25,243::multipath::85::Storage.Misc.excCmd::(isEnabled) SUCCESS:
>>>> <err>   = '';<rc>   = 0
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:25,244::multipath::109::Storage.Multipath::(isEnabled) multipath
>>>> Defaulting to False
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:25,244::misc::487::Storage.Misc::(rotateFiles) dir: /etc,
>>>> prefixName: multipath.conf, versions: 5
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:25,244::misc::508::Storage.Misc::(rotateFiles) versions found: [0]
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:25,244::multipath::118::Storage.Misc.excCmd::(setupMultipath)
>>>> '/usr/bin/sudo -n /bin/cp /etc/multipath.conf /etc/multipath.conf.1' (cwd
>>>> None)
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:25,255::multipath::118::Storage.Misc.excCmd::(setupMultipath)
>>>> FAILED:<err>   = 'sudo: unable to mkdir /var/db/sudo/vdsm: Read-only file
>>>> system\nsudo: sorry, a password is required to run sudo\n';<rc>   = 1
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:25,256::multipath::118::Storage.Misc.excCmd::(setupMultipath)
>>>> '/usr/bin/sudo -n /usr/sbin/persist /etc/multipath.conf.1' (cwd None)
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:25,269::multipath::118::Storage.Misc.excCmd::(setupMultipath)
>>>> FAILED:<err>   = 'sudo: unable to mkdir /var/db/sudo/vdsm: Read-only file
>>>> system\nsudo: sorry, a password is required to run sudo\n';<rc>   = 1
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:25,270::multipath::123::Storage.Misc.excCmd::(setupMultipath)
>>>> '/usr/bin/sudo -n /bin/cp /tmp/tmpnPcvWi /etc/multipath.conf' (cwd None)
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:25,283::multipath::123::Storage.Misc.excCmd::(setupMultipath)
>>>> SUCCESS:<err>   = '';<rc>   = 0
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:25,283::multipath::128::Storage.Misc.excCmd::(setupMultipath)
>>>> '/usr/bin/sudo -n /usr/sbin/persist /etc/multipath.conf' (cwd None)
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:25,294::multipath::128::Storage.Misc.excCmd::(setupMultipath)
>>>> FAILED:<err>   = 'sudo: unable to mkdir /var/db/sudo/vdsm: Read-only file
>>>> system\nsudo: sorry, a password is required to run sudo\n';<rc>   = 1
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:25,295::multipath::131::Storage.Misc.excCmd::(setupMultipath)
>>>> '/usr/bin/sudo -n /sbin/multipath -F' (cwd None)
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:25,323::multipath::131::Storage.Misc.excCmd::(setupMultipath)
>>>> FAILED:<err>   = '';<rc>   = 1
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:25,323::multipath::134::Storage.Misc.excCmd::(setupMultipath)
>>>> '/usr/bin/sudo -n /sbin/service multipathd restart' (cwd None)
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:26,397::multipath::134::Storage.Misc.excCmd::(setupMultipath)
>>>> SUCCESS:<err>   = '';<rc>   = 0
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:26,398::hsm::248::Storage.Misc.excCmd::(__validateLvmLockingType)
>>>> '/usr/bin/sudo -n /sbin/lvm dumpconfig global/locking_type' (cwd None)
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:26,443::hsm::248::Storage.Misc.excCmd::(__validateLvmLockingType)
>>>> SUCCESS:<err>   = '';<rc>   = 0
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:26,445::lvm::319::OperationMutex::(_reloadpvs) Operation 'lvm reload
>>>> operation' got the operation mutex
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:26,447::lvm::287::Storage.Misc.excCmd::(cmd) '/usr/bin/sudo -n
>>>> /sbin/lvm pvs --config " devices { preferred_names = [\\"^/dev/mapper/\\"]
>>>> ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
>>>> filter = [ \\"a%1SanDisk|3600605b00436bd80171b105c225377ce%\\",
>>>> \\"r%.*%\\" ] }  global {  locking_type=1  prioritise_write_locks=1
>>>> wait_for_locks=1 }  backup {  retain_min = 50  retain_days = 0 } "
>>>> --noheadings --units b --nosuffix --separator | -o
>>>> uuid,name,size,vg_name,vg_uuid,pe_start,pe_count,pe_alloc_count,mda_count,
>>>> d
>>>> ev_size' (cwd None)
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:26,811::lvm::287::Storage.Misc.excCmd::(cmd) SUCCESS:<err>   = '';
>>>> <rc>   = 0
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:26,811::lvm::342::OperationMutex::(_reloadpvs) Operation 'lvm reload
>>>> operation' released the operation mutex
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:26,812::lvm::352::OperationMutex::(_reloadvgs) Operation 'lvm reload
>>>> operation' got the operation mutex
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:26,812::lvm::287::Storage.Misc.excCmd::(cmd) '/usr/bin/sudo -n
>>>> /sbin/lvm vgs --config " devices { preferred_names = [\\"^/dev/mapper/\\"]
>>>> ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
>>>> filter = [ \\"a%1SanDisk|3600605b00436bd80171b105c225377ce%\\",
>>>> \\"r%.*%\\" ] }  global {  locking_type=1  prioritise_write_locks=1
>>>> wait_for_locks=1 }  backup {  retain_min = 50  retain_days = 0 } "
>>>> --noheadings --units b --nosuffix --separator | -o
>>>> uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_s
>>>> i
>>>> ze,vg_mda_free' (cwd None)
>>>> MainThread::INFO::2012-04-16 09:36:29,307::vdsm::71::vds::(run) I am the
>>>> actual vdsm 4.9-0
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:29,514::resourceManager::376::ResourceManager::(registerNamespace)
>>>> Registering namespace 'Storage'
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:29,515::threadPool::45::Misc.ThreadPool::(__init__) Enter -
>>>> numThreads: 10.0, waitTimeout: 3, maxTasks: 500.0
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:29,551::multipath::85::Storage.Misc.excCmd::(isEnabled)
>>>> '/usr/bin/sudo -n /bin/cat /etc/multipath.conf' (cwd None)
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:29,564::multipath::85::Storage.Misc.excCmd::(isEnabled) SUCCESS:
>>>> <err>   = '';<rc>   = 0
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:29,565::multipath::101::Storage.Multipath::(isEnabled) Current
>>>> revision of multipath.conf detected, preserving
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:29,565::hsm::248::Storage.Misc.excCmd::(__validateLvmLockingType)
>>>> '/usr/bin/sudo -n /sbin/lvm dumpconfig global/locking_type' (cwd None)
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:29,606::hsm::248::Storage.Misc.excCmd::(__validateLvmLockingType)
>>>> SUCCESS:<err>   = '';<rc>   = 0
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:29,606::lvm::319::OperationMutex::(_reloadpvs) Operation 'lvm reload
>>>> operation' got the operation mutex
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:29,608::lvm::287::Storage.Misc.excCmd::(cmd) '/usr/bin/sudo -n
>>>> /sbin/lvm pvs --config " devices { preferred_names = [\\"^/dev/mapper/\\"]
>>>> ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
>>>> filter = [ \\"a%1SanDisk|3600605b00436bd80171b105c225377ce%\\",
>>>> \\"r%.*%\\" ] }  global {  locking_type=1  prioritise_write_locks=1
>>>> wait_for_locks=1 }  backup {  retain_min = 50  retain_days = 0 } "
>>>> --noheadings --units b --nosuffix --separator | -o
>>>> uuid,name,size,vg_name,vg_uuid,pe_start,pe_count,pe_alloc_count,mda_count,
>>>> d
>>>> ev_size' (cwd None)
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:29,714::lvm::287::Storage.Misc.excCmd::(cmd) SUCCESS:<err>   = '';
>>>> <rc>   = 0
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:29,715::lvm::342::OperationMutex::(_reloadpvs) Operation 'lvm reload
>>>> operation' released the operation mutex
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:29,716::lvm::352::OperationMutex::(_reloadvgs) Operation 'lvm reload
>>>> operation' got the operation mutex
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:29,716::lvm::287::Storage.Misc.excCmd::(cmd) '/usr/bin/sudo -n
>>>> /sbin/lvm vgs --config " devices { preferred_names = [\\"^/dev/mapper/\\"]
>>>> ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
>>>> filter = [ \\"a%1SanDisk|3600605b00436bd80171b105c225377ce%\\",
>>>> \\"r%.*%\\" ] }  global {  locking_type=1  prioritise_write_locks=1
>>>> wait_for_locks=1 }  backup {  retain_min = 50  retain_days = 0 } "
>>>> --noheadings --units b --nosuffix --separator | -o
>>>> uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_s
>>>> i
>>>> ze,vg_mda_free' (cwd None)
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:29,813::lvm::287::Storage.Misc.excCmd::(cmd) SUCCESS:<err>   = '  No
>>>> volume groups found\n';<rc>   = 0
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:29,814::lvm::379::OperationMutex::(_reloadvgs) Operation 'lvm reload
>>>> operation' released the operation mutex
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:29,815::lvm::287::Storage.Misc.excCmd::(cmd) '/usr/bin/sudo -n
>>>> /sbin/lvm lvs --config " devices { preferred_names = [\\"^/dev/mapper/\\"]
>>>> ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
>>>> filter = [ \\"a%1SanDisk|3600605b00436bd80171b105c225377ce%\\",
>>>> \\"r%.*%\\" ] }  global {  locking_type=1  prioritise_write_locks=1
>>>> wait_for_locks=1 }  backup {  retain_min = 50  retain_days = 0 } "
>>>> --noheadings --units b --nosuffix --separator | -o
>>>> uuid,name,vg_name,attr,size,seg_start_pe,devices,tags' (cwd None)
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:29,916::lvm::287::Storage.Misc.excCmd::(cmd) SUCCESS:<err>   = '  No
>>>> volume groups found\n';<rc>   = 0
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:29,917::misc::1017::SamplingMethod::(__call__) Trying to enter
>>>> sampling method (storage.sdc.refreshStorage)
>>>> MainThread::INFO::2012-04-16
>>>> 09:36:29,919::dispatcher::121::Storage.Dispatcher::(__init__) Starting
>>>> StorageDispatcher...
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:29,919::misc::1019::SamplingMethod::(__call__) Got in to sampling
>>>> method
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:29,921::misc::1017::SamplingMethod::(__call__) Trying to enter
>>>> sampling method (storage.iscsi.rescan)
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:29,921::misc::1019::SamplingMethod::(__call__) Got in to sampling
>>>> method
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:29,921::iscsi::389::Storage.Misc.excCmd::(rescan) '/usr/bin/sudo -n
>>>> /sbin/iscsiadm -m session -R' (cwd None)
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:29,930::utils::595::Storage.Misc.excCmd::(execCmd) '/usr/bin/pgrep
>>>> -xf ksmd' (cwd None)
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:30,108::utils::595::Storage.Misc.excCmd::(execCmd) SUCCESS:<err>   =
>>>> '';<rc>   = 0
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:30,116::iscsi::389::Storage.Misc.excCmd::(rescan) FAILED:<err>   =
>>>> 'iscsiadm: No session found.\n';<rc>   = 21
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:30,116::misc::1027::SamplingMethod::(__call__) Returning last result
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:30,117::supervdsm::83::SuperVdsmProxy::(_killSupervdsm) Could not
>>>> kill old Super Vdsm [Errno 2] No such file or directory:
>>>> '/var/run/vdsm/svdsm.pid'
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:30,117::supervdsm::71::SuperVdsmProxy::(_launchSupervdsm) Launching
>>>> Super Vdsm
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:30,118::supervdsm::74::Storage.Misc.excCmd::(_launchSupervdsm)
>>>> '/usr/bin/sudo -n /usr/bin/python /usr/share/vdsm/supervdsmServer.pyc
>>>> bd4b3ae7-3e51-4d6b-b681-d5f6cb5bae07 2945' (cwd None)
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:30,254::supervdsmServer::170::SuperVdsm.Server::(main) Making sure
>>>> I'm root
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:30,255::supervdsmServer::174::SuperVdsm.Server::(main) Parsing cmd
>>>> args
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:30,255::supervdsmServer::177::SuperVdsm.Server::(main) Creating PID
>>>> file
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:30,255::supervdsmServer::181::SuperVdsm.Server::(main) Cleaning old
>>>> socket
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:30,255::supervdsmServer::185::SuperVdsm.Server::(main) Setting up
>>>> keep alive thread
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:30,256::supervdsmServer::190::SuperVdsm.Server::(main) Creating
>>>> remote object manager
>>>> MainThread::DEBUG::2012-04-16
>>>> 09:36:30,256::supervdsmServer::201::SuperVdsm.Server::(main) Started
>>>> serving super vdsm object
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:32,124::supervdsm::92::SuperVdsmProxy::(_connect) Trying to connect
>>>> to Super Vdsm
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:32,133::supervdsm::64::SuperVdsmProxy::(__init__) Connected to Super
>>>> Vdsm
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:34,070::multipath::71::Storage.Misc.excCmd::(rescan) '/usr/bin/sudo
>>>> -n /sbin/multipath' (cwd None)
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:34,130::multipath::71::Storage.Misc.excCmd::(rescan) SUCCESS:<err>
>>>> = '';<rc>   = 0
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:34,131::lvm::460::OperationMutex::(_invalidateAllPvs) Operation 'lvm
>>>> invalidate operation' got the operation mutex
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:34,131::lvm::462::OperationMutex::(_invalidateAllPvs) Operation 'lvm
>>>> invalidate operation' released the operation mutex
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:34,132::lvm::472::OperationMutex::(_invalidateAllVgs) Operation 'lvm
>>>> invalidate operation' got the operation mutex
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:34,132::lvm::474::OperationMutex::(_invalidateAllVgs) Operation 'lvm
>>>> invalidate operation' released the operation mutex
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:34,133::lvm::493::OperationMutex::(_invalidateAllLvs) Operation 'lvm
>>>> invalidate operation' got the operation mutex
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:34,133::lvm::495::OperationMutex::(_invalidateAllLvs) Operation 'lvm
>>>> invalidate operation' released the operation mutex
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:34,133::misc::1027::SamplingMethod::(__call__) Returning last result
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:34,133::hsm::272::Storage.HSM::(__cleanStorageRepository) Started
>>>> cleaning storage repository at '/rhev/data-center'
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:34,136::hsm::304::Storage.HSM::(__cleanStorageRepository) White
>>>> list: ['/rhev/data-center/hsm-tasks', '/rhev/data-center/hsm-tasks/*',
>>>> '/rhev/data-center/mnt']
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:34,136::hsm::305::Storage.HSM::(__cleanStorageRepository) Mount
>>>> list: ['/rhev/data-center']
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:34,136::hsm::307::Storage.HSM::(__cleanStorageRepository) Cleaning
>>>> leftovers
>>>> Thread-11::DEBUG::2012-04-16
>>>> 09:36:34,136::hsm::350::Storage.HSM::(__cleanStorageRepository) Finished
>>>> cleaning storage repository at '/rhev/data-center'
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> On 4/16/12 8:38 AM, "Mike Burns"<mburns at redhat.com>   wrote:
>>>>
>>>>> On Mon, 2012-04-16 at 08:14 -0500, Adam vonNieda wrote:
>>>>>>     Hi folks,
>>>>>>
>>>>>>
>>>>>>     I'm trying to install oVirt node v2.3.0 on A Dell C2100 server. I
>>>>>> can boot up just fine, but the two menu options I see are "Start oVirt
>>>>>> node", and "Troubleshooting". When I choose "Start oVirt node", it
>>>>>> does just that, and I am soon after given a console login prompt. I've
>>>>>> checked the docs, and I don't see what I'm supposed to do next, as in
>>>>>> a password etc. Am I missing something?
>>>>> Hi Adam,
>>>>>
>>>>> Something is breaking in the boot process.  You should be getting a TUI
>>>>> screen that will let you configure and install ovirt-node.
>>>>>
>>>>> I just added an entry on the Node Troublesooting wiki page[1] for you to
>>>>> follow.
>>>>>
>>>>> Mike
>>>>>
>>>>> [1] http://ovirt.org/wiki/Node_Troubleshooting#Boot_up_problems
>>>>>
>>>>>
>>>>>>     Thanks,
>>>>>>
>>>>>>
>>>>>>        -Adam
>>>>>> _______________________________________________
>>>>>> Users mailing list
>>>>>> Users at ovirt.org
>>>>>> http://lists.ovirt.org/mailman/listinfo/users
>>> _______________________________________________
>>> Users mailing list
>>> Users at ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/users
>> This is definitely the cause of the installer failing
>>
>> 2012-04-16 09:36:26,836 - DEBUG - ovirtfunctions - cat /proc/mounts|grep -q "none /live"
>> 2012-04-16 09:36:27,455 - ERROR - ovirtfunctions - Failed to mount_live()
>>
>>
>>
>> What kind of media are you installing from: usb/cd/remote console?
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users

I did go back and take a look at mount_live and made sure it contains a 
specific patch to handle usb drives properly. If you can get back to a 
shell prompt. run blkid and capture the output. If it's way too much to 
type then just the usb drive output should be ok.



More information about the Users mailing list