[Users] Booting oVirt node image 2.3.0, no install option
Adam vonNieda
adam at vonnieda.org
Tue Apr 17 18:48:54 UTC 2012
Turns out that there might be an issue with my thumb drive. I tried
another, and it worked fine. Thanks very much for the responses folks!
-Adam
On 4/17/12 10:11 AM, "Joey Boggs" <jboggs at redhat.com> wrote:
>On 04/17/2012 10:51 AM, Adam vonNieda wrote:
>> Thanks for the reply Joey. I saw that too, and thought maybe my USB
>>thumb drive was set to read only, but it's not. This box doesn't have a
>>DVD drive, I'll try a different USB drive, and if that doesn't work,
>>I'll dig up an external DVD drive.
>>
>> Thanks again,
>>
>> -Adam
>>
>> Adam vonNieda
>> Adam at vonNieda.org
>>
>> On Apr 17, 2012, at 9:07, Joey Boggs<jboggs at redhat.com> wrote:
>>
>>> On 04/17/2012 09:45 AM, Adam vonNieda wrote:
>>>> Hi folks,
>>>>
>>>> Still hoping someone can give me a hand with this. I can't install
>>>> overt-node 2.3.0 on a on a Dell C2100 server because it won't start
>>>>the
>>>> graphical interface. I booted up a standard F16 image this morning,
>>>>and
>>>> the graphical installer does start during that process. Logs are
>>>>below.
>>>>
>>>> Thanks very much,
>>>>
>>>> -Adam
>>>>
>>>>
>>>>> /tmp/ovirt.log
>>>>> ==============
>>>>>
>>>>> /sbin/restorecon set context
>>>>> /var/cache/yum->unconfined_u:object_r:rpm_var_cache_t:s0
>>>>>failed:'Read-only
>>>>> file system'
>>>>> /sbin/restorecon reset /var/cache/yum context
>>>>>
>>>>>unconfined_u:object_r:file_t:s0->unconfined_u:object_r:rpm_var_cache_t
>>>>>:s0
>>>>> /sbin/restorecon reset /etc/sysctl.conf context
>>>>>
>>>>>system_u:object_r:etc_runtime_t:s0->system_u:object_r:system_conf_t:s0
>>>>> /sbin/restorecon reset /boot-kdump context
>>>>> system_u:object_r:boot_t:s0->system_u:object_r:default_t:s0
>>>>> 2012-04-16 09:36:26,827 - INFO - ovirt-config-installer - ::::live
>>>>> device::::
>>>>> /dev/sdb
>>>>> 2012-04-16 09:36:26,836 - DEBUG - ovirtfunctions - cat
>>>>>/proc/mounts|grep
>>>>> -q "none /live"
>>>>> 2012-04-16 09:36:26,836 - DEBUG - ovirtfunctions -
>>>>> 2012-04-16 09:36:26,915 - DEBUG - ovirtfunctions - umount /live
>>>>> 2012-04-16 09:36:26,915 - DEBUG - ovirtfunctions -
>>>>> 2012-04-16 09:36:27,455 - ERROR - ovirtfunctions - Failed to
>>>>>mount_live()
>>>>>
>>>>> /var/log/ovirt.log
>>>>> ==================
>>>>>
>>>>> Apr 16 09:35:53 Starting ovirt-early
>>>>> oVirt Node Hypervisor release 2.3.0 (1.0.fc16)
>>>>> Apr 16 09:35:53 Updating /etc/default/ovirt
>>>>> Apr 16 09:35:54 Updating OVIRT_BOOTIF to ''
>>>>> Apr 16 09:35:54 Updating OVIRT_INIT to ''
>>>>> Apr 16 09:35:54 Updating OVIRT_UPGRADE to ''
>>>>> Apr 16 09:35:54 Updating OVIRT_STANDALONE to '1'
>>>>> Apr 16 09:35:54 Updating OVIRT_BOOTPARAMS to 'nomodeset
>>>>> crashkernel=512M-2G:64M,2G-:128M elevator=deadline quiet rd_NO_LVM
>>>>>rhgb
>>>>> rd.luks=0 rd.md=0 rd.dm=0'
>>>>> Apr 16 09:35:54 Updating OVIRT_RHN_TYPE to 'classic'
>>>>> Apr 16 09:35:54 Updating OVIRT_INSTALL to '1'
>>>>> Apr 16 09:35:54 Updating OVIRT_ISCSI_INSTALL to '1'
>>>>> Apr 16 09:36:08 Setting temporary admin password: F8Ax67kfRPSAw
>>>>> Apr 16 09:36:09 Setting temporary root password: F8Ax67kfRPSAw
>>>>> Apr 16 09:36:09 Skip runtime mode configuration.
>>>>> Apr 16 09:36:09 Completed ovirt-early
>>>>> Apr 16 09:36:09 Starting ovirt-awake.
>>>>> Apr 16 09:36:09 Node is operating in unmanaged mode.
>>>>> Apr 16 09:36:09 Completed ovirt-awake: RETVAL=0
>>>>> Apr 16 09:36:09 Starting ovirt
>>>>> Apr 16 09:36:09 Completed ovirt
>>>>> Apr 16 09:36:10 Starting ovirt-post
>>>>> Apr 16 09:36:20 Hardware virtualization detected
>>>>> Volume group "HostVG" not found
>>>>> Skipping volume group HostVG
>>>>> Restarting network (via systemctl): [ OK ]
>>>>> Apr 16 09:36:20 Starting ovirt-post
>>>>> Apr 16 09:36:21 Hardware virtualization detected
>>>>> Volume group "HostVG" not found
>>>>> Skipping volume group HostVG
>>>>> Restarting network (via systemctl): [ OK ]
>>>>> Apr 16 09:36:22 Starting ovirt-cim
>>>>> Apr 16 09:36:22 Completed ovirt-cim
>>>>> WARNING: persistent config storage not available
>>>>>
>>>>> /var/log/vdsm/vdsm.log
>>>>> =======================
>>>>>
>>>>> MainThread::INFO::2012-04-16 09:36:21,828::vdsm::71::vds::(run) I am
>>>>>the
>>>>> actual vdsm 4.9-0
>>>>> MainThread::DEBUG::2012-04-16
>>>>>
>>>>>09:36:23,873::resourceManager::376::ResourceManager::(registerNamespac
>>>>>e)
>>>>> Registering namespace 'Storage'
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:23,874::threadPool::45::Misc.ThreadPool::(__init__) Enter -
>>>>> numThreads: 10.0, waitTimeout: 3, maxTasks: 500.0
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:23,918::multipath::85::Storage.Misc.excCmd::(isEnabled)
>>>>> '/usr/bin/sudo -n /bin/cat /etc/multipath.conf' (cwd None)
>>>>> MainThread::INFO::2012-04-16 09:36:25,000::vdsm::71::vds::(run) I am
>>>>>the
>>>>> actual vdsm 4.9-0
>>>>> MainThread::DEBUG::2012-04-16
>>>>>
>>>>>09:36:25,199::resourceManager::376::ResourceManager::(registerNamespac
>>>>>e)
>>>>> Registering namespace 'Storage'
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:25,200::threadPool::45::Misc.ThreadPool::(__init__) Enter -
>>>>> numThreads: 10.0, waitTimeout: 3, maxTasks: 500.0
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:25,231::multipath::85::Storage.Misc.excCmd::(isEnabled)
>>>>> '/usr/bin/sudo -n /bin/cat /etc/multipath.conf' (cwd None)
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:25,243::multipath::85::Storage.Misc.excCmd::(isEnabled)
>>>>>SUCCESS:
>>>>> <err> = '';<rc> = 0
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:25,244::multipath::109::Storage.Multipath::(isEnabled)
>>>>>multipath
>>>>> Defaulting to False
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:25,244::misc::487::Storage.Misc::(rotateFiles) dir: /etc,
>>>>> prefixName: multipath.conf, versions: 5
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:25,244::misc::508::Storage.Misc::(rotateFiles) versions found:
>>>>>[0]
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:25,244::multipath::118::Storage.Misc.excCmd::(setupMultipath)
>>>>> '/usr/bin/sudo -n /bin/cp /etc/multipath.conf /etc/multipath.conf.1'
>>>>>(cwd
>>>>> None)
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:25,255::multipath::118::Storage.Misc.excCmd::(setupMultipath)
>>>>> FAILED:<err> = 'sudo: unable to mkdir /var/db/sudo/vdsm: Read-only
>>>>>file
>>>>> system\nsudo: sorry, a password is required to run sudo\n';<rc> = 1
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:25,256::multipath::118::Storage.Misc.excCmd::(setupMultipath)
>>>>> '/usr/bin/sudo -n /usr/sbin/persist /etc/multipath.conf.1' (cwd None)
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:25,269::multipath::118::Storage.Misc.excCmd::(setupMultipath)
>>>>> FAILED:<err> = 'sudo: unable to mkdir /var/db/sudo/vdsm: Read-only
>>>>>file
>>>>> system\nsudo: sorry, a password is required to run sudo\n';<rc> = 1
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:25,270::multipath::123::Storage.Misc.excCmd::(setupMultipath)
>>>>> '/usr/bin/sudo -n /bin/cp /tmp/tmpnPcvWi /etc/multipath.conf' (cwd
>>>>>None)
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:25,283::multipath::123::Storage.Misc.excCmd::(setupMultipath)
>>>>> SUCCESS:<err> = '';<rc> = 0
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:25,283::multipath::128::Storage.Misc.excCmd::(setupMultipath)
>>>>> '/usr/bin/sudo -n /usr/sbin/persist /etc/multipath.conf' (cwd None)
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:25,294::multipath::128::Storage.Misc.excCmd::(setupMultipath)
>>>>> FAILED:<err> = 'sudo: unable to mkdir /var/db/sudo/vdsm: Read-only
>>>>>file
>>>>> system\nsudo: sorry, a password is required to run sudo\n';<rc> = 1
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:25,295::multipath::131::Storage.Misc.excCmd::(setupMultipath)
>>>>> '/usr/bin/sudo -n /sbin/multipath -F' (cwd None)
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:25,323::multipath::131::Storage.Misc.excCmd::(setupMultipath)
>>>>> FAILED:<err> = '';<rc> = 1
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:25,323::multipath::134::Storage.Misc.excCmd::(setupMultipath)
>>>>> '/usr/bin/sudo -n /sbin/service multipathd restart' (cwd None)
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:26,397::multipath::134::Storage.Misc.excCmd::(setupMultipath)
>>>>> SUCCESS:<err> = '';<rc> = 0
>>>>> MainThread::DEBUG::2012-04-16
>>>>>
>>>>>09:36:26,398::hsm::248::Storage.Misc.excCmd::(__validateLvmLockingType
>>>>>)
>>>>> '/usr/bin/sudo -n /sbin/lvm dumpconfig global/locking_type' (cwd
>>>>>None)
>>>>> MainThread::DEBUG::2012-04-16
>>>>>
>>>>>09:36:26,443::hsm::248::Storage.Misc.excCmd::(__validateLvmLockingType
>>>>>)
>>>>> SUCCESS:<err> = '';<rc> = 0
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:26,445::lvm::319::OperationMutex::(_reloadpvs) Operation 'lvm
>>>>>reload
>>>>> operation' got the operation mutex
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:26,447::lvm::287::Storage.Misc.excCmd::(cmd) '/usr/bin/sudo -n
>>>>> /sbin/lvm pvs --config " devices { preferred_names =
>>>>>[\\"^/dev/mapper/\\"]
>>>>> ignore_suspended_devices=1 write_cache_state=0
>>>>>disable_after_error_count=3
>>>>> filter = [ \\"a%1SanDisk|3600605b00436bd80171b105c225377ce%\\",
>>>>> \\"r%.*%\\" ] } global { locking_type=1 prioritise_write_locks=1
>>>>> wait_for_locks=1 } backup { retain_min = 50 retain_days = 0 } "
>>>>> --noheadings --units b --nosuffix --separator | -o
>>>>>
>>>>>uuid,name,size,vg_name,vg_uuid,pe_start,pe_count,pe_alloc_count,mda_co
>>>>>unt,
>>>>> d
>>>>> ev_size' (cwd None)
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:26,811::lvm::287::Storage.Misc.excCmd::(cmd) SUCCESS:<err> =
>>>>>'';
>>>>> <rc> = 0
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:26,811::lvm::342::OperationMutex::(_reloadpvs) Operation 'lvm
>>>>>reload
>>>>> operation' released the operation mutex
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:26,812::lvm::352::OperationMutex::(_reloadvgs) Operation 'lvm
>>>>>reload
>>>>> operation' got the operation mutex
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:26,812::lvm::287::Storage.Misc.excCmd::(cmd) '/usr/bin/sudo -n
>>>>> /sbin/lvm vgs --config " devices { preferred_names =
>>>>>[\\"^/dev/mapper/\\"]
>>>>> ignore_suspended_devices=1 write_cache_state=0
>>>>>disable_after_error_count=3
>>>>> filter = [ \\"a%1SanDisk|3600605b00436bd80171b105c225377ce%\\",
>>>>> \\"r%.*%\\" ] } global { locking_type=1 prioritise_write_locks=1
>>>>> wait_for_locks=1 } backup { retain_min = 50 retain_days = 0 } "
>>>>> --noheadings --units b --nosuffix --separator | -o
>>>>>
>>>>>uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_m
>>>>>da_s
>>>>> i
>>>>> ze,vg_mda_free' (cwd None)
>>>>> MainThread::INFO::2012-04-16 09:36:29,307::vdsm::71::vds::(run) I am
>>>>>the
>>>>> actual vdsm 4.9-0
>>>>> MainThread::DEBUG::2012-04-16
>>>>>
>>>>>09:36:29,514::resourceManager::376::ResourceManager::(registerNamespac
>>>>>e)
>>>>> Registering namespace 'Storage'
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:29,515::threadPool::45::Misc.ThreadPool::(__init__) Enter -
>>>>> numThreads: 10.0, waitTimeout: 3, maxTasks: 500.0
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:29,551::multipath::85::Storage.Misc.excCmd::(isEnabled)
>>>>> '/usr/bin/sudo -n /bin/cat /etc/multipath.conf' (cwd None)
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:29,564::multipath::85::Storage.Misc.excCmd::(isEnabled)
>>>>>SUCCESS:
>>>>> <err> = '';<rc> = 0
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:29,565::multipath::101::Storage.Multipath::(isEnabled) Current
>>>>> revision of multipath.conf detected, preserving
>>>>> MainThread::DEBUG::2012-04-16
>>>>>
>>>>>09:36:29,565::hsm::248::Storage.Misc.excCmd::(__validateLvmLockingType
>>>>>)
>>>>> '/usr/bin/sudo -n /sbin/lvm dumpconfig global/locking_type' (cwd
>>>>>None)
>>>>> MainThread::DEBUG::2012-04-16
>>>>>
>>>>>09:36:29,606::hsm::248::Storage.Misc.excCmd::(__validateLvmLockingType
>>>>>)
>>>>> SUCCESS:<err> = '';<rc> = 0
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:29,606::lvm::319::OperationMutex::(_reloadpvs) Operation 'lvm
>>>>>reload
>>>>> operation' got the operation mutex
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:29,608::lvm::287::Storage.Misc.excCmd::(cmd) '/usr/bin/sudo -n
>>>>> /sbin/lvm pvs --config " devices { preferred_names =
>>>>>[\\"^/dev/mapper/\\"]
>>>>> ignore_suspended_devices=1 write_cache_state=0
>>>>>disable_after_error_count=3
>>>>> filter = [ \\"a%1SanDisk|3600605b00436bd80171b105c225377ce%\\",
>>>>> \\"r%.*%\\" ] } global { locking_type=1 prioritise_write_locks=1
>>>>> wait_for_locks=1 } backup { retain_min = 50 retain_days = 0 } "
>>>>> --noheadings --units b --nosuffix --separator | -o
>>>>>
>>>>>uuid,name,size,vg_name,vg_uuid,pe_start,pe_count,pe_alloc_count,mda_co
>>>>>unt,
>>>>> d
>>>>> ev_size' (cwd None)
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:29,714::lvm::287::Storage.Misc.excCmd::(cmd) SUCCESS:<err> =
>>>>>'';
>>>>> <rc> = 0
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:29,715::lvm::342::OperationMutex::(_reloadpvs) Operation 'lvm
>>>>>reload
>>>>> operation' released the operation mutex
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:29,716::lvm::352::OperationMutex::(_reloadvgs) Operation 'lvm
>>>>>reload
>>>>> operation' got the operation mutex
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:29,716::lvm::287::Storage.Misc.excCmd::(cmd) '/usr/bin/sudo -n
>>>>> /sbin/lvm vgs --config " devices { preferred_names =
>>>>>[\\"^/dev/mapper/\\"]
>>>>> ignore_suspended_devices=1 write_cache_state=0
>>>>>disable_after_error_count=3
>>>>> filter = [ \\"a%1SanDisk|3600605b00436bd80171b105c225377ce%\\",
>>>>> \\"r%.*%\\" ] } global { locking_type=1 prioritise_write_locks=1
>>>>> wait_for_locks=1 } backup { retain_min = 50 retain_days = 0 } "
>>>>> --noheadings --units b --nosuffix --separator | -o
>>>>>
>>>>>uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_m
>>>>>da_s
>>>>> i
>>>>> ze,vg_mda_free' (cwd None)
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:29,813::lvm::287::Storage.Misc.excCmd::(cmd) SUCCESS:<err> =
>>>>>' No
>>>>> volume groups found\n';<rc> = 0
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:29,814::lvm::379::OperationMutex::(_reloadvgs) Operation 'lvm
>>>>>reload
>>>>> operation' released the operation mutex
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:29,815::lvm::287::Storage.Misc.excCmd::(cmd) '/usr/bin/sudo -n
>>>>> /sbin/lvm lvs --config " devices { preferred_names =
>>>>>[\\"^/dev/mapper/\\"]
>>>>> ignore_suspended_devices=1 write_cache_state=0
>>>>>disable_after_error_count=3
>>>>> filter = [ \\"a%1SanDisk|3600605b00436bd80171b105c225377ce%\\",
>>>>> \\"r%.*%\\" ] } global { locking_type=1 prioritise_write_locks=1
>>>>> wait_for_locks=1 } backup { retain_min = 50 retain_days = 0 } "
>>>>> --noheadings --units b --nosuffix --separator | -o
>>>>> uuid,name,vg_name,attr,size,seg_start_pe,devices,tags' (cwd None)
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:29,916::lvm::287::Storage.Misc.excCmd::(cmd) SUCCESS:<err> =
>>>>>' No
>>>>> volume groups found\n';<rc> = 0
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:29,917::misc::1017::SamplingMethod::(__call__) Trying to enter
>>>>> sampling method (storage.sdc.refreshStorage)
>>>>> MainThread::INFO::2012-04-16
>>>>> 09:36:29,919::dispatcher::121::Storage.Dispatcher::(__init__)
>>>>>Starting
>>>>> StorageDispatcher...
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:29,919::misc::1019::SamplingMethod::(__call__) Got in to
>>>>>sampling
>>>>> method
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:29,921::misc::1017::SamplingMethod::(__call__) Trying to enter
>>>>> sampling method (storage.iscsi.rescan)
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:29,921::misc::1019::SamplingMethod::(__call__) Got in to
>>>>>sampling
>>>>> method
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:29,921::iscsi::389::Storage.Misc.excCmd::(rescan)
>>>>>'/usr/bin/sudo -n
>>>>> /sbin/iscsiadm -m session -R' (cwd None)
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:29,930::utils::595::Storage.Misc.excCmd::(execCmd)
>>>>>'/usr/bin/pgrep
>>>>> -xf ksmd' (cwd None)
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:30,108::utils::595::Storage.Misc.excCmd::(execCmd)
>>>>>SUCCESS:<err> =
>>>>> '';<rc> = 0
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:30,116::iscsi::389::Storage.Misc.excCmd::(rescan) FAILED:<err>
>>>>> =
>>>>> 'iscsiadm: No session found.\n';<rc> = 21
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:30,116::misc::1027::SamplingMethod::(__call__) Returning last
>>>>>result
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:30,117::supervdsm::83::SuperVdsmProxy::(_killSupervdsm) Could
>>>>>not
>>>>> kill old Super Vdsm [Errno 2] No such file or directory:
>>>>> '/var/run/vdsm/svdsm.pid'
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:30,117::supervdsm::71::SuperVdsmProxy::(_launchSupervdsm)
>>>>>Launching
>>>>> Super Vdsm
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:30,118::supervdsm::74::Storage.Misc.excCmd::(_launchSupervdsm)
>>>>> '/usr/bin/sudo -n /usr/bin/python /usr/share/vdsm/supervdsmServer.pyc
>>>>> bd4b3ae7-3e51-4d6b-b681-d5f6cb5bae07 2945' (cwd None)
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:30,254::supervdsmServer::170::SuperVdsm.Server::(main) Making
>>>>>sure
>>>>> I'm root
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:30,255::supervdsmServer::174::SuperVdsm.Server::(main) Parsing
>>>>>cmd
>>>>> args
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:30,255::supervdsmServer::177::SuperVdsm.Server::(main)
>>>>>Creating PID
>>>>> file
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:30,255::supervdsmServer::181::SuperVdsm.Server::(main)
>>>>>Cleaning old
>>>>> socket
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:30,255::supervdsmServer::185::SuperVdsm.Server::(main) Setting
>>>>>up
>>>>> keep alive thread
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:30,256::supervdsmServer::190::SuperVdsm.Server::(main) Creating
>>>>> remote object manager
>>>>> MainThread::DEBUG::2012-04-16
>>>>> 09:36:30,256::supervdsmServer::201::SuperVdsm.Server::(main) Started
>>>>> serving super vdsm object
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:32,124::supervdsm::92::SuperVdsmProxy::(_connect) Trying to
>>>>>connect
>>>>> to Super Vdsm
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:32,133::supervdsm::64::SuperVdsmProxy::(__init__) Connected to
>>>>>Super
>>>>> Vdsm
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:34,070::multipath::71::Storage.Misc.excCmd::(rescan)
>>>>>'/usr/bin/sudo
>>>>> -n /sbin/multipath' (cwd None)
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:34,130::multipath::71::Storage.Misc.excCmd::(rescan)
>>>>>SUCCESS:<err>
>>>>> = '';<rc> = 0
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:34,131::lvm::460::OperationMutex::(_invalidateAllPvs)
>>>>>Operation 'lvm
>>>>> invalidate operation' got the operation mutex
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:34,131::lvm::462::OperationMutex::(_invalidateAllPvs)
>>>>>Operation 'lvm
>>>>> invalidate operation' released the operation mutex
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:34,132::lvm::472::OperationMutex::(_invalidateAllVgs)
>>>>>Operation 'lvm
>>>>> invalidate operation' got the operation mutex
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:34,132::lvm::474::OperationMutex::(_invalidateAllVgs)
>>>>>Operation 'lvm
>>>>> invalidate operation' released the operation mutex
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:34,133::lvm::493::OperationMutex::(_invalidateAllLvs)
>>>>>Operation 'lvm
>>>>> invalidate operation' got the operation mutex
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:34,133::lvm::495::OperationMutex::(_invalidateAllLvs)
>>>>>Operation 'lvm
>>>>> invalidate operation' released the operation mutex
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:34,133::misc::1027::SamplingMethod::(__call__) Returning last
>>>>>result
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:34,133::hsm::272::Storage.HSM::(__cleanStorageRepository)
>>>>>Started
>>>>> cleaning storage repository at '/rhev/data-center'
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:34,136::hsm::304::Storage.HSM::(__cleanStorageRepository) White
>>>>> list: ['/rhev/data-center/hsm-tasks',
>>>>>'/rhev/data-center/hsm-tasks/*',
>>>>> '/rhev/data-center/mnt']
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:34,136::hsm::305::Storage.HSM::(__cleanStorageRepository) Mount
>>>>> list: ['/rhev/data-center']
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:34,136::hsm::307::Storage.HSM::(__cleanStorageRepository)
>>>>>Cleaning
>>>>> leftovers
>>>>> Thread-11::DEBUG::2012-04-16
>>>>> 09:36:34,136::hsm::350::Storage.HSM::(__cleanStorageRepository)
>>>>>Finished
>>>>> cleaning storage repository at '/rhev/data-center'
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> On 4/16/12 8:38 AM, "Mike Burns"<mburns at redhat.com> wrote:
>>>>>
>>>>>> On Mon, 2012-04-16 at 08:14 -0500, Adam vonNieda wrote:
>>>>>>> Hi folks,
>>>>>>>
>>>>>>>
>>>>>>> I'm trying to install oVirt node v2.3.0 on A Dell C2100
>>>>>>>server. I
>>>>>>> can boot up just fine, but the two menu options I see are "Start
>>>>>>>oVirt
>>>>>>> node", and "Troubleshooting". When I choose "Start oVirt node", it
>>>>>>> does just that, and I am soon after given a console login prompt.
>>>>>>>I've
>>>>>>> checked the docs, and I don't see what I'm supposed to do next, as
>>>>>>>in
>>>>>>> a password etc. Am I missing something?
>>>>>> Hi Adam,
>>>>>>
>>>>>> Something is breaking in the boot process. You should be getting a
>>>>>>TUI
>>>>>> screen that will let you configure and install ovirt-node.
>>>>>>
>>>>>> I just added an entry on the Node Troublesooting wiki page[1] for
>>>>>>you to
>>>>>> follow.
>>>>>>
>>>>>> Mike
>>>>>>
>>>>>> [1] http://ovirt.org/wiki/Node_Troubleshooting#Boot_up_problems
>>>>>>
>>>>>>
>>>>>>> Thanks,
>>>>>>>
>>>>>>>
>>>>>>> -Adam
>>>>>>> _______________________________________________
>>>>>>> Users mailing list
>>>>>>> Users at ovirt.org
>>>>>>> http://lists.ovirt.org/mailman/listinfo/users
>>>> _______________________________________________
>>>> Users mailing list
>>>> Users at ovirt.org
>>>> http://lists.ovirt.org/mailman/listinfo/users
>>> This is definitely the cause of the installer failing
>>>
>>> 2012-04-16 09:36:26,836 - DEBUG - ovirtfunctions - cat
>>>/proc/mounts|grep -q "none /live"
>>> 2012-04-16 09:36:27,455 - ERROR - ovirtfunctions - Failed to
>>>mount_live()
>>>
>>>
>>>
>>> What kind of media are you installing from: usb/cd/remote console?
>> _______________________________________________
>> Users mailing list
>> Users at ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>
>I did go back and take a look at mount_live and made sure it contains a
>specific patch to handle usb drives properly. If you can get back to a
>shell prompt. run blkid and capture the output. If it's way too much to
>type then just the usb drive output should be ok.
More information about the Users
mailing list