
On Tue, Jan 15, 2013 at 12:14:45PM +0100, Gianluca Cecchi wrote:
On Tue, Jan 15, 2013 at 11:49 AM, Dan Kenigsberg wrote:
vdsClient -s 0 getVdsCaps
would access localhost vdsm (assuming you are using the default ssl). If it barfs, we have a problem.
# systemctl status vdsmd.service vdsmd.service - Virtual Desktop Server Manager Loaded: loaded (/usr/lib/systemd/system/vdsmd.service; enabled) Active: inactive (dead)
So vdsm is enabled, it should've been up after reboot. But it has died. Maybe you can findout when it has happened, and correlate this to vdsm.log or libvirtd.log for clues. I'm out of guesses right now.
CGroup: name=systemd:/system/vdsmd.service
# vdsClient -s 0 getVdsCaps Connection to 10.4.4.59:54321 refused
# systemctl start vdsmd.service #
# systemctl status vdsmd.service vdsmd.service - Virtual Desktop Server Manager Loaded: loaded (/usr/lib/systemd/system/vdsmd.service; enabled) Active: active (running) since Tue, 2013-01-15 12:09:11 CET; 15s ago Process: 6566 ExecStart=/lib/systemd/systemd-vdsmd start (code=exited, status=0/SUCCESS) Main PID: 6834 (respawn) CGroup: name=systemd:/system/vdsmd.service ├ 6834 /bin/bash -e /usr/share/vdsm/respawn --minlifetime 10 --daemon --masterpid /var/run/vdsm/respawn.pid ... ├ 6837 /usr/bin/python /usr/share/vdsm/vdsm ├ 6855 /usr/bin/sudo -n /usr/bin/python /usr/share/vdsm/supervdsmServer.py d1587fa5-b439-4a5e-bca3-e3971dc08... └ 6856 /usr/bin/python /usr/share/vdsm/supervdsmServer.py d1587fa5-b439-4a5e-bca3-e3971dc08627 6837 /var/run...
Jan 15 12:09:11 f18ovn03runuser[6831]: pam_unix(runuser:session): session closed for user vdsm Jan 15 12:09:11 f18ovn03 systemd-vdsmd[6566]: [27B blob data] Jan 15 12:09:11 f18ovn03 systemd[1]: Started Virtual Desktop Server Manager. Jan 15 12:09:11 f18ovn03 python[6837]: DIGEST-MD5 client step 2 Jan 15 12:09:11 f18ovn03 python[6837]: DIGEST-MD5 client step 2 Jan 15 12:09:11 f18ovn03 python[6837]: DIGEST-MD5 client step 3 Jan 15 12:09:11 f18ovn03 vdsm[6837]: vdsm fileUtils WARNING Dir /rhev/data-center/mnt already exists Jan 15 12:09:14 f18ovn03 vdsm[6837]: vdsm Storage.LVM WARNING lvm pvs failed: 5 [] [' Skipping clustered vo...T01'] Jan 15 12:09:14 f18ovn03 vdsm[6837]: vdsm Storage.LVM WARNING lvm vgs failed: 5 [] [' Skipping clustered vo...T01'] Jan 15 12:09:14 f18ovn03 vdsm[6837]: vdsm vds WARNING Unable to load the json rpc server module. Please make...lled.
Is it normal the message regarding WARNING Dir /rhev/data-center/mnt already exists
yeah, just another log noise, I suppose :-(
?
Now # vdsClient -s 0 getVdsCaps HBAInventory = {'iSCSI': [{'InitiatorName': 'iqn.1994-05.com.redhat:3563e5612db4'}], 'FC': [{'wwpn': '50014380011bf958', 'wwnn': '50014380011bf959', 'model': 'QMH2462 - PCI-Express Dual Channel 4Gb Fibre Channel Mezzanine HBA'}, {'wwpn': '50014380011bf95a', 'wwnn': '50014380011bf95b', 'model': 'QMH2462 - PCI-Express Dual Channel 4Gb Fibre Channel Mezzanine HBA'}]} ISCSIInitiatorName = iqn.1994-05.com.redhat:3563e5612db4 bondings = {'bond4': {'addr': '', 'cfg': {}, 'mtu': '1500', 'netmask': '', 'slaves': [], 'hwaddr': '00:00:00:00:00:00'}, 'bond0': {'addr': '', 'cfg': {}, 'mtu': '1500', 'netmask': '', 'slaves': [], 'hwaddr': '00:00:00:00:00:00'}} bridges = {'ovirtmgmt': {'addr': '10.4.4.59', 'cfg': {'IPV6INIT': 'no', 'IPADDR': '10.4.4.59', 'ONBOOT': 'yes', 'DELAY': '0', 'NM_CONTROLLED': 'no', 'NETMASK': '255.255.255.0', 'BOOTPROTO': 'none', 'DEVICE': 'ovirtmgmt', 'TYPE': 'Bridge', 'GATEWAY': '10.4.4.250'}, 'mtu': '1500', 'netmask': '255.255.255.0', 'stp': 'off', 'ports': ['em3']}} clusterLevels = ['3.0', '3.1', '3.2'] cpuCores = 8 cpuFlags = fpu,vme,de,pse,tsc,msr,pae,mce,cx8,apic,sep,mtrr,pge,mca,cmov,pat,pse36,clflush,mmx,fxsr,sse,sse2,ht,syscall,nx,mmxext,fxsr_opt,rdtscp,lm,3dnowext,3dnow,rep_good,nopl,extd_apicid,pni,cx16,lahf_lm,cmp_legacy,svm,extapic,cr8_legacy,model_athlon,model_Opteron_G1,model_Opteron_G2 cpuModel = Dual-Core AMD Opteron(tm) Processor 8222 cpuSockets = 4 cpuSpeed = 3013.706 cpuThreads = 8 emulatedMachines = ['pc-1.2', 'none', 'pc', 'pc-1.1', 'pc-1.0', 'pc-0.15', 'pc-0.14', 'pc-0.13', 'pc-0.12', 'pc-0.11', 'pc-0.10', 'isapc'] guestOverhead = 65 hooks = {} kvmEnabled = true lastClient = 10.4.4.60 lastClientIface = ovirtmgmt management_ip = memSize = 32176 netConfigDirty = False networks = {'ovirtmgmt': {'iface': 'ovirtmgmt', 'addr': '10.4.4.59', 'cfg': {'IPV6INIT': 'no', 'IPADDR': '10.4.4.59', 'ONBOOT': 'yes', 'DELAY': '0', 'NM_CONTROLLED': 'no', 'NETMASK': '255.255.255.0', 'BOOTPROTO': 'none', 'DEVICE': 'ovirtmgmt', 'TYPE': 'Bridge', 'GATEWAY': '10.4.4.250'}, 'mtu': '1500', 'netmask': '255.255.255.0', 'stp': 'off', 'bridged': True, 'gateway': '10.4.4.250', 'ports': ['em3']}} nics = {'em4': {'addr': '', 'cfg': {'DEVICE': 'em4', 'HWADDR': '00:1c:c4:ab:3a:de', 'ONBOOT': 'yes', 'NM_CONTROLLED': 'no'}, 'mtu': '1500', 'netmask': '', 'hwaddr': '00:1c:c4:ab:3a:de', 'speed': 1000}, 'em1': {'addr': '', 'cfg': {'DEVICE': 'em1', 'NM_CONTROLLED': 'no', 'TYPE': 'Ethernet', 'ONBOOT': 'yes', 'HWADDR': '00:1E:0B:21:B8:C4'}, 'mtu': '1500', 'netmask': '', 'hwaddr': '00:1e:0b:21:b8:c4', 'speed': 1000}, 'em3': {'addr': '', 'cfg': {'BRIDGE': 'ovirtmgmt', 'IPV6INIT': 'no', 'NM_CONTROLLED': 'no', 'DEVICE': 'em3', 'HWADDR': '00:1c:c4:ab:3a:dd', 'ONBOOT': 'yes'}, 'mtu': '1500', 'netmask': '', 'hwaddr': '00:1c:c4:ab:3a:dd', 'speed': 1000}, 'em2': {'addr': '', 'cfg': {'DEVICE': 'em2', 'NM_CONTROLLED': 'no', 'TYPE': 'Ethernet', 'ONBOOT': 'yes', 'HWADDR': '00:1E:0B:21:B8:C6'}, 'mtu': '1500', 'netmask': '', 'hwaddr': '00:1e:0b:21:b8:c6', 'speed': 1000}} operatingSystem = {'release': '1', 'version': '18', 'name': 'Fedora'} packages2 = {'kernel': {'release': '3.fc18.x86_64', 'buildtime': 1355776539.0, 'version': '3.6.11'}, 'spice-server': {'release': '1.fc18', 'buildtime': 1356035501, 'version': '0.12.2'}, 'vdsm': {'release': '0.78.gitb005b54.fc18', 'buildtime': 1358090637, 'version': '4.10.3'}, 'qemu-kvm': {'release': '1.fc18', 'buildtime': 1355702442, 'version': '1.2.2'}, 'libvirt': {'release': '3.fc18', 'buildtime': 1355788803, 'version': '0.10.2.2'}, 'qemu-img': {'release': '1.fc18', 'buildtime': 1355702442, 'version': '1.2.2'}, 'mom': {'release': '1.fc18', 'buildtime': 1349470214, 'version': '0.3.0'}} reservedMem = 321 software_revision = 0.78 software_version = 4.10 supportedENGINEs = ['3.0', '3.1'] supportedProtocols = ['2.2', '2.3'] uuid = 34353439-3036-435A-4A38-303330393338 version_name = Snow Man vlans = {} vmTypes = ['kvm']
and from webadmin I see the host up...
Is it repeatable that vdsm is down after boot? Dan.