Please package the logs (tar or zip) and send them.
On Mon, Feb 6, 2017 at 12:05 PM, Shalabh Goel <shalabhgoel13(a)gmail.com>
wrote:
Yes, I am using OVS as the switch type and I did not know that it was
not
supported officially.
The output of ovs-vsctl show is as follows:
f634d53e-4849-488b-8454-6b1fafa7c6ac
ovs_version: "2.6.90"
I am attaching OVS switch logs below:
/var/log/openvswitch/ovsdb-server.log
2017-02-06T09:46:07.788Z|00001|vlog|INFO|opened log file
/var/log/openvswitch/ovsdb-server.log
2017-02-06T09:46:07.791Z|00002|ovsdb_server|INFO|ovsdb-server (Open
vSwitch) 2.6.90
2017-02-06T09:46:17.802Z|00003|memory|INFO|2296 kB peak resident set size
after 10.0 seconds
2017-02-06T09:46:17.802Z|00004|memory|INFO|cells:16 json-caches:1
monitors:1 sessions:1
ovs-vswitchd.log
2017-02-06T09:46:07.999Z|00001|vlog|INFO|opened log file
/var/log/openvswitch/ovs-vswitchd.log
2017-02-06T09:46:08.036Z|00002|ovs_numa|INFO|Discovered 24 CPU cores on
NUMA node 0
2017-02-06T09:46:08.036Z|00003|ovs_numa|INFO|Discovered 24 CPU cores on
NUMA node 1
2017-02-06T09:46:08.036Z|00004|ovs_numa|INFO|Discovered 2 NUMA nodes and
48 CPU cores
2017-02-06T09:46:08.037Z|00005|reconnect|INFO|unix:/var/run/openvswitch/db.sock:
connecting...
2017-02-06T09:46:08.037Z|00006|reconnect|INFO|unix:/var/run/openvswitch/db.sock:
connected
2017-02-06T09:46:08.039Z|00007|bridge|INFO|ovs-vswitchd (Open vSwitch)
2.6.90
What should I do now?
The engine says that "Host host2 does not comply with the cluster Default
networks, the following networks are missing on host: 'ovirtmgmt'"
What other logs should I attach?
Thanks
Shalabh Goel
On Sun, Feb 5, 2017 at 1:10 PM, Edward Haas <ehaas(a)redhat.com> wrote:
> Based on what I can see, you used OVS as the switch type and it seems ovs
> (openvswitch) is not properly installed on your host.
> Make sure that you have ovs operational by issuing "ovs-vsctl show".
>
> You should note that OVS network support is not an official release
> feature, and you should use it on 4.1 and up versions.
> Fixes will be probably submitted to master (appearing in nightly builds).
>
> Next time please include the mailing-list in your replies and attach the
> log files, it is less spamming.
>
> Thanks,
> Edy.
>
> On Fri, Feb 3, 2017 at 5:07 AM, Shalabh Goel <shalabhgoel13(a)gmail.com>
> wrote:
>
>> log from messages
>>
>> Feb 3 08:27:53 ovirtnode3 ovs-vsctl:
ovs|00001|db_ctl_base|ERR|unix:/var/run/openvswitch/db.sock:
>> database connection failed (No such file or directory)
>> Feb 3 08:27:53 ovirtnode3 journal: vdsm vds ERROR Executing commands
>> failed: ovs-vsctl: unix:/var/run/openvswitch/db.sock: database
>> connection failed (No su
>> ch file or directory)#012Traceback (most recent call last):#012 File
>> "/usr/share/vdsm/API.py", line 1531, in setupNetworks#012
>> supervdsm.getProxy().setup
>> Networks(networks, bondings, options)#012 File
>> "/usr/lib/python2.7/site-packages/vdsm/supervdsm.py", line 53, in
>> __call__#012 return callMethod()#012 Fi
>> le "/usr/lib/python2.7/site-packages/vdsm/supervdsm.py", line 51, in
>> <lambda>#012 **kwargs)#012 File "<string>", line 2, in
>> setupNetworks#012 File "/usr
>> /lib64/python2.7/multiprocessing/managers.py", line 773, in
>> _callmethod#012 raise convert_to_error(kind,
>> result)#012ConfigNetworkError: (21, 'Executing co
>> mmands failed: ovs-vsctl: unix:/var/run/openvswitch/db.sock: database
>> connection failed (No such file or directory)')
>>
>> Log from vdsm.log
>>
>> Reactor thread::INFO::2017-02-03 08:32:24,638::protocoldetector
>> ::72::ProtocolDetector.AcceptorImpl::(handle_accept) Accepted
>> connection from ::1:53082
>> Reactor thread::DEBUG::2017-02-03 08:32:24,643::protocoldetector
>> ::91::ProtocolDetector.Detector::(__init__) Using required_size=11
>> Reactor thread::INFO::2017-02-03 08:32:24,643::protocoldetector
>> ::127::ProtocolDetector.Detector::(handle_read) Detected protocol xml
>> from ::1:53082
>> Reactor thread::DEBUG::2017-02-03 08:32:24,643::bindingxmlrpc::1
>> 317::XmlDetector::(handle_socket) xml over http detected from ('::1',
>> 53082)
>> BindingXMLRPC::INFO::2017-02-03
08:32:24,643::xmlrpc::73::vds.XMLRPCServer::(handle_request)
>> Starting request handler for ::1:53082
>> Thread-20::INFO::2017-02-03 08:32:24,644::xmlrpc::83::vds.
>> XMLRPCServer::(_process_requests) Request handler for ::1:53082 started
>> Thread-20::DEBUG::2017-02-03 08:32:24,644::bindingxmlrpc::1263::vds::(wrapper)
>> client [::1]::call getAllVmStats with () {}
>> Thread-20::DEBUG::2017-02-03 08:32:24,644::bindingxmlrpc::1270::vds::(wrapper)
>> return getAllVmStats with {'status': {'message': 'Done',
'code': 0},
>> 'statsLis
>> t': (suppressed)}
>> Thread-20::INFO::2017-02-03 08:32:24,644::bindingxmlrpc::1297::vds::(wrapper)
>> RPC call getAllVmStats finished (code=0) in 0.00 seconds
>> Thread-20::INFO::2017-02-03 08:32:24,645::xmlrpc::91::vds.
>> XMLRPCServer::(_process_requests) Request handler for ::1:53082 stopped
>> periodic/0::DEBUG::2017-02-03 08:32:24,907::sampling::508::v
>> irt.sampling.VMBulkSampler::(__call__) sampled timestamp 4347942.15
>> elapsed 0.010 acquired True d
>> omains all
>> periodic/2::DEBUG::2017-02-03 08:32:24,908::task::599::stora
>> ge.TaskManager.Task::(_updateState)
(Task='bb07f89c-6165-42bc-902f-1b7701b813fa')
>> moving from sta
>> te init -> state preparing
>> periodic/2::INFO::2017-02-03 08:32:24,908::logUtils::49::dispatcher::(wrapper)
>> Run and protect: repoStats(options=None)
>> periodic/2::INFO::2017-02-03 08:32:24,908::logUtils::52::dispatcher::(wrapper)
>> Run and protect: repoStats, Return response: {}
>> periodic/2::DEBUG::2017-02-03 08:32:24,909::task::1195::stor
>> age.TaskManager.Task::(prepare)
(Task='bb07f89c-6165-42bc-902f-1b7701b813fa')
>> finished: {}
>> periodic/2::DEBUG::2017-02-03 08:32:24,909::task::599::stora
>> ge.TaskManager.Task::(_updateState)
(Task='bb07f89c-6165-42bc-902f-1b7701b813fa')
>> moving from sta
>> te preparing -> state finished
>> periodic/2::DEBUG::2017-02-03 08:32:24,909::resourceManager:
>> :910::storage.ResourceManager.Owner::(releaseAll) Owner.releaseAll
>> requests {} resources {}
>> periodic/2::DEBUG::2017-02-03 08:32:24,909::resourceManager:
>> :947::storage.ResourceManager.Owner::(cancelAll) Owner.cancelAll
>> requests {}
>> periodic/2::DEBUG::2017-02-03
08:32:24,909::task::997::storage.TaskManager.Task::(_decref)
>> (Task='bb07f89c-6165-42bc-902f-1b7701b813fa') ref 0 aborting False
>> jsonrpc/4::DEBUG::2017-02-03 08:32:26,653::__init__::532::j
>> sonrpc.JsonRpcServer::(_handle_request) Calling 'Host.getAllVmStats' in
>> bridge with {}
>> jsonrpc/4::DEBUG::2017-02-03 08:32:26,653::__init__::557::j
>> sonrpc.JsonRpcServer::(_handle_request) Return 'Host.getAllVmStats' in
>> bridge with (suppressed)
>> jsonrpc/4::INFO::2017-02-03 08:32:26,653::__init__::515::j
>> sonrpc.JsonRpcServer::(_serveRequest) RPC call Host.getAllVmStats
>> succeeded in 0.00 seconds
>> jsonrpc/6::DEBUG::2017-02-03 08:32:27,706::__init__::532::j
>> sonrpc.JsonRpcServer::(_handle_request) Calling 'Host.getStats' in
>> bridge with {}
>> jsonrpc/6::DEBUG::2017-02-03 08:32:27,706::__init__::532::j
>> sonrpc.JsonRpcServer::(_handle_request) Calling 'Host.getStats' in
>> bridge with {}
>> jsonrpc/6::DEBUG::2017-02-03 08:32:27,707::task::599::stora
>> ge.TaskManager.Task::(_updateState)
(Task='9285329e-485b-4165-8bdc-4f411f394356')
>> moving from state init -> state preparing
>> jsonrpc/6::INFO::2017-02-03 08:32:27,707::logUtils::49::dispatcher::(wrapper)
>> Run and protect: repoStats(options=None)
>> jsonrpc/6::INFO::2017-02-03 08:32:27,707::logUtils::52::dispatcher::(wrapper)
>> Run and protect: repoStats, Return response: {}
>> jsonrpc/6::DEBUG::2017-02-03 08:32:27,707::task::1195::stor
>> age.TaskManager.Task::(prepare)
(Task='9285329e-485b-4165-8bdc-4f411f394356')
>> finished: {}
>> jsonrpc/6::DEBUG::2017-02-03 08:32:27,707::task::599::stora
>> ge.TaskManager.Task::(_updateState)
(Task='9285329e-485b-4165-8bdc-4f411f394356')
>> moving from state preparing -> state finished
>> jsonrpc/6::DEBUG::2017-02-03 08:32:27,708::resourceManager:
>> :910::storage.ResourceManager.Owner::(releaseAll) Owner.releaseAll
>> requests {} resources {}
>> jsonrpc/6::DEBUG::2017-02-03 08:32:27,708::resourceManager:
>> :947::storage.ResourceManager.Owner::(cancelAll) Owner.cancelAll
>> requests {}
>> jsonrpc/6::DEBUG::2017-02-03
08:32:27,708::task::997::storage.TaskManager.Task::(_decref)
>> (Task='9285329e-485b-4165-8bdc-4f411f394356') ref 0 aborting False
>> jsonrpc/6::DEBUG::2017-02-03 08:32:27,712::__init__::557::j
>> sonrpc.JsonRpcServer::(_handle_request) Return 'Host.getStats' in
>> bridge with {'cpuStatistics': {'28': {'cpuUser':
'0.13', 'nodeIndex': 0,
>> 'cpuSys': '0.00', 'cpuIdle': '99.87'},
'29': {'cpuUser': '0.00',
>> 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle':
'100.00'}, '24': {'cpuUser':
>> '0.00', 'nodeIndex': 0, 'cpuSys': '0.00',
'cpuIdle': '100.00'}, '25':
>> {'cpuUser': '0.13', 'nodeIndex': 0, 'cpuSys':
'0.07', 'cpuIdle': '99.80'},
>> '26': {'cpuUser': '0.00', 'nodeIndex': 0,
'cpuSys': '0.00', 'cpuIdle':
>> '100.00'}, '27': {'cpuUser': '0.00',
'nodeIndex': 0, 'cpuSys': '0.00',
>> 'cpuIdle': '100.00'}, '20': {'cpuUser':
'0.00', 'nodeIndex': 1, 'cpuSys':
>> '0.00', 'cpuIdle': '100.00'}, '21':
{'cpuUser': '0.00', 'nodeIndex': 1,
>> 'cpuSys': '0.00', 'cpuIdle': '100.00'},
'22': {'cpuUser': '0.00',
>> 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle':
'100.00'}, '23': {'cpuUser':
>> '0.00', 'nodeIndex': 1, 'cpuSys': '0.00',
'cpuIdle': '100.00'}, '46':
>> {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys':
'0.00', 'cpuIdle': '100.00'},
>> '47': {'cpuUser': '0.00', 'nodeIndex': 1,
'cpuSys': '0.00', 'cpuIdle':
>> '100.00'}, '44': {'cpuUser': '0.00',
'nodeIndex': 1, 'cpuSys': '0.00',
>> 'cpuIdle': '100.00'}, '45': {'cpuUser':
'0.00', 'nodeIndex': 1, 'cpuSys':
>> '0.00', 'cpuIdle': '100.00'}, '42':
{'cpuUser': '0.00', 'nodeIndex': 1,
>> 'cpuSys': '0.00', 'cpuIdle': '100.00'},
'43': {'cpuUser': '0.00',
>> 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle':
'100.00'}, '40': {'cpuUser':
>> '0.00', 'nodeIndex': 1, 'cpuSys': '0.00',
'cpuIdle': '100.00'}, '41':
>> {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys':
'0.00', 'cpuIdle': '100.00'},
>> '1': {'cpuUser': '0.13', 'nodeIndex': 0,
'cpuSys': '0.07', 'cpuIdle':
>> '99.80'}, '0': {'cpuUser': '0.00',
'nodeIndex': 0, 'cpuSys': '0.00',
>> 'cpuIdle': '100.00'}, '3': {'cpuUser':
'0.07', 'nodeIndex': 0, 'cpuSys':
>> '0.07', 'cpuIdle': '99.86'}, '2':
{'cpuUser': '0.07', 'nodeIndex': 0,
>> 'cpuSys': '0.00', 'cpuIdle': '99.93'},
'5': {'cpuUser': '0.00',
>> 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle':
'100.00'}, '4': {'cpuUser':
>> '0.00', 'nodeIndex': 0, 'cpuSys': '0.07',
'cpuIdle': '99.93'}, '7':
>> {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys':
'0.00', 'cpuIdle': '100.00'},
>> '6': {'cpuUser': '0.20', 'nodeIndex': 0,
'cpuSys': '0.00', 'cpuIdle':
>> '99.80'}, '9': {'cpuUser': '0.00',
'nodeIndex': 0, 'cpuSys': '0.00',
>> 'cpuIdle': '100.00'}, '8': {'cpuUser':
'0.00', 'nodeIndex': 0, 'cpuSys':
>> '0.00', 'cpuIdle': '100.00'}, '39':
{'cpuUser': '0.00', 'nodeIndex': 1,
>> 'cpuSys': '0.00', 'cpuIdle': '100.00'},
'38': {'cpuUser': '0.00',
>> 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle':
'100.00'}, '11': {'cpuUser':
>> '0.00', 'nodeIndex': 0, 'cpuSys': '0.00',
'cpuIdle': '100.00'}, '10':
>> {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys':
'0.00', 'cpuIdle': '100.00'},
>> '13': {'cpuUser': '0.00', 'nodeIndex': 1,
'cpuSys': '0.00', 'cpuIdle':
>> '100.00'}, '12': {'cpuUser': '0.00',
'nodeIndex': 1, 'cpuSys': '0.00',
>> 'cpuIdle': '100.00'}, '15': {'cpuUser':
'0.00', 'nodeIndex': 1, 'cpuSys':
>> '0.00', 'cpuIdle': '100.00'}, '14':
{'cpuUser': '0.00', 'nodeIndex': 1,
>> 'cpuSys': '0.00', 'cpuIdle': '100.00'},
'17': {'cpuUser': '0.00',
>> 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle':
'100.00'}, '16': {'cpuUser':
>> '0.00', 'nodeIndex': 1, 'cpuSys': '0.00',
'cpuIdle': '100.00'}, '19':
>> {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys':
'0.00', 'cpuIdle': '100.00'},
>> '18': {'cpuUser': '0.00', 'nodeIndex': 1,
'cpuSys': '0.00', 'cpuIdle':
>> '100.00'}, '31': {'cpuUser': '0.00',
'nodeIndex': 0, 'cpuSys': '0.00',
>> 'cpuIdle': '100.00'}, '30': {'cpuUser':
'0.00', 'nodeIndex': 0, 'cpuSys':
>> '0.00', 'cpuIdle': '100.00'}, '37':
{'cpuUser': '0.00', 'nodeIndex': 1,
>> 'cpuSys': '0.00', 'cpuIdle': '100.00'},
'36': {'cpuUser': '0.00',
>> 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle':
'100.00'}, '35': {'cpuUser':
>> '0.00', 'nodeIndex': 0, 'cpuSys': '0.00',
'cpuIdle': '100.00'}, '34':
>> {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys':
'0.00', 'cpuIdle': '100.00'},
>> '33': {'cpuUser': '0.00', 'nodeIndex': 0,
'cpuSys': '0.00', 'cpuIdle':
>> '100.00'}, '32': {'cpuUser': '0.00',
'nodeIndex': 0, 'cpuSys': '0.00',
>> 'cpuIdle': '100.00'}}, 'numaNodeMemFree': {'1':
{'memPercent': 3,
>> 'memFree': '63780'}, '0': {'memPercent': 4,
'memFree': '63426'}},
>> 'memShared': 0, 'thpState': 'always',
'ksmMergeAcrossNodes': True,
>> 'vmCount': 0, 'memUsed': '2', 'storageDomains':
{}, 'incomingVmMigrations':
>> 0, 'network': {'ib0': {'txErrors': '0',
'state': 'down', 'sampleTime':
>> 1486090944.903051, 'name': 'ib0', 'tx': '0',
'txDropped': '0', 'rx': '0',
>> 'rxErrors': '0', 'speed': '1000',
'rxDropped': '0'}, 'lo': {'txErrors':
>> '0', 'state': 'up', 'sampleTime':
1486090944.903051, 'name': 'lo', 'tx':
>> '56174388', 'txDropped': '0', 'rx':
'56174388', 'rxErrors': '0', 'speed':
>> '1000', 'rxDropped': '0'}, 'eno2':
{'txErrors': '0', 'state': 'down',
>> 'sampleTime': 1486090944.903051, 'name': 'eno2',
'tx': '0', 'txDropped':
>> '0', 'rx': '0', 'rxErrors': '0',
'speed': '1000', 'rxDropped': '0'},
>> 'eno1': {'txErrors': '0', 'state': 'up',
'sampleTime': 1486090944.903051,
>> 'name': 'eno1', 'tx': '26032823',
'txDropped': '0', 'rx': '174722400',
>> 'rxErrors': '0', 'speed': '1000',
'rxDropped': '5624'}, ';vdsmdummy;':
>> {'txErrors': '0', 'state': 'down',
'sampleTime': 1486090944.903051, 'name':
>> ';vdsmdummy;', 'tx': '0', 'txDropped':
'0', 'rx': '0', 'rxErrors': '0',
>> 'speed': '1000', 'rxDropped': '0'}},
'txDropped': '0', 'anonHugePages':
>> '34', 'ksmPages': 100, 'elapsedTime': '287.81',
'cpuLoad': '0.07',
>> 'cpuSys': '0.00', 'diskStats': {'/var/log':
{'free': '15072'},
>> '/var/log/core': {'free': '15072'},
'/var/run/vdsm/': {'free': '64418'},
>> '/tmp': {'free': '914756'}}, 'cpuUserVdsmd':
'0.13', 'netConfigDirty':
>> 'True', 'memCommitted': 0, 'ksmState': False,
'vmMigrating': 0, 'ksmCpu':
>> 0, 'memAvailable': 127200, 'bootTime': '1486037670',
'haStats': {'active':
>> False, 'configured': False, 'score': 0,
'localMaintenance': False,
>> 'globalMaintenance': False}, 'momStatus': 'active',
'rxDropped': '5624',
>> 'outgoingVmMigrations': 0, 'swapTotal': 4095, 'swapFree':
4095, 'dateTime':
>> '2017-02-03T03:02:27 GMT', 'cpuUser': '0.02',
'memFree': 127456, 'cpuIdle':
>> '99.98', 'vmActive': 0, 'v2vJobs': {},
'cpuSysVdsmd': '0.07'}
>> jsonrpc/6::INFO::2017-02-03 08:32:27,713::__init__::515::j
>> sonrpc.JsonRpcServer::(_serveRequest) RPC call Host.getStats succeeded
>> in 0.00 seconds
>> Reactor thread::INFO::2017-02-03 08:32:39,661::protocoldetector
>> ::72::ProtocolDetector.AcceptorImpl::(handle_accept) Accepted
>> connection from ::1:53084
>>
>> supervdsm.log
>>
>>
>> File "/usr/share/vdsm/supervdsmServer", line 95, in wrapper
>> res = func(*args, **kwargs)
>> File "/usr/share/vdsm/supervdsmServer", line 143, in umount
>> timeout=timeout)
>> File "/usr/lib/python2.7/site-packages/vdsm/storage/mount.py", line
>> 281, in _umount
>> _runcmd(cmd, timeout)
>> File "/usr/lib/python2.7/site-packages/vdsm/storage/mount.py", line
>> 297, in _runcmd
>> raise MountError(rc, ";".join((out, err)))
>> MountError: (32, ';umount: /rhev/data-center/mnt/10.1.235.6:_iso:
>> mountpoint not found\n')
>> MainProcess|jsonrpc/5::DEBUG::2017-02-03 08:26:59,495::supervdsmServer:
>> :93::SuperVdsm.ServerCallback::(wrapper) call hbaRescan with () {}
>> MainProcess|jsonrpc/5::DEBUG::2017-02-03
08:26:59,495::commands::69::storage.HBA::(execCmd)
>> /usr/bin/taskset --cpu-list 0-47 /usr/libexec/vdsm/fc-scan (cwd N
>> one)
>> MainProcess|jsonrpc/5::DEBUG::2017-02-03 08:26:59,579::supervdsmServer:
>> :100::SuperVdsm.ServerCallback::(wrapper) return hbaRescan with None
>> MainThread::DEBUG::2017-02-03
08:27:26,367::supervdsmServer::311::SuperVdsm.Server::(main)
>> Terminated normally
>> MainThread::DEBUG::2017-02-03
08:27:36,148::__init__::47::blivet::(register_device_format)
>> registered device format class DeviceFormat as None
>> MainThread::DEBUG::2017-02-03
08:27:36,149::__init__::47::blivet::(register_device_format)
>> registered device format class BIOSBoot as biosboot
>> MainThread::DEBUG::2017-02-03
08:27:36,150::__init__::47::blivet::(register_device_format)
>> registered device format class DiskLabel as disklabel
>> MainThread::DEBUG::2017-02-03
08:27:36,150::__init__::47::blivet::(register_device_format)
>> registered device format class DMRaidMember as dmraidmember
>> MainThread::DEBUG::2017-02-03
08:27:36,151::__init__::47::blivet::(register_device_format)
>> registered device format class Ext2FS as ext2
>> MainThread::DEBUG::2017-02-03
08:27:36,152::__init__::47::blivet::(register_device_format)
>> registered device format class Ext3FS as ext3
>> MainThread::DEBUG::2017-02-03
08:27:36,152::__init__::47::blivet::(register_device_format)
>> registered device format class Ext4FS as ext4
>> MainThread::DEBUG::2017-02-03
08:27:36,152::__init__::47::blivet::(register_device_format)
>> registered device format class FATFS as vfat
>> MainThread::DEBUG::2017-02-03
08:27:36,152::__init__::47::blivet::(register_device_format)
>> registered device format class EFIFS as efi
>> MainThread::DEBUG::2017-02-03
08:27:36,152::__init__::47::blivet::(register_device_format)
>> registered device format class BTRFS as btrfs
>> MainThread::DEBUG::2017-02-03
08:27:36,153::__init__::47::blivet::(register_device_format)
>> registered device format class GFS2 as gfs2
>> MainThread::DEBUG::2017-02-03
08:27:36,153::__init__::47::blivet::(register_device_format)
>> registered device format class JFS as jfs
>> MainThread::DEBUG::2017-02-03
08:27:36,153::__init__::47::blivet::(register_device_format)
>> registered device format class ReiserFS as reiserfs
>> MainThread::DEBUG::2017-02-03
08:27:36,153::__init__::47::blivet::(register_device_format)
>> registered device format class XFS as xfs
>> MainThread::DEBUG::2017-02-03
08:27:36,153::__init__::47::blivet::(register_device_format)
>> registered device format class HFS as hfs
>> MainThread::DEBUG::2017-02-03
08:27:36,153::__init__::47::blivet::(register_device_format)
>> registered device format class AppleBootstrapFS as appleboot
>> MainThread::DEBUG::2017-02-03
08:27:36,154::__init__::47::blivet::(register_device_format)
>> registered device format class HFSPlus as hfs+
>> MainThread::DEBUG::2017-02-03
08:27:36,154::__init__::47::blivet::(register_device_format)
>> registered device format class MacEFIFS as macefi
>> MainThread::DEBUG::2017-02-03
08:27:36,154::__init__::47::blivet::(register_device_format)
>> registered device format class NTFS as ntfs
>> MainThread::DEBUG::2017-02-03
08:27:36,154::__init__::47::blivet::(register_device_format)
>> registered device format class NFS as nfs
>> MainThread::DEBUG::2017-02-03
08:27:36,154::__init__::47::blivet::(register_device_format)
>> registered device format class NFSv4 as nfs4
>> MainThread::DEBUG::2017-02-03
08:27:36,154::__init__::47::blivet::(register_device_format)
>> registered device format class Iso9660FS as iso9660
>> MainThread::DEBUG::2017-02-03
08:27:36,154::__init__::47::blivet::(register_device_format)
>> registered device format class NoDevFS as nodev
>> MainThread::DEBUG::2017-02-03
08:27:36,155::__init__::47::blivet::(register_device_format)
>> registered device format class DevPtsFS as devpts
>> MainThread::DEBUG::2017-02-03
08:27:36,155::__init__::47::blivet::(register_device_format)
>> registered device format class ProcFS as proc
>> MainThread::DEBUG::2017-02-03
08:27:36,155::__init__::47::blivet::(register_device_format)
>> registered device format class SysFS as sysfs
>> MainThread::DEBUG::2017-02-03
08:27:36,155::__init__::47::blivet::(register_device_format)
>> registered device format class TmpFS as tmpfs
>> MainThread::DEBUG::2017-02-03
08:27:36,155::__init__::47::blivet::(register_device_format)
>> registered device format class BindFS as bind
>> MainThread::DEBUG::2017-02-03
08:27:36,155::__init__::47::blivet::(register_device_format)
>> registered device format class SELinuxFS as selinuxfs
>> MainThread::DEBUG::2017-02-03
08:27:36,155::__init__::47::blivet::(register_device_format)
>> registered device format class USBFS as usbfs
>> MainThread::DEBUG::2017-02-03
08:27:36,155::__init__::47::blivet::(register_device_format)
>> registered device format class EFIVarFS as efivarfs
>> MainThread::DEBUG::2017-02-03
08:27:36,179::__init__::47::blivet::(register_device_format)
>> registered device format class LUKS as luks
>> MainThread::DEBUG::2017-02-03
08:27:36,181::__init__::47::blivet::(register_device_format)
>> registered device format class LVMPhysicalVolume as lvmpv
>> MainThread::DEBUG::2017-02-03
08:27:36,181::__init__::47::blivet::(register_device_format)
>> registered device format class MDRaidMember as mdmember
>> MainThread::DEBUG::2017-02-03
08:27:36,181::__init__::47::blivet::(register_device_format)
>> registered device format class MultipathMember as multipath_member
>> MainThread::DEBUG::2017-02-03
08:27:36,181::__init__::47::blivet::(register_device_format)
>> registered device format class PPCPRePBoot as prepboot
>> MainThread::DEBUG::2017-02-03
08:27:36,182::__init__::47::blivet::(register_device_format)
>> registered device format class SwapSpace as swap
>> MainThread::DEBUG::2017-02-03
08:27:36,216::storage_log::69::blivet::(log_exception_info)
>> IGNORED: Caught exception, continuing.
>> MainThread::DEBUG::2017-02-03
08:27:36,216::storage_log::72::blivet::(log_exception_info)
>> IGNORED: Problem description: failed to get initiator name from
>> iscsi firmware
>> MainThread::DEBUG::2017-02-03
08:27:36,216::storage_log::73::blivet::(log_exception_info)
>> IGNORED: Begin exception details.
>> MainThread::DEBUG::2017-02-03
08:27:36,216::storage_log::76::blivet::(log_exception_info)
>> IGNORED: Traceback (most recent call last):
>> MainThread::DEBUG::2017-02-03
08:27:36,216::storage_log::76::blivet::(log_exception_info)
>> IGNORED: File "/usr/lib/python2.7/site-packages/blivet
>> /iscsi.py", line 146, in __init__
>> MainThread::DEBUG::2017-02-03
08:27:36,216::storage_log::76::blivet::(log_exception_info)
>> IGNORED: initiatorname = libiscsi.get_firmware_initi
>> ator_name()
>> MainThread::DEBUG::2017-02-03
08:27:36,216::storage_log::76::blivet::(log_exception_info)
>> IGNORED: IOError: Unknown error
>> MainThread::DEBUG::2017-02-03
08:27:36,216::storage_log::77::blivet::(log_exception_info)
>> IGNORED: End exception details.
>> MainThread::DEBUG::2017-02-03
08:27:36,227::supervdsmServer::271::SuperVdsm.Server::(main)
>> Making sure I'm root - SuperVdsm
>> MainThread::DEBUG::2017-02-03
08:27:36,227::supervdsmServer::280::SuperVdsm.Server::(main)
>> Parsing cmd args
>> MainThread::DEBUG::2017-02-03
08:27:36,227::supervdsmServer::283::SuperVdsm.Server::(main)
>> Cleaning old socket /var/run/vdsm/svdsm.sock
>> MainThread::DEBUG::2017-02-03
08:27:36,227::supervdsmServer::287::SuperVdsm.Server::(main)
>> Setting up keep alive thread
>> MainThread::DEBUG::2017-02-03
08:27:36,228::supervdsmServer::293::SuperVdsm.Server::(main)
>> Creating remote object manager
>> MainThread::DEBUG::2017-02-03
08:27:36,228::fileUtils::246::storage.fileUtils::(chown)
>> Changing owner for /var/run/vdsm/svdsm.sock, to (36:36)
>> MainThread::DEBUG::2017-02-03
08:27:36,228::supervdsmServer::303::SuperVdsm.Server::(main)
>> Started serving super vdsm object
>> sourceRoute::DEBUG::2017-02-03 08:27:36,229::sourceroutethrea
>> d::79::root::(_subscribeToInotifyLoop) sourceRouteThread.subscribeToInotifyLoop
>> started
>> restore-net::INFO::2017-02-03
08:27:37,160::vdsm-restore-net-config::465::root::(restore)
>> networks already restored. doing nothing.
>> MainProcess|hsm/init::DEBUG::2017-02-03 08:27:39,601::supervdsmServer:
>> :93::SuperVdsm.ServerCallback::(wrapper) call hbaRescan with () {}
>> MainProcess|hsm/init::DEBUG::2017-02-03
08:27:39,602::commands::69::storage.HBA::(execCmd)
>> /usr/bin/taskset --cpu-list 0-47 /usr/libexec/vdsm/fc-scan (cwd No
>> ne)
>> MainProcess|hsm/init::DEBUG::2017-02-03 08:27:39,691::supervdsmServer:
>> :100::SuperVdsm.ServerCallback::(wrapper) return hbaRescan with None
>> MainProcess|jsonrpc/4::DEBUG::2017-02-03 08:27:48,355::supervdsmServer:
>> :93::SuperVdsm.ServerCallback::(wrapper) call network_caps with () {}
>> MainProcess|jsonrpc/4::DEBUG::2017-02-03
08:27:48,360::commands::69::root::(execCmd)
>> /usr/bin/taskset --cpu-list 0-47 /sbin/ip route show to 0.0.0.0/0 table
>> main (cwd None)
>> MainProcess|jsonrpc/4::DEBUG::2017-02-03
08:27:48,364::commands::93::root::(execCmd)
>> SUCCESS: <err> = ''; <rc> = 0
>> MainProcess|jsonrpc/4::DEBUG::2017-02-03
08:27:48,408::libvirtconnection::160::root::(get)
>> trying to connect libvirt
>> MainProcess|jsonrpc/4::DEBUG::2017-02-03
08:27:48,450::commands::69::root::(execCmd)
>> /usr/bin/taskset --cpu-list 0-47 /usr/sbin/tc qdisc show (cwd None)
>> MainProcess|jsonrpc/4::DEBUG::2017-02-03
08:27:48,453::commands::93::root::(execCmd)
>> SUCCESS: <err> = ''; <rc> = 0
>> MainProcess|jsonrpc/4::DEBUG::2017-02-03
08:27:48,454::commands::69::root::(execCmd)
>> /usr/bin/taskset --cpu-list 0-47 /bin/systemctl --no-pager list-unit-fil
>> es (cwd None)
>> MainProcess|jsonrpc/4::DEBUG::2017-02-03
08:27:48,530::commands::93::root::(execCmd)
>> SUCCESS: <err> = ''; <rc> = 0
>> MainProcess|jsonrpc/4::DEBUG::2017-02-03
08:27:48,530::commands::69::root::(execCmd)
>> /usr/bin/taskset --cpu-list 0-47 /bin/systemctl status openvswitch.servi
>> ce (cwd None)
>> MainProcess|jsonrpc/4::DEBUG::2017-02-03
08:27:48,561::commands::93::root::(execCmd)
>> FAILED: <err> = ''; <rc> = 3
>> MainProcess|jsonrpc/4::DEBUG::2017-02-03 08:27:48,561::supervdsmServer:
>> :100::SuperVdsm.ServerCallback::(wrapper) return network_caps with
>> {'bridges': {}, 'bo
>> ndings': {}, 'nameservers': ['10.1.1.8'], 'nics':
{'ib0':
>> {'ipv6autoconf': True, 'addr': '', 'ipv6gateway':
'::', 'dhcpv6': False,
>> 'ipv6addrs': [], 'mtu': '4
>> 092', 'dhcpv4': False, 'netmask': '',
'ipv4defaultroute': False,
>> 'ipv4addrs': [], 'hwaddr': '80:00:02:08:fe:80:00:00:00:00
>> :00:00:0c:c4:7a:ff:ff:0f:d8:85', 's
>> peed': 0, 'gateway': ''}, 'eno1':
{'ipv6autoconf': False, 'addr':
>> '10.1.235.5', 'ipv6gateway': '::', 'dhcpv6':
False, 'ipv6addrs': [], 'mtu':
>> '1500', 'dhcpv4
>> ': False, 'netmask': '255.255.0.0',
'ipv4defaultroute': True,
>> 'ipv4addrs': ['10.1.235.5/16'], 'hwaddr':
'0c:c4:7a:0f:6b:2a', 'speed':
>> 1000, 'gateway': '10.1.
>> 0.1'}, 'eno2': {'ipv6autoconf': True, 'addr':
'', 'ipv6gateway': '::',
>> 'dhcpv6': False, 'ipv6addrs': [], 'mtu': '1500',
'dhcpv4': False,
>> 'netmask': '', 'ipv4
>> defaultroute': False, 'ipv4addrs': [], 'hwaddr':
'0c:c4:7a:0f:6b:2b',
>> 'speed': 0, 'gateway': ''}}, 'supportsIPv6':
True, 'vlans': {}, 'networks':
>> {}}
>> MainProcess|jsonrpc/5::DEBUG::2017-02-03 08:27:53,690::supervdsmServer:
>> :93::SuperVdsm.ServerCallback::(wrapper) call setupNetworks with ({},
>> {u'connectivityC
>> heck': u'true', u'connectivityTimeout': 120}) {}
>> MainProcess|jsonrpc/5::DEBUG::2017-02-03
08:27:53,691::api::247::root::(setupNetworks)
>> Setting up network according to configuration: networks:{u'ovirtmgmt':
>> {u'ipv6autoconf': False, u'nic': u'eno1',
u'ipaddr': u'10.1.235.5',
>> u'netmask': u'255.255.0.0', u'mtu': 1500,
u'switch': u'ovs', u'dhcpv6':
>> False, u'STP': u'no', u'bridged': u'true',
u'gateway': u'10.1.0.1',
>> u'defaultRoute': True}}, bondings:{},
options:{u'connectivityCheck':
>> u'true', u'connectivityTimeout': 120}
>> MainProcess|jsonrpc/5::DEBUG::2017-02-03
08:27:53,691::api::252::root::(setupNetworks)
>> Validating configuration
>> MainProcess|jsonrpc/5::DEBUG::2017-02-03
08:27:53,694::commands::69::root::(execCmd)
>> /usr/bin/taskset --cpu-list 0-47 /sbin/ip route show to 0.0.0.0/0 table
>> main (cwd None)
>> MainProcess|jsonrpc/5::DEBUG::2017-02-03
08:27:53,698::commands::93::root::(execCmd)
>> SUCCESS: <err> = ''; <rc> = 0
>> MainProcess|jsonrpc/5::DEBUG::2017-02-03
08:27:53,771::commands::69::root::(execCmd)
>> /usr/bin/taskset --cpu-list 0-47 /usr/sbin/tc qdisc show (cwd None)
>> MainProcess|jsonrpc/5::DEBUG::2017-02-03
08:27:53,774::commands::93::root::(execCmd)
>> SUCCESS: <err> = ''; <rc> = 0
>> MainProcess|jsonrpc/5::DEBUG::2017-02-03 08:27:53,774::vsctl::57::root::(commit)
>> Executing commands: /usr/bin/ovs-vsctl --oneline --format=json -- list
>> Bridge -- list Port -- list Interface
>> MainProcess|jsonrpc/5::DEBUG::2017-02-03
08:27:53,774::commands::69::root::(execCmd)
>> /usr/bin/taskset --cpu-list 0-47 /usr/bin/ovs-vsctl --oneline --format=json
>> -- list Bridge -- list Port -- list Interface (cwd None)
>> MainProcess|jsonrpc/5::DEBUG::2017-02-03
08:27:53,807::commands::93::root::(execCmd)
>> FAILED: <err> = 'ovs-vsctl: unix:/var/run/openvswitch/db.sock:
database
>> connection failed (No such file or directory)\n'; <rc> = 1
>> MainProcess|jsonrpc/5::ERROR::2017-02-03 08:27:53,808::supervdsmServer:
>> :97::SuperVdsm.ServerCallback::(wrapper) Error in setupNetworks
>> Traceback (most recent call last):
>> File "/usr/share/vdsm/supervdsmServer", line 95, in wrapper
>> res = func(*args, **kwargs)
>> File "/usr/lib/python2.7/site-packages/vdsm/network/api.py", line
>> 254, in setupNetworks
>> netswitch.validate(networks, bondings)
>> File "/usr/lib/python2.7/site-packages/vdsm/network/netswitch.py",
>> line 121, in validate
>> ovs_switch.validate_network_setup(ovs_nets, ovs_bonds)
>> File "/usr/lib/python2.7/site-packages/vdsm/network/ovs/switch.py",
>> line 40, in validate_network_setup
>> ovs_networks = info.create_netinfo(info.OvsInfo())['networks']
>> File "/usr/lib/python2.7/site-packages/vdsm/network/ovs/info.py",
>> line 73, in __init__
>> ovs_db = OvsDB(driver.create())
>> File "/usr/lib/python2.7/site-packages/vdsm/network/ovs/info.py",
>> line 64, in __init__
>> transaction.add(ifaces_command)
>> File
"/usr/lib/python2.7/site-packages/vdsm/network/ovs/driver/__init__.py",
>> line 54, in __exit__
>> self.result = self.commit()
>> File
"/usr/lib/python2.7/site-packages/vdsm/network/ovs/driver/vsctl.py",
>> line 63, in commit
>> 'Executing commands failed: %s' % '\n'.join(err))
>> ConfigNetworkError: (21, 'Executing commands failed: ovs-vsctl:
>> unix:/var/run/openvswitch/db.sock: database connection failed (No such
>> file or directory)')
>> MainProcess|Thread-2::DEBUG::2017-02-03 08:27:54,440::supervdsmServer:
>> :93::SuperVdsm.ServerCallback::(wrapper) call ksmTune with () {}
>> MainProcess|Thread-2::DEBUG::2017-02-03 08:27:54,440::supervdsmServer:
>> :100::SuperVdsm.ServerCallback::(wrapper) return ksmTune with None
>> MainProcess|jsonrpc/0::DEBUG::2017-02-03 08:30:00,918::supervdsmServer:
>> :93::SuperVdsm.ServerCallback::(wrapper) call network_caps with () {}
>> MainProcess|jsonrpc/0::DEBUG::2017-02-03
08:30:00,922::commands::69::root::(execCmd)
>> /usr/bin/taskset --cpu-list 0-47 /sbin/ip route show to 0.0.0.0/0 table
>> main (cwd None)
>> MainProcess|jsonrpc/0::DEBUG::2017-02-03
08:30:00,927::commands::93::root::(execCmd)
>> SUCCESS: <err> = ''; <rc> = 0
>> MainProcess|jsonrpc/0::DEBUG::2017-02-03
08:30:01,013::commands::69::root::(execCmd)
>> /usr/bin/taskset --cpu-list 0-47 /usr/sbin/tc qdisc show (cwd None)
>> MainProcess|jsonrpc/0::DEBUG::2017-02-03
08:30:01,016::commands::93::root::(execCmd)
>> SUCCESS: <err> = ''; <rc> = 0
>> MainProcess|jsonrpc/0::DEBUG::2017-02-03 08:30:01,017::supervdsmServer:
>> :100::SuperVdsm.ServerCallback::(wrapper) return network_caps with
>> {'bridges': {}, 'bondings': {}, 'nameservers':
['10.1.1.8'], 'nics':
>> {'ib0': {'ipv6autoconf': True, 'addr': '',
'ipv6gateway': '::', 'dhcpv6':
>> False, 'ipv6addrs': [], 'mtu': '4092', 'dhcpv4':
False, 'netmask': '',
>> 'ipv4defaultroute': False, 'ipv4addrs': [], 'hwaddr':
>> '80:00:02:08:fe:80:00:00:00:00:00:00:0c:c4:7a:ff:ff:0f:d8:85',
'speed':
>> 0, 'gateway': ''}, 'eno1': {'ipv6autoconf':
False, 'addr': '10.1.235.5',
>> 'ipv6gateway': '::', 'dhcpv6': False,
'ipv6addrs': [], 'mtu': '1500',
>> 'dhcpv4': False, 'netmask': '255.255.0.0',
'ipv4defaultroute': True,
>> 'ipv4addrs': ['10.1.235.5/16'], 'hwaddr':
'0c:c4:7a:0f:6b:2a', 'speed':
>> 1000, 'gateway': '10.1.0.1'}, 'eno2':
{'ipv6autoconf': True, 'addr': '',
>> 'ipv6gateway': '::', 'dhcpv6': False,
'ipv6addrs': [], 'mtu': '1500',
>> 'dhcpv4': False, 'netmask': '',
'ipv4defaultroute': False, 'ipv4addrs': [],
>> 'hwaddr': '0c:c4:7a:0f:6b:2b', 'speed': 0,
'gateway': ''}}, 'supportsIPv6':
>> True, 'vlans': {}, 'networks': {}}
>> MainProcess|jsonrpc/1::DEBUG::2017-02-03 08:30:01,036::supervdsmServer:
>> :93::SuperVdsm.ServerCallback::(wrapper) call getHardwareInfo with () {}
>> MainProcess|jsonrpc/1::DEBUG::2017-02-03 08:30:01,066::supervdsmServer:
>> :100::SuperVdsm.ServerCallback::(wrapper) return getHardwareInfo with
>> {'systemProductName': 'X9DRT', 'systemSerialNumber':
'0123456789',
>> 'systemFamily': 'To be filled by O.E.M.',
'systemVersion': '0123456789',
>> 'systemUUID': '00000000-0000-0000-0000-0CC47A0F6B2A',
>> 'systemManufacturer': 'Supermicro'}
>> (END)
>>
>>
>>
>> On Fri, Feb 3, 2017 at 12:05 AM, Edward Haas <ehaas(a)redhat.com> wrote:
>>
>>> Hello Shalabh,
>>>
>>> Please provide the logs from your node:
>>> - messages
>>> - vdsm/vdsm.log, vdsm/supervdsm.log
>>>
>>> It may be that you are missing openvswitch installed, although VDSM
>>> should not require it for its operation.
>>>
>>> Thanks,
>>> Edy.
>>>
>>>
>>> On Thu, Feb 2, 2017 at 2:10 PM, Shalabh Goel <shalabhgoel13(a)gmail.com>
>>> wrote:
>>>
>>>> HI,
>>>>
>>>> I am getting the following error on my node after rebooting it.
>>>>
>>>> VDSM ovirtnode2 command HostSetupNetworksVDS failed: Executing
>>>> commands failed: ovs-vsctl: unix:/var/run/openvswitch/db.sock:
>>>> database connection failed (No such file or directory)
>>>>
>>>>
>>>> To solve this, I am trying to restart ovsdb-server using the following
>>>> command,
>>>>
>>>> ovsdb-server --remote=punix:*/var/run/openvswitch/db.sock*
>>>> --remote=db:Open_vSwitch,Open_vSwitch,manager_options
>>>>
--private-key=db:Open_vSwitch,SSL,private_key--certificate=db:Open_vSwitch,SSL,certificate
>>>> --bootstrap-ca-cert=db:Open_vSwitch,SSL,ca_cert --pidfile --detach
>>>>
>>>> But I am getting the following error.
>>>>
>>>> ovsdb-server: /var/run/openvswitch/ovsdb-server.pid.tmp: create
>>>> failed (No such file or directory)
>>>>
>>>> How to restart the ovsdb-server?? Also ovirtmgmt network is missing
>>>> from my node. It happened after I rebooted my node after it got upgraded
to
>>>> Ovirt 4.1
>>>>
>>>> --
>>>> Shalabh Goel
>>>>
>>>> _______________________________________________
>>>> Users mailing list
>>>> Users(a)ovirt.org
>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>
>>>>
>>>
>>
>>
>> --
>> Shalabh Goel
>>
>
>
--
Shalabh Goel