[ovirt-users] problem after rebooting the node

Shalabh Goel shalabhgoel13 at gmail.com
Mon Feb 6 11:14:16 UTC 2017


which all log files? Actually I am new to Ovirt, so it would be really
helpful if  you can tell me which ones??

Thanks

On Mon, Feb 6, 2017 at 4:39 PM, Edward Haas <ehaas at redhat.com> wrote:

> Please package the logs (tar or zip) and send them.
>
> On Mon, Feb 6, 2017 at 12:05 PM, Shalabh Goel <shalabhgoel13 at gmail.com>
> wrote:
>
>> Yes, I am using OVS as the switch type and I did not know that it was not
>> supported officially.
>>
>> The output of ovs-vsctl show is as follows:
>>
>> f634d53e-4849-488b-8454-6b1fafa7c6ac
>>     ovs_version: "2.6.90"
>>
>> I am attaching OVS switch logs below:
>>
>> /var/log/openvswitch/ovsdb-server.log
>>
>>
>> 2017-02-06T09:46:07.788Z|00001|vlog|INFO|opened log file
>> /var/log/openvswitch/ovsdb-server.log
>> 2017-02-06T09:46:07.791Z|00002|ovsdb_server|INFO|ovsdb-server (Open
>> vSwitch) 2.6.90
>> 2017-02-06T09:46:17.802Z|00003|memory|INFO|2296 kB peak resident set
>> size after 10.0 seconds
>> 2017-02-06T09:46:17.802Z|00004|memory|INFO|cells:16 json-caches:1
>> monitors:1 sessions:1
>>
>> ovs-vswitchd.log
>>
>>
>> 2017-02-06T09:46:07.999Z|00001|vlog|INFO|opened log file
>> /var/log/openvswitch/ovs-vswitchd.log
>> 2017-02-06T09:46:08.036Z|00002|ovs_numa|INFO|Discovered 24 CPU cores on
>> NUMA node 0
>> 2017-02-06T09:46:08.036Z|00003|ovs_numa|INFO|Discovered 24 CPU cores on
>> NUMA node 1
>> 2017-02-06T09:46:08.036Z|00004|ovs_numa|INFO|Discovered 2 NUMA nodes and
>> 48 CPU cores
>> 2017-02-06T09:46:08.037Z|00005|reconnect|INFO|unix:/var/run/openvswitch/db.sock:
>> connecting...
>> 2017-02-06T09:46:08.037Z|00006|reconnect|INFO|unix:/var/run/openvswitch/db.sock:
>> connected
>> 2017-02-06T09:46:08.039Z|00007|bridge|INFO|ovs-vswitchd (Open vSwitch)
>> 2.6.90
>>
>> What should I do now?
>>
>> The engine says that "Host host2 does not comply with the cluster Default
>> networks, the following networks are missing on host: 'ovirtmgmt'"
>>
>> What other logs should I attach?
>>
>> Thanks
>>
>> Shalabh Goel
>>
>> On Sun, Feb 5, 2017 at 1:10 PM, Edward Haas <ehaas at redhat.com> wrote:
>>
>>> Based on what I can see, you used OVS as the switch type and it seems
>>> ovs (openvswitch) is not properly installed on your host.
>>> Make sure that you have ovs operational by issuing "ovs-vsctl show".
>>>
>>> You should note that OVS network support is not an official release
>>> feature, and you should use it on 4.1 and up versions.
>>> Fixes will be probably submitted to master (appearing in nightly builds).
>>>
>>> Next time please include the mailing-list in your replies and attach the
>>> log files, it is less spamming.
>>>
>>> Thanks,
>>> Edy.
>>>
>>> On Fri, Feb 3, 2017 at 5:07 AM, Shalabh Goel <shalabhgoel13 at gmail.com>
>>> wrote:
>>>
>>>> log from messages
>>>>
>>>> Feb  3 08:27:53 ovirtnode3 ovs-vsctl: ovs|00001|db_ctl_base|ERR|unix:/var/run/openvswitch/db.sock:
>>>> database connection failed (No such file or directory)
>>>> Feb  3 08:27:53 ovirtnode3 journal: vdsm vds ERROR Executing commands
>>>> failed: ovs-vsctl: unix:/var/run/openvswitch/db.sock: database
>>>> connection failed (No su
>>>> ch file or directory)#012Traceback (most recent call last):#012  File
>>>> "/usr/share/vdsm/API.py", line 1531, in setupNetworks#012
>>>> supervdsm.getProxy().setup
>>>> Networks(networks, bondings, options)#012  File
>>>> "/usr/lib/python2.7/site-packages/vdsm/supervdsm.py", line 53, in
>>>> __call__#012    return callMethod()#012  Fi
>>>> le "/usr/lib/python2.7/site-packages/vdsm/supervdsm.py", line 51, in
>>>> <lambda>#012    **kwargs)#012  File "<string>", line 2, in
>>>> setupNetworks#012  File "/usr
>>>> /lib64/python2.7/multiprocessing/managers.py", line 773, in
>>>> _callmethod#012    raise convert_to_error(kind,
>>>> result)#012ConfigNetworkError: (21, 'Executing co
>>>> mmands failed: ovs-vsctl: unix:/var/run/openvswitch/db.sock: database
>>>> connection failed (No such file or directory)')
>>>>
>>>> Log from vdsm.log
>>>>
>>>> Reactor thread::INFO::2017-02-03 08:32:24,638::protocoldetector
>>>> ::72::ProtocolDetector.AcceptorImpl::(handle_accept) Accepted
>>>> connection from ::1:53082
>>>> Reactor thread::DEBUG::2017-02-03 08:32:24,643::protocoldetector
>>>> ::91::ProtocolDetector.Detector::(__init__) Using required_size=11
>>>> Reactor thread::INFO::2017-02-03 08:32:24,643::protocoldetector
>>>> ::127::ProtocolDetector.Detector::(handle_read) Detected protocol xml
>>>> from ::1:53082
>>>> Reactor thread::DEBUG::2017-02-03 08:32:24,643::bindingxmlrpc::1
>>>> 317::XmlDetector::(handle_socket) xml over http detected from ('::1',
>>>> 53082)
>>>> BindingXMLRPC::INFO::2017-02-03 08:32:24,643::xmlrpc::73::vds.XMLRPCServer::(handle_request)
>>>> Starting request handler for ::1:53082
>>>> Thread-20::INFO::2017-02-03 08:32:24,644::xmlrpc::83::vds.
>>>> XMLRPCServer::(_process_requests) Request handler for ::1:53082 started
>>>> Thread-20::DEBUG::2017-02-03 08:32:24,644::bindingxmlrpc::1263::vds::(wrapper)
>>>> client [::1]::call getAllVmStats with () {}
>>>> Thread-20::DEBUG::2017-02-03 08:32:24,644::bindingxmlrpc::1270::vds::(wrapper)
>>>> return getAllVmStats with {'status': {'message': 'Done', 'code': 0},
>>>> 'statsLis
>>>> t': (suppressed)}
>>>> Thread-20::INFO::2017-02-03 08:32:24,644::bindingxmlrpc::1297::vds::(wrapper)
>>>> RPC call getAllVmStats finished (code=0) in 0.00 seconds
>>>> Thread-20::INFO::2017-02-03 08:32:24,645::xmlrpc::91::vds.
>>>> XMLRPCServer::(_process_requests) Request handler for ::1:53082 stopped
>>>> periodic/0::DEBUG::2017-02-03 08:32:24,907::sampling::508::v
>>>> irt.sampling.VMBulkSampler::(__call__) sampled timestamp 4347942.15
>>>> elapsed 0.010 acquired True d
>>>> omains all
>>>> periodic/2::DEBUG::2017-02-03 08:32:24,908::task::599::stora
>>>> ge.TaskManager.Task::(_updateState) (Task='bb07f89c-6165-42bc-902f-1b7701b813fa')
>>>> moving from sta
>>>> te init -> state preparing
>>>> periodic/2::INFO::2017-02-03 08:32:24,908::logUtils::49::dispatcher::(wrapper)
>>>> Run and protect: repoStats(options=None)
>>>> periodic/2::INFO::2017-02-03 08:32:24,908::logUtils::52::dispatcher::(wrapper)
>>>> Run and protect: repoStats, Return response: {}
>>>> periodic/2::DEBUG::2017-02-03 08:32:24,909::task::1195::stor
>>>> age.TaskManager.Task::(prepare) (Task='bb07f89c-6165-42bc-902f-1b7701b813fa')
>>>> finished: {}
>>>> periodic/2::DEBUG::2017-02-03 08:32:24,909::task::599::stora
>>>> ge.TaskManager.Task::(_updateState) (Task='bb07f89c-6165-42bc-902f-1b7701b813fa')
>>>> moving from sta
>>>> te preparing -> state finished
>>>> periodic/2::DEBUG::2017-02-03 08:32:24,909::resourceManager:
>>>> :910::storage.ResourceManager.Owner::(releaseAll) Owner.releaseAll
>>>> requests {} resources {}
>>>> periodic/2::DEBUG::2017-02-03 08:32:24,909::resourceManager:
>>>> :947::storage.ResourceManager.Owner::(cancelAll) Owner.cancelAll
>>>> requests {}
>>>> periodic/2::DEBUG::2017-02-03 08:32:24,909::task::997::storage.TaskManager.Task::(_decref)
>>>> (Task='bb07f89c-6165-42bc-902f-1b7701b813fa') ref 0 aborting False
>>>> jsonrpc/4::DEBUG::2017-02-03 08:32:26,653::__init__::532::j
>>>> sonrpc.JsonRpcServer::(_handle_request) Calling 'Host.getAllVmStats'
>>>> in bridge with {}
>>>> jsonrpc/4::DEBUG::2017-02-03 08:32:26,653::__init__::557::j
>>>> sonrpc.JsonRpcServer::(_handle_request) Return 'Host.getAllVmStats' in
>>>> bridge with (suppressed)
>>>> jsonrpc/4::INFO::2017-02-03 08:32:26,653::__init__::515::j
>>>> sonrpc.JsonRpcServer::(_serveRequest) RPC call Host.getAllVmStats
>>>> succeeded in 0.00 seconds
>>>> jsonrpc/6::DEBUG::2017-02-03 08:32:27,706::__init__::532::j
>>>> sonrpc.JsonRpcServer::(_handle_request) Calling 'Host.getStats' in
>>>> bridge with {}
>>>> jsonrpc/6::DEBUG::2017-02-03 08:32:27,706::__init__::532::j
>>>> sonrpc.JsonRpcServer::(_handle_request) Calling 'Host.getStats' in
>>>> bridge with {}
>>>> jsonrpc/6::DEBUG::2017-02-03 08:32:27,707::task::599::stora
>>>> ge.TaskManager.Task::(_updateState) (Task='9285329e-485b-4165-8bdc-4f411f394356')
>>>> moving from state init -> state preparing
>>>> jsonrpc/6::INFO::2017-02-03 08:32:27,707::logUtils::49::dispatcher::(wrapper)
>>>> Run and protect: repoStats(options=None)
>>>> jsonrpc/6::INFO::2017-02-03 08:32:27,707::logUtils::52::dispatcher::(wrapper)
>>>> Run and protect: repoStats, Return response: {}
>>>> jsonrpc/6::DEBUG::2017-02-03 08:32:27,707::task::1195::stor
>>>> age.TaskManager.Task::(prepare) (Task='9285329e-485b-4165-8bdc-4f411f394356')
>>>> finished: {}
>>>> jsonrpc/6::DEBUG::2017-02-03 08:32:27,707::task::599::stora
>>>> ge.TaskManager.Task::(_updateState) (Task='9285329e-485b-4165-8bdc-4f411f394356')
>>>> moving from state preparing -> state finished
>>>> jsonrpc/6::DEBUG::2017-02-03 08:32:27,708::resourceManager:
>>>> :910::storage.ResourceManager.Owner::(releaseAll) Owner.releaseAll
>>>> requests {} resources {}
>>>> jsonrpc/6::DEBUG::2017-02-03 08:32:27,708::resourceManager:
>>>> :947::storage.ResourceManager.Owner::(cancelAll) Owner.cancelAll
>>>> requests {}
>>>> jsonrpc/6::DEBUG::2017-02-03 08:32:27,708::task::997::storage.TaskManager.Task::(_decref)
>>>> (Task='9285329e-485b-4165-8bdc-4f411f394356') ref 0 aborting False
>>>> jsonrpc/6::DEBUG::2017-02-03 08:32:27,712::__init__::557::j
>>>> sonrpc.JsonRpcServer::(_handle_request) Return 'Host.getStats' in
>>>> bridge with {'cpuStatistics': {'28': {'cpuUser': '0.13', 'nodeIndex': 0,
>>>> 'cpuSys': '0.00', 'cpuIdle': '99.87'}, '29': {'cpuUser': '0.00',
>>>> 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '24': {'cpuUser':
>>>> '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '25':
>>>> {'cpuUser': '0.13', 'nodeIndex': 0, 'cpuSys': '0.07', 'cpuIdle': '99.80'},
>>>> '26': {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle':
>>>> '100.00'}, '27': {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.00',
>>>> 'cpuIdle': '100.00'}, '20': {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys':
>>>> '0.00', 'cpuIdle': '100.00'}, '21': {'cpuUser': '0.00', 'nodeIndex': 1,
>>>> 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '22': {'cpuUser': '0.00',
>>>> 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '23': {'cpuUser':
>>>> '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '46':
>>>> {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'},
>>>> '47': {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle':
>>>> '100.00'}, '44': {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': '0.00',
>>>> 'cpuIdle': '100.00'}, '45': {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys':
>>>> '0.00', 'cpuIdle': '100.00'}, '42': {'cpuUser': '0.00', 'nodeIndex': 1,
>>>> 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '43': {'cpuUser': '0.00',
>>>> 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '40': {'cpuUser':
>>>> '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '41':
>>>> {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'},
>>>> '1': {'cpuUser': '0.13', 'nodeIndex': 0, 'cpuSys': '0.07', 'cpuIdle':
>>>> '99.80'}, '0': {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.00',
>>>> 'cpuIdle': '100.00'}, '3': {'cpuUser': '0.07', 'nodeIndex': 0, 'cpuSys':
>>>> '0.07', 'cpuIdle': '99.86'}, '2': {'cpuUser': '0.07', 'nodeIndex': 0,
>>>> 'cpuSys': '0.00', 'cpuIdle': '99.93'}, '5': {'cpuUser': '0.00',
>>>> 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '4': {'cpuUser':
>>>> '0.00', 'nodeIndex': 0, 'cpuSys': '0.07', 'cpuIdle': '99.93'}, '7':
>>>> {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': '100.00'},
>>>> '6': {'cpuUser': '0.20', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle':
>>>> '99.80'}, '9': {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.00',
>>>> 'cpuIdle': '100.00'}, '8': {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys':
>>>> '0.00', 'cpuIdle': '100.00'}, '39': {'cpuUser': '0.00', 'nodeIndex': 1,
>>>> 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '38': {'cpuUser': '0.00',
>>>> 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '11': {'cpuUser':
>>>> '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '10':
>>>> {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': '100.00'},
>>>> '13': {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle':
>>>> '100.00'}, '12': {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': '0.00',
>>>> 'cpuIdle': '100.00'}, '15': {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys':
>>>> '0.00', 'cpuIdle': '100.00'}, '14': {'cpuUser': '0.00', 'nodeIndex': 1,
>>>> 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '17': {'cpuUser': '0.00',
>>>> 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '16': {'cpuUser':
>>>> '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '19':
>>>> {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'},
>>>> '18': {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle':
>>>> '100.00'}, '31': {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.00',
>>>> 'cpuIdle': '100.00'}, '30': {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys':
>>>> '0.00', 'cpuIdle': '100.00'}, '37': {'cpuUser': '0.00', 'nodeIndex': 1,
>>>> 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '36': {'cpuUser': '0.00',
>>>> 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '35': {'cpuUser':
>>>> '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '34':
>>>> {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': '100.00'},
>>>> '33': {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle':
>>>> '100.00'}, '32': {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.00',
>>>> 'cpuIdle': '100.00'}}, 'numaNodeMemFree': {'1': {'memPercent': 3,
>>>> 'memFree': '63780'}, '0': {'memPercent': 4, 'memFree': '63426'}},
>>>> 'memShared': 0, 'thpState': 'always', 'ksmMergeAcrossNodes': True,
>>>> 'vmCount': 0, 'memUsed': '2', 'storageDomains': {}, 'incomingVmMigrations':
>>>> 0, 'network': {'ib0': {'txErrors': '0', 'state': 'down', 'sampleTime':
>>>> 1486090944.903051, 'name': 'ib0', 'tx': '0', 'txDropped': '0', 'rx': '0',
>>>> 'rxErrors': '0', 'speed': '1000', 'rxDropped': '0'}, 'lo': {'txErrors':
>>>> '0', 'state': 'up', 'sampleTime': 1486090944.903051, 'name': 'lo', 'tx':
>>>> '56174388', 'txDropped': '0', 'rx': '56174388', 'rxErrors': '0', 'speed':
>>>> '1000', 'rxDropped': '0'}, 'eno2': {'txErrors': '0', 'state': 'down',
>>>> 'sampleTime': 1486090944.903051, 'name': 'eno2', 'tx': '0', 'txDropped':
>>>> '0', 'rx': '0', 'rxErrors': '0', 'speed': '1000', 'rxDropped': '0'},
>>>> 'eno1': {'txErrors': '0', 'state': 'up', 'sampleTime': 1486090944.903051,
>>>> 'name': 'eno1', 'tx': '26032823', 'txDropped': '0', 'rx': '174722400',
>>>> 'rxErrors': '0', 'speed': '1000', 'rxDropped': '5624'}, ';vdsmdummy;':
>>>> {'txErrors': '0', 'state': 'down', 'sampleTime': 1486090944.903051, 'name':
>>>> ';vdsmdummy;', 'tx': '0', 'txDropped': '0', 'rx': '0', 'rxErrors': '0',
>>>> 'speed': '1000', 'rxDropped': '0'}}, 'txDropped': '0', 'anonHugePages':
>>>> '34', 'ksmPages': 100, 'elapsedTime': '287.81', 'cpuLoad': '0.07',
>>>> 'cpuSys': '0.00', 'diskStats': {'/var/log': {'free': '15072'},
>>>> '/var/log/core': {'free': '15072'}, '/var/run/vdsm/': {'free': '64418'},
>>>> '/tmp': {'free': '914756'}}, 'cpuUserVdsmd': '0.13', 'netConfigDirty':
>>>> 'True', 'memCommitted': 0, 'ksmState': False, 'vmMigrating': 0, 'ksmCpu':
>>>> 0, 'memAvailable': 127200, 'bootTime': '1486037670', 'haStats': {'active':
>>>> False, 'configured': False, 'score': 0, 'localMaintenance': False,
>>>> 'globalMaintenance': False}, 'momStatus': 'active', 'rxDropped': '5624',
>>>> 'outgoingVmMigrations': 0, 'swapTotal': 4095, 'swapFree': 4095, 'dateTime':
>>>> '2017-02-03T03:02:27 GMT', 'cpuUser': '0.02', 'memFree': 127456, 'cpuIdle':
>>>> '99.98', 'vmActive': 0, 'v2vJobs': {}, 'cpuSysVdsmd': '0.07'}
>>>> jsonrpc/6::INFO::2017-02-03 08:32:27,713::__init__::515::j
>>>> sonrpc.JsonRpcServer::(_serveRequest) RPC call Host.getStats succeeded
>>>> in 0.00 seconds
>>>> Reactor thread::INFO::2017-02-03 08:32:39,661::protocoldetector
>>>> ::72::ProtocolDetector.AcceptorImpl::(handle_accept) Accepted
>>>> connection from ::1:53084
>>>>
>>>> supervdsm.log
>>>>
>>>>
>>>>   File "/usr/share/vdsm/supervdsmServer", line 95, in wrapper
>>>>     res = func(*args, **kwargs)
>>>>   File "/usr/share/vdsm/supervdsmServer", line 143, in umount
>>>>     timeout=timeout)
>>>>   File "/usr/lib/python2.7/site-packages/vdsm/storage/mount.py", line
>>>> 281, in _umount
>>>>     _runcmd(cmd, timeout)
>>>>   File "/usr/lib/python2.7/site-packages/vdsm/storage/mount.py", line
>>>> 297, in _runcmd
>>>>     raise MountError(rc, ";".join((out, err)))
>>>> MountError: (32, ';umount: /rhev/data-center/mnt/10.1.235.6:_iso:
>>>> mountpoint not found\n')
>>>> MainProcess|jsonrpc/5::DEBUG::2017-02-03 08:26:59,495::supervdsmServer:
>>>> :93::SuperVdsm.ServerCallback::(wrapper) call hbaRescan with () {}
>>>> MainProcess|jsonrpc/5::DEBUG::2017-02-03 08:26:59,495::commands::69::storage.HBA::(execCmd)
>>>> /usr/bin/taskset --cpu-list 0-47 /usr/libexec/vdsm/fc-scan (cwd N
>>>> one)
>>>> MainProcess|jsonrpc/5::DEBUG::2017-02-03 08:26:59,579::supervdsmServer:
>>>> :100::SuperVdsm.ServerCallback::(wrapper) return hbaRescan with None
>>>> MainThread::DEBUG::2017-02-03 08:27:26,367::supervdsmServer::311::SuperVdsm.Server::(main)
>>>> Terminated normally
>>>> MainThread::DEBUG::2017-02-03 08:27:36,148::__init__::47::blivet::(register_device_format)
>>>> registered device format class DeviceFormat as None
>>>> MainThread::DEBUG::2017-02-03 08:27:36,149::__init__::47::blivet::(register_device_format)
>>>> registered device format class BIOSBoot as biosboot
>>>> MainThread::DEBUG::2017-02-03 08:27:36,150::__init__::47::blivet::(register_device_format)
>>>> registered device format class DiskLabel as disklabel
>>>> MainThread::DEBUG::2017-02-03 08:27:36,150::__init__::47::blivet::(register_device_format)
>>>> registered device format class DMRaidMember as dmraidmember
>>>> MainThread::DEBUG::2017-02-03 08:27:36,151::__init__::47::blivet::(register_device_format)
>>>> registered device format class Ext2FS as ext2
>>>> MainThread::DEBUG::2017-02-03 08:27:36,152::__init__::47::blivet::(register_device_format)
>>>> registered device format class Ext3FS as ext3
>>>> MainThread::DEBUG::2017-02-03 08:27:36,152::__init__::47::blivet::(register_device_format)
>>>> registered device format class Ext4FS as ext4
>>>> MainThread::DEBUG::2017-02-03 08:27:36,152::__init__::47::blivet::(register_device_format)
>>>> registered device format class FATFS as vfat
>>>> MainThread::DEBUG::2017-02-03 08:27:36,152::__init__::47::blivet::(register_device_format)
>>>> registered device format class EFIFS as efi
>>>> MainThread::DEBUG::2017-02-03 08:27:36,152::__init__::47::blivet::(register_device_format)
>>>> registered device format class BTRFS as btrfs
>>>> MainThread::DEBUG::2017-02-03 08:27:36,153::__init__::47::blivet::(register_device_format)
>>>> registered device format class GFS2 as gfs2
>>>> MainThread::DEBUG::2017-02-03 08:27:36,153::__init__::47::blivet::(register_device_format)
>>>> registered device format class JFS as jfs
>>>> MainThread::DEBUG::2017-02-03 08:27:36,153::__init__::47::blivet::(register_device_format)
>>>> registered device format class ReiserFS as reiserfs
>>>> MainThread::DEBUG::2017-02-03 08:27:36,153::__init__::47::blivet::(register_device_format)
>>>> registered device format class XFS as xfs
>>>> MainThread::DEBUG::2017-02-03 08:27:36,153::__init__::47::blivet::(register_device_format)
>>>> registered device format class HFS as hfs
>>>> MainThread::DEBUG::2017-02-03 08:27:36,153::__init__::47::blivet::(register_device_format)
>>>> registered device format class AppleBootstrapFS as appleboot
>>>> MainThread::DEBUG::2017-02-03 08:27:36,154::__init__::47::blivet::(register_device_format)
>>>> registered device format class HFSPlus as hfs+
>>>> MainThread::DEBUG::2017-02-03 08:27:36,154::__init__::47::blivet::(register_device_format)
>>>> registered device format class MacEFIFS as macefi
>>>> MainThread::DEBUG::2017-02-03 08:27:36,154::__init__::47::blivet::(register_device_format)
>>>> registered device format class NTFS as ntfs
>>>> MainThread::DEBUG::2017-02-03 08:27:36,154::__init__::47::blivet::(register_device_format)
>>>> registered device format class NFS as nfs
>>>> MainThread::DEBUG::2017-02-03 08:27:36,154::__init__::47::blivet::(register_device_format)
>>>> registered device format class NFSv4 as nfs4
>>>> MainThread::DEBUG::2017-02-03 08:27:36,154::__init__::47::blivet::(register_device_format)
>>>> registered device format class Iso9660FS as iso9660
>>>> MainThread::DEBUG::2017-02-03 08:27:36,154::__init__::47::blivet::(register_device_format)
>>>> registered device format class NoDevFS as nodev
>>>> MainThread::DEBUG::2017-02-03 08:27:36,155::__init__::47::blivet::(register_device_format)
>>>> registered device format class DevPtsFS as devpts
>>>> MainThread::DEBUG::2017-02-03 08:27:36,155::__init__::47::blivet::(register_device_format)
>>>> registered device format class ProcFS as proc
>>>> MainThread::DEBUG::2017-02-03 08:27:36,155::__init__::47::blivet::(register_device_format)
>>>> registered device format class SysFS as sysfs
>>>> MainThread::DEBUG::2017-02-03 08:27:36,155::__init__::47::blivet::(register_device_format)
>>>> registered device format class TmpFS as tmpfs
>>>> MainThread::DEBUG::2017-02-03 08:27:36,155::__init__::47::blivet::(register_device_format)
>>>> registered device format class BindFS as bind
>>>> MainThread::DEBUG::2017-02-03 08:27:36,155::__init__::47::blivet::(register_device_format)
>>>> registered device format class SELinuxFS as selinuxfs
>>>> MainThread::DEBUG::2017-02-03 08:27:36,155::__init__::47::blivet::(register_device_format)
>>>> registered device format class USBFS as usbfs
>>>> MainThread::DEBUG::2017-02-03 08:27:36,155::__init__::47::blivet::(register_device_format)
>>>> registered device format class EFIVarFS as efivarfs
>>>> MainThread::DEBUG::2017-02-03 08:27:36,179::__init__::47::blivet::(register_device_format)
>>>> registered device format class LUKS as luks
>>>> MainThread::DEBUG::2017-02-03 08:27:36,181::__init__::47::blivet::(register_device_format)
>>>> registered device format class LVMPhysicalVolume as lvmpv
>>>> MainThread::DEBUG::2017-02-03 08:27:36,181::__init__::47::blivet::(register_device_format)
>>>> registered device format class MDRaidMember as mdmember
>>>> MainThread::DEBUG::2017-02-03 08:27:36,181::__init__::47::blivet::(register_device_format)
>>>> registered device format class MultipathMember as multipath_member
>>>> MainThread::DEBUG::2017-02-03 08:27:36,181::__init__::47::blivet::(register_device_format)
>>>> registered device format class PPCPRePBoot as prepboot
>>>> MainThread::DEBUG::2017-02-03 08:27:36,182::__init__::47::blivet::(register_device_format)
>>>> registered device format class SwapSpace as swap
>>>> MainThread::DEBUG::2017-02-03 08:27:36,216::storage_log::69::blivet::(log_exception_info)
>>>> IGNORED:        Caught exception, continuing.
>>>> MainThread::DEBUG::2017-02-03 08:27:36,216::storage_log::72::blivet::(log_exception_info)
>>>> IGNORED:        Problem description: failed to get initiator name from
>>>> iscsi firmware
>>>> MainThread::DEBUG::2017-02-03 08:27:36,216::storage_log::73::blivet::(log_exception_info)
>>>> IGNORED:        Begin exception details.
>>>> MainThread::DEBUG::2017-02-03 08:27:36,216::storage_log::76::blivet::(log_exception_info)
>>>> IGNORED:            Traceback (most recent call last):
>>>> MainThread::DEBUG::2017-02-03 08:27:36,216::storage_log::76::blivet::(log_exception_info)
>>>> IGNORED:              File "/usr/lib/python2.7/site-packages/blivet
>>>> /iscsi.py", line 146, in __init__
>>>> MainThread::DEBUG::2017-02-03 08:27:36,216::storage_log::76::blivet::(log_exception_info)
>>>> IGNORED:                initiatorname = libiscsi.get_firmware_initi
>>>> ator_name()
>>>> MainThread::DEBUG::2017-02-03 08:27:36,216::storage_log::76::blivet::(log_exception_info)
>>>> IGNORED:            IOError: Unknown error
>>>> MainThread::DEBUG::2017-02-03 08:27:36,216::storage_log::77::blivet::(log_exception_info)
>>>> IGNORED:        End exception details.
>>>> MainThread::DEBUG::2017-02-03 08:27:36,227::supervdsmServer::271::SuperVdsm.Server::(main)
>>>> Making sure I'm root - SuperVdsm
>>>> MainThread::DEBUG::2017-02-03 08:27:36,227::supervdsmServer::280::SuperVdsm.Server::(main)
>>>> Parsing cmd args
>>>> MainThread::DEBUG::2017-02-03 08:27:36,227::supervdsmServer::283::SuperVdsm.Server::(main)
>>>> Cleaning old socket /var/run/vdsm/svdsm.sock
>>>> MainThread::DEBUG::2017-02-03 08:27:36,227::supervdsmServer::287::SuperVdsm.Server::(main)
>>>> Setting up keep alive thread
>>>> MainThread::DEBUG::2017-02-03 08:27:36,228::supervdsmServer::293::SuperVdsm.Server::(main)
>>>> Creating remote object manager
>>>> MainThread::DEBUG::2017-02-03 08:27:36,228::fileUtils::246::storage.fileUtils::(chown)
>>>> Changing owner for /var/run/vdsm/svdsm.sock, to (36:36)
>>>> MainThread::DEBUG::2017-02-03 08:27:36,228::supervdsmServer::303::SuperVdsm.Server::(main)
>>>> Started serving super vdsm object
>>>> sourceRoute::DEBUG::2017-02-03 08:27:36,229::sourceroutethrea
>>>> d::79::root::(_subscribeToInotifyLoop) sourceRouteThread.subscribeToInotifyLoop
>>>> started
>>>> restore-net::INFO::2017-02-03 08:27:37,160::vdsm-restore-net-config::465::root::(restore)
>>>> networks already restored. doing nothing.
>>>> MainProcess|hsm/init::DEBUG::2017-02-03 08:27:39,601::supervdsmServer:
>>>> :93::SuperVdsm.ServerCallback::(wrapper) call hbaRescan with () {}
>>>> MainProcess|hsm/init::DEBUG::2017-02-03 08:27:39,602::commands::69::storage.HBA::(execCmd)
>>>> /usr/bin/taskset --cpu-list 0-47 /usr/libexec/vdsm/fc-scan (cwd No
>>>> ne)
>>>> MainProcess|hsm/init::DEBUG::2017-02-03 08:27:39,691::supervdsmServer:
>>>> :100::SuperVdsm.ServerCallback::(wrapper) return hbaRescan with None
>>>> MainProcess|jsonrpc/4::DEBUG::2017-02-03 08:27:48,355::supervdsmServer:
>>>> :93::SuperVdsm.ServerCallback::(wrapper) call network_caps with () {}
>>>> MainProcess|jsonrpc/4::DEBUG::2017-02-03 08:27:48,360::commands::69::root::(execCmd)
>>>> /usr/bin/taskset --cpu-list 0-47 /sbin/ip route show to 0.0.0.0/0
>>>> table
>>>> main (cwd None)
>>>> MainProcess|jsonrpc/4::DEBUG::2017-02-03 08:27:48,364::commands::93::root::(execCmd)
>>>> SUCCESS: <err> = ''; <rc> = 0
>>>> MainProcess|jsonrpc/4::DEBUG::2017-02-03 08:27:48,408::libvirtconnection::160::root::(get)
>>>> trying to connect libvirt
>>>> MainProcess|jsonrpc/4::DEBUG::2017-02-03 08:27:48,450::commands::69::root::(execCmd)
>>>> /usr/bin/taskset --cpu-list 0-47 /usr/sbin/tc qdisc show (cwd None)
>>>> MainProcess|jsonrpc/4::DEBUG::2017-02-03 08:27:48,453::commands::93::root::(execCmd)
>>>> SUCCESS: <err> = ''; <rc> = 0
>>>> MainProcess|jsonrpc/4::DEBUG::2017-02-03 08:27:48,454::commands::69::root::(execCmd)
>>>> /usr/bin/taskset --cpu-list 0-47 /bin/systemctl --no-pager list-unit-fil
>>>> es (cwd None)
>>>> MainProcess|jsonrpc/4::DEBUG::2017-02-03 08:27:48,530::commands::93::root::(execCmd)
>>>> SUCCESS: <err> = ''; <rc> = 0
>>>> MainProcess|jsonrpc/4::DEBUG::2017-02-03 08:27:48,530::commands::69::root::(execCmd)
>>>> /usr/bin/taskset --cpu-list 0-47 /bin/systemctl status openvswitch.servi
>>>> ce (cwd None)
>>>> MainProcess|jsonrpc/4::DEBUG::2017-02-03 08:27:48,561::commands::93::root::(execCmd)
>>>> FAILED: <err> = ''; <rc> = 3
>>>> MainProcess|jsonrpc/4::DEBUG::2017-02-03 08:27:48,561::supervdsmServer:
>>>> :100::SuperVdsm.ServerCallback::(wrapper) return network_caps with
>>>> {'bridges': {}, 'bo
>>>> ndings': {}, 'nameservers': ['10.1.1.8'], 'nics': {'ib0':
>>>> {'ipv6autoconf': True, 'addr': '', 'ipv6gateway': '::', 'dhcpv6': False,
>>>> 'ipv6addrs': [], 'mtu': '4
>>>> 092', 'dhcpv4': False, 'netmask': '', 'ipv4defaultroute': False,
>>>> 'ipv4addrs': [], 'hwaddr': '80:00:02:08:fe:80:00:00:00:00
>>>> :00:00:0c:c4:7a:ff:ff:0f:d8:85', 's
>>>> peed': 0, 'gateway': ''}, 'eno1': {'ipv6autoconf': False, 'addr':
>>>> '10.1.235.5', 'ipv6gateway': '::', 'dhcpv6': False, 'ipv6addrs': [], 'mtu':
>>>> '1500', 'dhcpv4
>>>> ': False, 'netmask': '255.255.0.0', 'ipv4defaultroute': True,
>>>> 'ipv4addrs': ['10.1.235.5/16'], 'hwaddr': '0c:c4:7a:0f:6b:2a',
>>>> 'speed': 1000, 'gateway': '10.1.
>>>> 0.1'}, 'eno2': {'ipv6autoconf': True, 'addr': '', 'ipv6gateway': '::',
>>>> 'dhcpv6': False, 'ipv6addrs': [], 'mtu': '1500', 'dhcpv4': False,
>>>> 'netmask': '', 'ipv4
>>>> defaultroute': False, 'ipv4addrs': [], 'hwaddr': '0c:c4:7a:0f:6b:2b',
>>>> 'speed': 0, 'gateway': ''}}, 'supportsIPv6': True, 'vlans': {}, 'networks':
>>>> {}}
>>>> MainProcess|jsonrpc/5::DEBUG::2017-02-03 08:27:53,690::supervdsmServer:
>>>> :93::SuperVdsm.ServerCallback::(wrapper) call setupNetworks with ({},
>>>> {u'connectivityC
>>>> heck': u'true', u'connectivityTimeout': 120}) {}
>>>> MainProcess|jsonrpc/5::DEBUG::2017-02-03 08:27:53,691::api::247::root::(setupNetworks)
>>>> Setting up network according to configuration: networks:{u'ovirtmgmt':
>>>> {u'ipv6autoconf': False, u'nic': u'eno1', u'ipaddr': u'10.1.235.5',
>>>> u'netmask': u'255.255.0.0', u'mtu': 1500, u'switch': u'ovs', u'dhcpv6':
>>>> False, u'STP': u'no', u'bridged': u'true', u'gateway': u'10.1.0.1',
>>>> u'defaultRoute': True}}, bondings:{}, options:{u'connectivityCheck':
>>>> u'true', u'connectivityTimeout': 120}
>>>> MainProcess|jsonrpc/5::DEBUG::2017-02-03 08:27:53,691::api::252::root::(setupNetworks)
>>>> Validating configuration
>>>> MainProcess|jsonrpc/5::DEBUG::2017-02-03 08:27:53,694::commands::69::root::(execCmd)
>>>> /usr/bin/taskset --cpu-list 0-47 /sbin/ip route show to 0.0.0.0/0
>>>> table main (cwd None)
>>>> MainProcess|jsonrpc/5::DEBUG::2017-02-03 08:27:53,698::commands::93::root::(execCmd)
>>>> SUCCESS: <err> = ''; <rc> = 0
>>>> MainProcess|jsonrpc/5::DEBUG::2017-02-03 08:27:53,771::commands::69::root::(execCmd)
>>>> /usr/bin/taskset --cpu-list 0-47 /usr/sbin/tc qdisc show (cwd None)
>>>> MainProcess|jsonrpc/5::DEBUG::2017-02-03 08:27:53,774::commands::93::root::(execCmd)
>>>> SUCCESS: <err> = ''; <rc> = 0
>>>> MainProcess|jsonrpc/5::DEBUG::2017-02-03 08:27:53,774::vsctl::57::root::(commit)
>>>> Executing commands: /usr/bin/ovs-vsctl --oneline --format=json -- list
>>>> Bridge -- list Port -- list Interface
>>>> MainProcess|jsonrpc/5::DEBUG::2017-02-03 08:27:53,774::commands::69::root::(execCmd)
>>>> /usr/bin/taskset --cpu-list 0-47 /usr/bin/ovs-vsctl --oneline --format=json
>>>> -- list Bridge -- list Port -- list Interface (cwd None)
>>>> MainProcess|jsonrpc/5::DEBUG::2017-02-03 08:27:53,807::commands::93::root::(execCmd)
>>>> FAILED: <err> = 'ovs-vsctl: unix:/var/run/openvswitch/db.sock:
>>>> database connection failed (No such file or directory)\n'; <rc> = 1
>>>> MainProcess|jsonrpc/5::ERROR::2017-02-03 08:27:53,808::supervdsmServer:
>>>> :97::SuperVdsm.ServerCallback::(wrapper) Error in setupNetworks
>>>> Traceback (most recent call last):
>>>>   File "/usr/share/vdsm/supervdsmServer", line 95, in wrapper
>>>>     res = func(*args, **kwargs)
>>>>   File "/usr/lib/python2.7/site-packages/vdsm/network/api.py", line
>>>> 254, in setupNetworks
>>>>     netswitch.validate(networks, bondings)
>>>>   File "/usr/lib/python2.7/site-packages/vdsm/network/netswitch.py",
>>>> line 121, in validate
>>>>     ovs_switch.validate_network_setup(ovs_nets, ovs_bonds)
>>>>   File "/usr/lib/python2.7/site-packages/vdsm/network/ovs/switch.py",
>>>> line 40, in validate_network_setup
>>>>     ovs_networks = info.create_netinfo(info.OvsInfo())['networks']
>>>>   File "/usr/lib/python2.7/site-packages/vdsm/network/ovs/info.py",
>>>> line 73, in __init__
>>>>     ovs_db = OvsDB(driver.create())
>>>>   File "/usr/lib/python2.7/site-packages/vdsm/network/ovs/info.py",
>>>> line 64, in __init__
>>>>     transaction.add(ifaces_command)
>>>>   File "/usr/lib/python2.7/site-packages/vdsm/network/ovs/driver/__init__.py",
>>>> line 54, in __exit__
>>>>     self.result = self.commit()
>>>>   File "/usr/lib/python2.7/site-packages/vdsm/network/ovs/driver/vsctl.py",
>>>> line 63, in commit
>>>>     'Executing commands failed: %s' % '\n'.join(err))
>>>> ConfigNetworkError: (21, 'Executing commands failed: ovs-vsctl:
>>>> unix:/var/run/openvswitch/db.sock: database connection failed (No such
>>>> file or directory)')
>>>> MainProcess|Thread-2::DEBUG::2017-02-03 08:27:54,440::supervdsmServer:
>>>> :93::SuperVdsm.ServerCallback::(wrapper) call ksmTune with () {}
>>>> MainProcess|Thread-2::DEBUG::2017-02-03 08:27:54,440::supervdsmServer:
>>>> :100::SuperVdsm.ServerCallback::(wrapper) return ksmTune with None
>>>> MainProcess|jsonrpc/0::DEBUG::2017-02-03 08:30:00,918::supervdsmServer:
>>>> :93::SuperVdsm.ServerCallback::(wrapper) call network_caps with () {}
>>>> MainProcess|jsonrpc/0::DEBUG::2017-02-03 08:30:00,922::commands::69::root::(execCmd)
>>>> /usr/bin/taskset --cpu-list 0-47 /sbin/ip route show to 0.0.0.0/0
>>>> table main (cwd None)
>>>> MainProcess|jsonrpc/0::DEBUG::2017-02-03 08:30:00,927::commands::93::root::(execCmd)
>>>> SUCCESS: <err> = ''; <rc> = 0
>>>> MainProcess|jsonrpc/0::DEBUG::2017-02-03 08:30:01,013::commands::69::root::(execCmd)
>>>> /usr/bin/taskset --cpu-list 0-47 /usr/sbin/tc qdisc show (cwd None)
>>>> MainProcess|jsonrpc/0::DEBUG::2017-02-03 08:30:01,016::commands::93::root::(execCmd)
>>>> SUCCESS: <err> = ''; <rc> = 0
>>>> MainProcess|jsonrpc/0::DEBUG::2017-02-03 08:30:01,017::supervdsmServer:
>>>> :100::SuperVdsm.ServerCallback::(wrapper) return network_caps with
>>>> {'bridges': {}, 'bondings': {}, 'nameservers': ['10.1.1.8'], 'nics':
>>>> {'ib0': {'ipv6autoconf': True, 'addr': '', 'ipv6gateway': '::', 'dhcpv6':
>>>> False, 'ipv6addrs': [], 'mtu': '4092', 'dhcpv4': False, 'netmask': '',
>>>> 'ipv4defaultroute': False, 'ipv4addrs': [], 'hwaddr':
>>>> '80:00:02:08:fe:80:00:00:00:00:00:00:0c:c4:7a:ff:ff:0f:d8:85',
>>>> 'speed': 0, 'gateway': ''}, 'eno1': {'ipv6autoconf': False, 'addr':
>>>> '10.1.235.5', 'ipv6gateway': '::', 'dhcpv6': False, 'ipv6addrs': [], 'mtu':
>>>> '1500', 'dhcpv4': False, 'netmask': '255.255.0.0', 'ipv4defaultroute':
>>>> True, 'ipv4addrs': ['10.1.235.5/16'], 'hwaddr': '0c:c4:7a:0f:6b:2a',
>>>> 'speed': 1000, 'gateway': '10.1.0.1'}, 'eno2': {'ipv6autoconf': True,
>>>> 'addr': '', 'ipv6gateway': '::', 'dhcpv6': False, 'ipv6addrs': [], 'mtu':
>>>> '1500', 'dhcpv4': False, 'netmask': '', 'ipv4defaultroute': False,
>>>> 'ipv4addrs': [], 'hwaddr': '0c:c4:7a:0f:6b:2b', 'speed': 0, 'gateway':
>>>> ''}}, 'supportsIPv6': True, 'vlans': {}, 'networks': {}}
>>>> MainProcess|jsonrpc/1::DEBUG::2017-02-03 08:30:01,036::supervdsmServer:
>>>> :93::SuperVdsm.ServerCallback::(wrapper) call getHardwareInfo with ()
>>>> {}
>>>> MainProcess|jsonrpc/1::DEBUG::2017-02-03 08:30:01,066::supervdsmServer:
>>>> :100::SuperVdsm.ServerCallback::(wrapper) return getHardwareInfo with
>>>> {'systemProductName': 'X9DRT', 'systemSerialNumber': '0123456789',
>>>> 'systemFamily': 'To be filled by O.E.M.', 'systemVersion': '0123456789',
>>>> 'systemUUID': '00000000-0000-0000-0000-0CC47A0F6B2A',
>>>> 'systemManufacturer': 'Supermicro'}
>>>> (END)
>>>>
>>>>
>>>>
>>>> On Fri, Feb 3, 2017 at 12:05 AM, Edward Haas <ehaas at redhat.com> wrote:
>>>>
>>>>> Hello Shalabh,
>>>>>
>>>>> Please provide the logs from your node:
>>>>> - messages
>>>>> - vdsm/vdsm.log, vdsm/supervdsm.log
>>>>>
>>>>> It may be that you are missing openvswitch installed, although VDSM
>>>>> should not require it for its operation.
>>>>>
>>>>> Thanks,
>>>>> Edy.
>>>>>
>>>>>
>>>>> On Thu, Feb 2, 2017 at 2:10 PM, Shalabh Goel <shalabhgoel13 at gmail.com>
>>>>> wrote:
>>>>>
>>>>>> HI,
>>>>>>
>>>>>> I am getting the following error on my node after rebooting it.
>>>>>>
>>>>>> VDSM ovirtnode2 command HostSetupNetworksVDS failed: Executing
>>>>>> commands failed: ovs-vsctl: unix:/var/run/openvswitch/db.sock:
>>>>>> database connection failed (No such file or directory)
>>>>>>
>>>>>>
>>>>>> To solve this, I am trying to restart ovsdb-server using the
>>>>>> following command,
>>>>>>
>>>>>> ovsdb-server --remote=punix:*/var/run/openvswitch/db.sock*
>>>>>> --remote=db:Open_vSwitch,Open_vSwitch,manager_options
>>>>>> --private-key=db:Open_vSwitch,SSL,private_key--certificate=db:Open_vSwitch,SSL,certificate
>>>>>> --bootstrap-ca-cert=db:Open_vSwitch,SSL,ca_cert --pidfile --detach
>>>>>>
>>>>>> But I am getting the following error.
>>>>>>
>>>>>> ovsdb-server: /var/run/openvswitch/ovsdb-server.pid.tmp: create
>>>>>> failed (No such file or directory)
>>>>>>
>>>>>> How to restart the ovsdb-server?? Also ovirtmgmt network is missing
>>>>>> from my node. It happened after I rebooted my node after it got upgraded to
>>>>>> Ovirt 4.1
>>>>>>
>>>>>> --
>>>>>> Shalabh Goel
>>>>>>
>>>>>> _______________________________________________
>>>>>> Users mailing list
>>>>>> Users at ovirt.org
>>>>>> http://lists.ovirt.org/mailman/listinfo/users
>>>>>>
>>>>>>
>>>>>
>>>>
>>>>
>>>> --
>>>> Shalabh Goel
>>>>
>>>
>>>
>>
>>
>> --
>> Shalabh Goel
>>
>
>


-- 
Shalabh Goel
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20170206/2d16b6d5/attachment-0001.html>


More information about the Users mailing list