Hi Strahil,
I remember during after creating the volume I applied the virt group to it.
Volume info:
----------------
Volume Name: data
Type: Replicate
Volume ID: 05842cd6-7f16-4329-9ffd-64a0b4366fbe
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: host1storage:/gluster_bricks/data/data
Brick2: host2storage:/gluster_bricks/data/data
Brick3: host3storage:/gluster_bricks/data/data
Options Reconfigured:
performance.client-io-threads: on
nfs.disable: on
transport.address-family: inet
storage.owner-gid: 36
storage.owner-uid: 36
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.low-prio-threads: 32
network.remote-dio: enable
cluster.eager-lock: enable
cluster.quorum-type: auto
cluster.server-quorum-type: server
cluster.data-self-heal-algorithm: full
cluster.locking-scheme: granular
cluster.shd-max-threads: 8
cluster.shd-wait-qlength: 10000
features.shard: on
user.cifs: off
cluster.choose-local: off
client.event-threads: 4
server.event-threads: 4
I do not have full logs but I have some saved.
/var/log/messages:
-------------------------
Sep 14 08:36:20 host1 vdsm[4301]: ERROR Unhandled exception in <Task discardable
<UpdateVolumes vm=fd5123cb-9364-448d-b41c-8a48fb1826c5 at 0x7f1244099050>
timeout=30.0, duration=0.01 at 0x7f1244099210>#012Traceback (most recent call
last):#012 File "/usr/lib/python2.7/site-packages/vdsm/executor.py", line 315,
in _execute_task#012 task()#012 File
"/usr/lib/python2.7/site-packages/vdsm/executor.py", line 391, in __call__#012
self._callable()#012 File
"/usr/lib/python2.7/site-packages/vdsm/virt/periodic.py", line 315, in
__call__#012 self._execute()#012 File
"/usr/lib/python2.7/site-packages/vdsm/virt/periodic.py", line 357, in
_execute#012 self._vm.updateDriveVolume(drive)#012 File
"/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 4189, in
updateDriveVolume#012 vmDrive.volumeID)#012 File
"/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 6101, in
_getVolumeSize#012 (domainID, volumeID))#012StorageUnavailableError: Unable to get
volume size for doma
in 88f5972f-58bd-469f-bc77-5bf3b1802291 volume cdf313d7-bed3-4fae-a803-1297cdf8c82f
Sep 14 08:37:20 host1 vdsm[4301]: ERROR Unhandled exception in <Task discardable
<UpdateVolumes vm=fd5123cb-9364-448d-b41c-8a48fb1826c5 at 0x7f12442efad0>
timeout=30.0, duration=0.00 at 0x7f1244078490>#012Traceback (most recent call
last):#012 File "/usr/lib/python2.7/site-packages/vdsm/executor.py", line 315,
in _execute_task#012 task()#012 File
"/usr/lib/python2.7/site-packages/vdsm/executor.py", line 391, in __call__#012
self._callable()#012 File
"/usr/lib/python2.7/site-packages/vdsm/virt/periodic.py", line 315, in
__call__#012 self._execute()#012 File
"/usr/lib/python2.7/site-packages/vdsm/virt/periodic.py", line 357, in
_execute#012 self._vm.updateDriveVolume(drive)#012 File
"/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 4189, in
updateDriveVolume#012 vmDrive.volumeID)#012 File
"/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 6101, in
_getVolumeSize#012 (domainID, volumeID))#012StorageUnavailableError: Unable to get
volume size for doma
in 88f5972f-58bd-469f-bc77-5bf3b1802291 volume cdf313d7-bed3-4fae-a803-1297cdf8c82f
Sep 14 08:38:20 host1 vdsm[4301]: ERROR Unhandled exception in <Task discardable
<UpdateVolumes vm=fd5123cb-9364-448d-b41c-8a48fb1826c5 at 0x7f12045aa550>
timeout=30.0, duration=0.00 at 0x7f12045aaa90>#012Traceback (most recent call
last):#012 File "/usr/lib/python2.7/site-packages/vdsm/executor.py", line 315,
in _execute_task#012 task()#012 File
"/usr/lib/python2.7/site-packages/vdsm/executor.py", line 391, in __call__#012
self._callable()#012 File
"/usr/lib/python2.7/site-packages/vdsm/virt/periodic.py", line 315, in
__call__#012 self._execute()#012 File
"/usr/lib/python2.7/site-packages/vdsm/virt/periodic.py", line 357, in
_execute#012 self._vm.updateDriveVolume(drive)#012 File
"/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 4189, in
updateDriveVolume#012 vmDrive.volumeID)#012 File
"/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 6101, in
_getVolumeSize#012 (domainID, volumeID))#012StorageUnavailableError: Unable to get
volume size for doma
in 88f5972f-58bd-469f-bc77-5bf3b1802291 volume cdf313d7-bed3-4fae-a803-1297cdf8c82f
Sep 14 08:39:20 host1 vdsm[4301]: ERROR Unhandled exception in <Task discardable
<UpdateVolumes vm=fd5123cb-9364-448d-b41c-8a48fb1826c5 at 0x7f12441d6f50>
timeout=30.0, duration=0.01 at 0x7f1287f189d0>#012Traceback (most recent call
last):#012 File "/usr/lib/python2.7/site-packages/vdsm/executor.py", line 315,
in _execute_task#012 task()#012 File
"/usr/lib/python2.7/site-packages/vdsm/executor.py", line 391, in __call__#012
self._callable()#012 File
"/usr/lib/python2.7/site-packages/vdsm/virt/periodic.py", line 315, in
__call__#012 self._execute()#012 File
"/usr/lib/python2.7/site-packages/vdsm/virt/periodic.py", line 357, in
_execute#012 self._vm.updateDriveVolume(drive)#012 File
"/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 4189, in
updateDriveVolume#012 vmDrive.volumeID)#012 File
"/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 6101, in
_getVolumeSize#012 (domainID, volumeID))#012StorageUnavailableError: Unable to get
volume size for doma
in 88f5972f-58bd-469f-bc77-5bf3b1802291 volume cdf313d7-bed3-4fae-a803-1297cdf8c82f
Sep 14 08:40:01 host1 systemd: Created slice User Slice of root.
Sep 14 08:40:01 host1 systemd: Started Session 52513 of user root.
Sep 14 08:40:01 host1 systemd: Removed slice User Slice of root.
Sep 14 08:40:20 host1 vdsm[4301]: ERROR Unhandled exception in <Task discardable
<UpdateVolumes vm=fd5123cb-9364-448d-b41c-8a48fb1826c5 at 0x7f12442ef290>
timeout=30.0, duration=0.00 at 0x7f11e73d5b50>#012Traceback (most recent call
last):#012 File "/usr/lib/python2.7/site-packages/vdsm/executor.py", line 315,
in _execute_task#012 task()#012 File
"/usr/lib/python2.7/site-packages/vdsm/executor.py", line 391, in __call__#012
self._callable()#012 File
"/usr/lib/python2.7/site-packages/vdsm/virt/periodic.py", line 315, in
__call__#012 self._execute()#012 File
"/usr/lib/python2.7/site-packages/vdsm/virt/periodic.py", line 357, in
_execute#012 self._vm.updateDriveVolume(drive)#012 File
"/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 4189, in
updateDriveVolume#012 vmDrive.volumeID)#012 File
"/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 6101, in
_getVolumeSize#012 (domainID, volumeID))#012StorageUnavailableError: Unable to get
volume size for doma
in 88f5972f-58bd-469f-bc77-5bf3b1802291 volume cdf313d7-bed3-4fae-a803-1297cdf8c82f
Sep 14 08:40:42 host1 systemd: Created slice User Slice of root.
Sep 14 08:40:42 host1 systemd: Started Session c69119 of user root.
Sep 14 08:40:42 host1 systemd: Removed slice User Slice of root.
rhev-data-center-mnt-glusterSD-host1storage::_data.log
----------------------------------------------------------------
2020-09-14 08:40:39.648159] I [MSGID: 114018] [client.c:2331:client_rpc_notify]
0-data-client-2: disconnected from data-client-2. Client process will keep trying to
connect to glusterd until brick's port is available
[2020-09-14 08:40:39.648183] W [MSGID: 108001] [afr-common.c:5613:afr_notify]
0-data-replicate-0: Client-quorum is not met
[2020-09-14 08:40:50.243632] E [MSGID: 114058]
[client-handshake.c:1449:client_query_portmap_cbk] 0-data-client-2: failed to get the port
number for remote subvolume. Please run 'gluster volume status' on server to see
if brick process is running.
[2020-09-14 08:40:50.243686] I [MSGID: 114018] [client.c:2331:client_rpc_notify]
0-data-client-2: disconnected from data-client-2. Client process will keep trying to
connect to glusterd until brick's port is available
[2020-09-14 08:41:05.147093] I [MSGID: 114018] [client.c:2331:client_rpc_notify]
0-data-client-1: disconnected from data-client-1. Client process will keep trying to
connect to glusterd until brick's port is available
[2020-09-14 08:41:05.147139] E [MSGID: 108006]
[afr-common.c:5323:__afr_handle_child_down_event] 0-data-replicate-0: All subvolumes are
down. Going offline until at least one of them comes back up.
[2020-09-14 08:41:05.156727] I [MSGID: 108006] [afr-common.c:5669:afr_local_init]
0-data-replicate-0: no subvolumes up
The message "I [MSGID: 108006] [afr-common.c:5669:afr_local_init] 0-data-replicate-0:
no subvolumes up" repeated 195 times between [2020-09-14 08:41:05.156727] and
[2020-09-14 08:41:15.222500]
[2020-09-14 08:41:15.288560] E [MSGID: 114058]
[client-handshake.c:1449:client_query_portmap_cbk] 0-data-client-1: failed to get the port
number for remote subvolume. Please run 'gluster volume status' on server to see
if brick process is running.
[2020-09-14 08:41:15.288608] I [MSGID: 114018] [client.c:2331:client_rpc_notify]
0-data-client-1: disconnected from data-client-1. Client process will keep trying to
connect to glusterd until brick's port is available
[2020-09-14 08:41:15.722793] I [MSGID: 108006] [afr-common.c:5669:afr_local_init]
0-data-replicate-0: no subvolumes up
The message "I [MSGID: 108006] [afr-common.c:5669:afr_local_init] 0-data-replicate-0:
no subvolumes up" repeated 170 times between [2020-09-14 08:41:15.722793] and
[2020-09-14 08:41:24.230828]
[2020-09-14 08:41:24.731222] I [MSGID: 108006] [afr-common.c:5669:afr_local_init]
0-data-replicate-0: no subvolumes up
[2020-09-14 08:41:41.352791] I [socket.c:811:__socket_shutdown] 0-data-client-2:
intentional socket shutdown(6)
[2020-09-14 08:41:44.363054] I [socket.c:811:__socket_shutdown] 0-data-client-0:
intentional socket shutdown(6)
[2020-09-14 08:42:42.512364] I [socket.c:811:__socket_shutdown] 0-data-client-1:
intentional socket shutdown(6)
The message "I [MSGID: 108006] [afr-common.c:5669:afr_local_init] 0-data-replicate-0:
no subvolumes up" repeated 708 times between [2020-09-14 08:41:24.731222] and
[2020-09-14 08:43:23.874840]
[2020-09-14 08:43:26.411190] I [MSGID: 108006] [afr-common.c:5669:afr_local_init]
0-data-replicate-0: no subvolumes up
...
...
...
[2020-09-14 12:19:59.517980] I [MSGID: 108006] [afr-common.c:5669:afr_local_init]
0-data-replicate-0: no subvolumes up
[2020-09-14 12:20:29.093669] W [socket.c:721:__socket_rwv] 0-glusterfs: readv on
192.168.0.101:24007 failed (No data available)
[2020-09-14 12:20:29.093695] I [glusterfsd-mgmt.c:2443:mgmt_rpc_notify] 0-glusterfsd-mgmt:
disconnected from remote-host:
host1storage.mydomain.com
[2020-09-14 12:20:29.093707] I [glusterfsd-mgmt.c:2483:mgmt_rpc_notify] 0-glusterfsd-mgmt:
connecting to next volfile server
host2.mydomain.com
[2020-09-14 12:20:39.630645] I [socket.c:811:__socket_shutdown] 0-data-client-2:
intentional socket shutdown(7)
[2020-09-14 12:20:40.635471] I [rpc-clnt.c:2028:rpc_clnt_reconfig] 0-data-client-1:
changing port to 49152 (from 0)
[2020-09-14 12:20:40.639281] I [MSGID: 114046]
[client-handshake.c:1106:client_setvolume_cbk] 0-data-client-1: Connected to
data-client-1, attached to remote volume '/gluster_bricks/data/data'.
[2020-09-14 12:20:40.639299] I [MSGID: 114042]
[client-handshake.c:930:client_post_handshake] 0-data-client-1: 5 fds open - Delaying
child_up until they are re-opened
[2020-09-14 12:20:40.639945] I [MSGID: 114041]
[client-handshake.c:318:client_child_up_reopen_done] 0-data-client-1: last fd
open'd/lock-self-heal'd - notifying CHILD-UP
[2020-09-14 12:20:40.639967] I [MSGID: 108005]
[afr-common.c:5245:__afr_handle_child_up_event] 0-data-replicate-0: Subvolume
'data-client-1' came back up; going online.
[2020-09-14 12:20:42.640257] I [rpc-clnt.c:2028:rpc_clnt_reconfig] 0-data-client-0:
changing port to 49152 (from 0)
[2020-09-14 12:20:42.640288] I [socket.c:811:__socket_shutdown] 0-data-client-0:
intentional socket shutdown(10)
[2020-09-14 12:20:42.643664] I [MSGID: 114046]
[client-handshake.c:1106:client_setvolume_cbk] 0-data-client-0: Connected to
data-client-0, attached to remote volume '/gluster_bricks/data/data'.
[2020-09-14 12:20:42.643683] I [MSGID: 114042]
[client-handshake.c:930:client_post_handshake] 0-data-client-0: 5 fds open - Delaying
child_up until they are re-opened
[2020-09-14 12:20:42.644327] I [MSGID: 114041]
[client-handshake.c:318:client_child_up_reopen_done] 0-data-client-0: last fd
open'd/lock-self-heal'd - notifying CHILD-UP
[2020-09-14 12:20:42.644347] I [MSGID: 108002] [afr-common.c:5607:afr_notify]
0-data-replicate-0: Client-quorum is met
[2020-09-14 12:20:42.843176] I [rpc-clnt.c:2028:rpc_clnt_reconfig] 0-data-client-2:
changing port to 49152 (from 0)
[2020-09-14 12:20:42.846562] I [MSGID: 114046]
[client-handshake.c:1106:client_setvolume_cbk] 0-data-client-2: Connected to
data-client-2, attached to remote volume '/gluster_bricks/data/data'.
[2020-09-14 12:20:42.846598] I [MSGID: 114042]
[client-handshake.c:930:client_post_handshake] 0-data-client-2: 5 fds open - Delaying
child_up until they are re-opened
[2020-09-14 12:20:42.847429] I [MSGID: 114041]
[client-handshake.c:318:client_child_up_reopen_done] 0-data-client-2: last fd
open'd/lock-self-heal'd - notifying CHILD-UP
[2020-09-14 12:20:49.640811] I [MSGID: 133022] [shard.c:3674:shard_delete_shards]
0-data-shard: Deleted shards of gfid=18250f19-3820-4a98-9c49-37ba23c08dfd from backend
[2020-09-14 12:20:50.244771] E [MSGID: 114031]
[client-rpc-fops_v2.c:2540:client4_0_opendir_cbk] 0-data-client-1: remote operation
failed. Path: /88f5972f-58bd-469f-bc77-5bf3b1802291/images
(5fdbb512-e924-4945-a633-10820133e5ff) [Permission denied]
[2020-09-14 12:20:50.244829] E [MSGID: 114031]
[client-rpc-fops_v2.c:2540:client4_0_opendir_cbk] 0-data-client-2: remote operation
failed. Path: /88f5972f-58bd-469f-bc77-5bf3b1802291/images
(5fdbb512-e924-4945-a633-10820133e5ff) [Permission denied]
[2020-09-14 12:20:50.244880] W [MSGID: 114061]
[client-common.c:3325:client_pre_readdirp_v2] 0-data-client-1:
(5fdbb512-e924-4945-a633-10820133e5ff) remote_fd is -1. EBADFD [File descriptor in bad
state]
[2020-09-14 12:21:20.364907] W [MSGID: 114031]
[client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data-client-1: remote operation failed.
Path: /88f5972f-58bd-469f-bc77-5bf3b1802291/images/4e79467c-4707-4e7d-8a2a-909b208d4b97
(b0f8bb67-09e5-431b-acea-c03f0280fb34) [Permission denied]
[2020-09-14 12:21:20.365181] W [MSGID: 114031]
[client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data-client-2: remote operation failed.
Path: /88f5972f-58bd-469f-bc77-5bf3b1802291/images/4e79467c-4707-4e7d-8a2a-909b208d4b97
(b0f8bb67-09e5-431b-acea-c03f0280fb34) [Permission denied]
[2020-09-14 12:21:20.368381] W [MSGID: 114031]
[client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data-client-2: remote operation failed.
Path: (null) (00000000-0000-0000-0000-000000000000) [Permission denied]
[2020-09-14 12:21:20.368401] W [MSGID: 114031]
[client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data-client-1: remote operation failed.
Path: (null) (00000000-0000-0000-0000-000000000000) [Permission denied]
[2020-09-14 12:21:20.368462] W [MSGID: 108027]
[afr-common.c:2274:afr_attempt_readsubvol_set] 0-data-replicate-0: no read subvols for
/88f5972f-58bd-469f-bc77-5bf3b1802291/images/4e79467c-4707-4e7d-8a2a-909b208d4b97
[2020-09-14 12:21:20.369001] W [MSGID: 114031]
[client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data-client-1: remote operation failed.
Path: /88f5972f-58bd-469f-bc77-5bf3b1802291/images/4e79467c-4707-4e7d-8a2a-909b208d4b97
(00000000-0000-0000-0000-000000000000) [Permission denied]
[2020-09-14 12:21:20.369002] W [MSGID: 114031]
[client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data-client-2: remote operation failed.
Path: /88f5972f-58bd-469f-bc77-5bf3b1802291/images/4e79467c-4707-4e7d-8a2a-909b208d4b97
(00000000-0000-0000-0000-000000000000) [Permission denied]
[2020-09-14 12:21:20.370979] W [MSGID: 114031]
[client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data-client-2: remote operation failed.
Path: (null) (00000000-0000-0000-0000-000000000000) [Permission denied]
[2020-09-14 12:21:20.371009] W [MSGID: 114031]
[client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data-client-1: remote operation failed.
Path: (null) (00000000-0000-0000-0000-000000000000) [Permission denied]
[2020-09-14 12:21:20.371075] W [MSGID: 108027]
[afr-common.c:2274:afr_attempt_readsubvol_set] 0-data-replicate-0: no read subvols for
/88f5972f-58bd-469f-bc77-5bf3b1802291/images/4e79467c-4707-4e7d-8a2a-909b208d4b97
The message "I [MSGID: 108006] [afr-common.c:5669:afr_local_init] 0-data-replicate-0:
no subvolumes up" repeated 102 times between [2020-09-14 12:19:59.517980] and
[2020-09-14 12:20:39.539124]
[2020-09-14 12:22:11.510751] I [MSGID: 133022] [shard.c:3674:shard_delete_shards]
0-data-shard: Deleted shards of gfid=40104663-41d6-4210-b9aa-065e0ba48c1f from backend
[2020-09-14 12:22:11.853348] E [MSGID: 114031]
[client-rpc-fops_v2.c:2540:client4_0_opendir_cbk] 0-data-client-2: remote operation
failed. Path: /88f5972f-58bd-469f-bc77-5bf3b1802291/images
(5fdbb512-e924-4945-a633-10820133e5ff) [Permission denied]
glusterd.log
----------------
[2020-09-14 08:40:39.750074] E [socket.c:2282:__socket_read_frag] 0-rpc: wrong MSG-TYPE
(9) received from 10.0.0.116:30052
[2020-09-14 08:42:36.920865] C [rpcsvc.c:1029:rpcsvc_notify] 0-rpcsvc: got MAP_XID event,
which should have not come
[2020-09-14 08:42:56.426832] E [socket.c:2282:__socket_read_frag] 0-rpc: wrong MSG-TYPE
(3866624) received from 10.0.0.116:53501
[2020-09-14 08:48:50.776839] E [socket.c:2282:__socket_read_frag] 0-rpc: wrong MSG-TYPE
(-66911352) received from 10.0.0.116:46153
[2020-09-14 09:00:51.692183] E [socket.c:2282:__socket_read_frag] 0-rpc: wrong MSG-TYPE
(7602176) received from 10.0.0.116:64053
[2020-09-14 12:17:50.501625] W [MSGID: 101095] [xlator.c:210:xlator_volopt_dynload]
0-xlator: /usr/lib64/glusterfs/6.6/xlator/encryption/crypt.so: cannot open shared object
file: No such file or directory
[2020-09-14 12:17:50.508662] E [MSGID: 101097] [xlator.c:218:xlator_volopt_dynload]
0-xlator: dlsym(xlator_api) missing: /usr/lib64/glusterfs/6.6/rpc-transport/socket.so:
undefined symbol: xlator_api
[2020-09-14 12:17:50.510694] W [MSGID: 101095] [xlator.c:210:xlator_volopt_dynload]
0-xlator: /usr/lib64/glusterfs/6.6/xlator/nfs/server.so: cannot open shared object file:
No such file or directory
[2020-09-14 12:17:50.518341] W [MSGID: 101095] [xlator.c:210:xlator_volopt_dynload]
0-xlator: /usr/lib64/glusterfs/6.6/xlator/storage/bd.so: cannot open shared object file:
No such file or directory
[2020-09-14 12:18:08.102606] I [MSGID: 106499]
[glusterd-handler.c:4429:__glusterd_handle_status_volume] 0-management: Received status
volume req for volume data
[2020-09-14 12:18:08.118107] I [MSGID: 106499]
[glusterd-handler.c:4429:__glusterd_handle_status_volume] 0-management: Received status
volume req for volume engine
[2020-09-14 12:18:37.847581] E [MSGID: 106537]
[glusterd-volume-ops.c:1763:glusterd_op_stage_start_volume] 0-management: Volume engine
already started
[2020-09-14 12:18:37.847606] W [MSGID: 106121]
[glusterd-mgmt.c:178:gd_mgmt_v3_pre_validate_fn] 0-management: Volume start prevalidation
failed.
[2020-09-14 12:18:37.847618] E [MSGID: 106121]
[glusterd-mgmt.c:1079:glusterd_mgmt_v3_pre_validate] 0-management: Pre Validation failed
for operation Start on local node
[2020-09-14 12:18:37.847625] E [MSGID: 106121]
[glusterd-mgmt.c:2457:glusterd_mgmt_v3_initiate_all_phases] 0-management: Pre Validation
Failed
The message "W [MSGID: 101095] [xlator.c:210:xlator_volopt_dynload] 0-xlator:
/usr/lib64/glusterfs/6.6/xlator/encryption/crypt.so: cannot open shared object file: No
such file or directory" repeated 2 times between [2020-09-14 12:17:50.501625] and
[2020-09-14 12:17:50.501663]
The message "E [MSGID: 101097] [xlator.c:218:xlator_volopt_dynload] 0-xlator:
dlsym(xlator_api) missing: /usr/lib64/glusterfs/6.6/rpc-transport/socket.so: undefined
symbol: xlator_api" repeated 7 times between [2020-09-14 12:17:50.508662] and
[2020-09-14 12:17:50.508709]
The message "W [MSGID: 101095] [xlator.c:210:xlator_volopt_dynload] 0-xlator:
/usr/lib64/glusterfs/6.6/xlator/nfs/server.so: cannot open shared object file: No such
file or directory" repeated 30 times between [2020-09-14 12:17:50.510694] and
[2020-09-14 12:17:50.510909]
[2020-09-14 12:19:41.946661] I [MSGID: 106487]
[glusterd-handler.c:1516:__glusterd_handle_cli_list_friends] 0-glusterd: Received cli list
req
[2020-09-14 12:19:49.871903] I [MSGID: 106533]
[glusterd-volume-ops.c:982:__glusterd_handle_cli_heal_volume] 0-management: Received heal
vol req for volume engine
[2020-09-14 12:20:29.091554] W [glusterfsd.c:1570:cleanup_and_exit]
(-->/lib64/libpthread.so.0(+0x7e65) [0x7fabf0a81e65]
-->/usr/sbin/glusterd(glusterfs_sigwaiter+0xe5) [0x5643e96971f5]
-->/usr/sbin/glusterd(cleanup_and_exit+0x6b) [0x5643e969705b] ) 0-: received signum
(15), shutting down
[2020-09-14 12:20:29.123582] I [MSGID: 100030] [glusterfsd.c:2847:main]
0-/usr/sbin/glusterd: Started running /usr/sbin/glusterd version 6.6 (args:
/usr/sbin/glusterd -p /var/run/glusterd.pid --log-level INFO)
[2020-09-14 12:20:29.124349] I [glusterfsd.c:2556:daemonize] 0-glusterfs: Pid of current
running process is 33290
[2020-09-14 12:20:29.129801] I [MSGID: 106478] [glusterd.c:1422:init] 0-management:
Maximum allowed open file descriptors set to 65536
[2020-09-14 12:20:29.129827] I [MSGID: 106479] [glusterd.c:1478:init] 0-management: Using
/var/lib/glusterd as working directory
[2020-09-14 12:20:29.129834] I [MSGID: 106479] [glusterd.c:1484:init] 0-management: Using
/var/run/gluster as pid file working directory
[2020-09-14 12:20:29.133500] I [socket.c:961:__socket_server_bind] 0-socket.management:
process started listening on port (24007)
[2020-09-14 12:20:29.136005] W [MSGID: 103071] [rdma.c:4472:__gf_rdma_ctx_create]
0-rpc-transport/rdma: rdma_cm event channel creation failed [No such device]
[2020-09-14 12:20:29.136024] W [MSGID: 103055] [rdma.c:4782:init] 0-rdma.management:
Failed to initialize IB Device
[2020-09-14 12:20:29.136032] W [rpc-transport.c:363:rpc_transport_load] 0-rpc-transport:
'rdma' initialization failed
[2020-09-14 12:20:29.136102] W [rpcsvc.c:1985:rpcsvc_create_listener] 0-rpc-service:
cannot create listener, initing the transport failed
[2020-09-14 12:20:29.136112] E [MSGID: 106244] [glusterd.c:1785:init] 0-management:
creation of 1 listeners failed, continuing with succeeded transport
[2020-09-14 12:20:29.137180] I [socket.c:904:__socket_server_bind] 0-socket.management:
closing (AF_UNIX) reuse check socket 12
[2020-09-14 12:20:29.137478] I [MSGID: 106059] [glusterd.c:1865:init] 0-management:
max-port override: 60999
[2020-09-14 12:20:30.969620] I [MSGID: 106513]
[glusterd-store.c:2394:glusterd_restore_op_version] 0-glusterd: retrieved op-version:
60000
[2020-09-14 12:20:31.061282] I [MSGID: 106544] [glusterd.c:152:glusterd_uuid_init]
0-management: retrieved UUID: 6fe3d6e3-ab45-4004-af59-93c2fc3afc93
[2020-09-14 12:20:31.086867] I [MSGID: 106498]
[glusterd-handler.c:3687:glusterd_friend_add_from_peerinfo] 0-management: connect returned
0
[2020-09-14 12:20:31.087760] I [MSGID: 106498]
[glusterd-handler.c:3687:glusterd_friend_add_from_peerinfo] 0-management: connect returned
0
[2020-09-14 12:20:31.087814] W [MSGID: 106061]
[glusterd-handler.c:3490:glusterd_transport_inet_options_build] 0-glusterd: Failed to get
tcp-user-timeout
[2020-09-14 12:20:31.087844] I [rpc-clnt.c:1005:rpc_clnt_connection_init] 0-management:
setting frame-timeout to 600
[2020-09-14 12:20:31.091059] I [rpc-clnt.c:1005:rpc_clnt_connection_init] 0-management:
setting frame-timeout to 600
[2020-09-14 12:20:31.091052] W [MSGID: 106061]
[glusterd-handler.c:3490:glusterd_transport_inet_options_build] 0-glusterd: Failed to get
tcp-user-timeout
[2020-09-14 12:20:31.095394] I [MSGID: 101190]
[event-epoll.c:680:event_dispatch_epoll_worker] 0-epoll: Started thread with index 0
[2020-09-14 12:20:40.073223] I [MSGID: 106493]
[glusterd-rpc-ops.c:468:__glusterd_friend_add_cbk] 0-glusterd: Received ACC from uuid:
1c2f1776-4830-4b6f-8e9e-c29d0133a0e9, host:
myhost3.mydomain.com, port: 0
[2020-09-14 12:20:40.076519] C [MSGID: 106003]
[glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action] 0-management: Server
quorum regained for volume data. Starting local bricks.
[2020-09-14 12:20:40.077376] I [glusterd-utils.c:6312:glusterd_brick_start] 0-management:
starting a fresh brick process for brick /gluster_bricks/data/data
[2020-09-14 12:20:40.080459] I [rpc-clnt.c:1005:rpc_clnt_connection_init] 0-management:
setting frame-timeout to 600
[2020-09-14 12:20:40.093965] C [MSGID: 106003]
[glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action] 0-management: Server
quorum regained for volume engine. Starting local bricks.
[2020-09-14 12:20:40.094129] I [glusterd-utils.c:6312:glusterd_brick_start] 0-management:
starting a fresh brick process for brick /gluster_bricks/engine/engine
[2020-09-14 12:20:40.095873] I [rpc-clnt.c:1005:rpc_clnt_connection_init] 0-management:
setting frame-timeout to 600
[2020-09-14 12:20:40.112487] I [rpc-clnt.c:1005:rpc_clnt_connection_init] 0-management:
setting frame-timeout to 600
[2020-09-14 12:20:40.125181] I [MSGID: 106493]
[glusterd-rpc-ops.c:468:__glusterd_friend_add_cbk] 0-glusterd: Received ACC from uuid:
b5a0980f-015b-4720-9b5b-7792ac022b54, host:
myhost2.mydomain.com, port: 0
[2020-09-14 12:20:40.163361] I [rpc-clnt.c:1005:rpc_clnt_connection_init] 0-nfs: setting
frame-timeout to 600
[2020-09-14 12:20:40.163464] I [MSGID: 106131]
[glusterd-proc-mgmt.c:86:glusterd_proc_stop] 0-management: nfs already stopped
[2020-09-14 12:20:40.163482] I [MSGID: 106568] [glusterd-svc-mgmt.c:253:glusterd_svc_stop]
0-management: nfs service is stopped
[2020-09-14 12:20:40.163493] I [MSGID: 106599]
[glusterd-nfs-svc.c:81:glusterd_nfssvc_manager] 0-management: nfs/server.so xlator is not
installed
[2020-09-14 12:20:40.163520] I [rpc-clnt.c:1005:rpc_clnt_connection_init] 0-glustershd:
setting frame-timeout to 600
[2020-09-14 12:20:40.165762] I [MSGID: 106568]
[glusterd-proc-mgmt.c:92:glusterd_proc_stop] 0-management: Stopping glustershd daemon
running in pid: 28027
[2020-09-14 12:20:40.170812] I [MSGID: 106492]
[glusterd-handler.c:2796:__glusterd_handle_friend_update] 0-glusterd: Received friend
update from uuid: 1c2f1776-4830-4b6f-8e9e-c29d0133a0e9
[2020-09-14 12:20:40.170841] I [MSGID: 106502]
[glusterd-handler.c:2837:__glusterd_handle_friend_update] 0-management: Received my uuid
as Friend
[2020-09-14 12:20:40.175230] I [MSGID: 106493]
[glusterd-rpc-ops.c:681:__glusterd_friend_update_cbk] 0-management: Received ACC from
uuid: b5a0980f-015b-4720-9b5b-7792ac022b54
[2020-09-14 12:20:40.175329] I [MSGID: 106492]
[glusterd-handler.c:2796:__glusterd_handle_friend_update] 0-glusterd: Received friend
update from uuid: b5a0980f-015b-4720-9b5b-7792ac022b54
[2020-09-14 12:20:40.175353] I [MSGID: 106502]
[glusterd-handler.c:2837:__glusterd_handle_friend_update] 0-management: Received my uuid
as Friend
[2020-09-14 12:20:40.183400] I [MSGID: 106163]
[glusterd-handshake.c:1389:__glusterd_mgmt_hndsk_versions_ack] 0-management: using the
op-version 60000
[2020-09-14 12:20:40.188715] I [MSGID: 106490]
[glusterd-handler.c:2611:__glusterd_handle_incoming_friend_req] 0-glusterd: Received probe
from uuid: b5a0980f-015b-4720-9b5b-7792ac022b54
[2020-09-14 12:20:40.193866] I [MSGID: 106493]
[glusterd-handler.c:3883:glusterd_xfer_friend_add_resp] 0-glusterd: Responded to
myhost2.mydomain.com (0), ret: 0, op_ret: 0
[2020-09-14 12:20:40.198118] I [MSGID: 106492]
[glusterd-handler.c:2796:__glusterd_handle_friend_update] 0-glusterd: Received friend
update from uuid: b5a0980f-015b-4720-9b5b-7792ac022b54
[2020-09-14 12:20:40.198144] I [MSGID: 106502]
[glusterd-handler.c:2837:__glusterd_handle_friend_update] 0-management: Received my uuid
as Friend
[2020-09-14 12:20:40.200072] I [MSGID: 106142] [glusterd-pmap.c:290:pmap_registry_bind]
0-pmap: adding brick /gluster_bricks/data/data on port 49152
[2020-09-14 12:20:40.200198] I [MSGID: 106142] [glusterd-pmap.c:290:pmap_registry_bind]
0-pmap: adding brick /gluster_bricks/engine/engine on port 49153
[2020-09-14 12:20:40.200245] I [MSGID: 106493]
[glusterd-rpc-ops.c:681:__glusterd_friend_update_cbk] 0-management: Received ACC from
uuid: b5a0980f-015b-4720-9b5b-7792ac022b54
[2020-09-14 12:20:40.875962] I [MSGID: 106493]
[glusterd-rpc-ops.c:681:__glusterd_friend_update_cbk] 0-management: Received ACC from
uuid: 1c2f1776-4830-4b6f-8e9e-c29d0133a0e9
[2020-09-14 12:20:40.888913] I [MSGID: 106163]
[glusterd-handshake.c:1389:__glusterd_mgmt_hndsk_versions_ack] 0-management: using the
op-version 60000
[2020-09-14 12:20:40.892888] I [MSGID: 106490]
[glusterd-handler.c:2611:__glusterd_handle_incoming_friend_req] 0-glusterd: Received probe
from uuid: 1c2f1776-4830-4b6f-8e9e-c29d0133a0e9
[2020-09-14 12:20:40.896510] I [MSGID: 106493]
[glusterd-handler.c:3883:glusterd_xfer_friend_add_resp] 0-glusterd: Responded to
myhost3.mydomain.com (0), ret: 0, op_ret: 0
[2020-09-14 12:20:40.928796] I [MSGID: 106492]
[glusterd-handler.c:2796:__glusterd_handle_friend_update] 0-glusterd: Received friend
update from uuid: 1c2f1776-4830-4b6f-8e9e-c29d0133a0e9
[2020-09-14 12:20:40.928820] I [MSGID: 106502]
[glusterd-handler.c:2837:__glusterd_handle_friend_update] 0-management: Received my uuid
as Friend
[2020-09-14 12:20:40.997183] I [MSGID: 106493]
[glusterd-rpc-ops.c:681:__glusterd_friend_update_cbk] 0-management: Received ACC from
uuid: 1c2f1776-4830-4b6f-8e9e-c29d0133a0e9
[2020-09-14 12:20:41.165913] I [MSGID: 106568] [glusterd-svc-mgmt.c:253:glusterd_svc_stop]
0-management: glustershd service is stopped
[2020-09-14 12:20:41.165965] I [MSGID: 106567]
[glusterd-svc-mgmt.c:220:glusterd_svc_start] 0-management: Starting glustershd service
[2020-09-14 12:20:42.168619] I [rpc-clnt.c:1005:rpc_clnt_connection_init] 0-quotad:
setting frame-timeout to 600
[2020-09-14 12:20:42.168901] I [MSGID: 106131]
[glusterd-proc-mgmt.c:86:glusterd_proc_stop] 0-management: quotad already stopped
[2020-09-14 12:20:42.168917] I [MSGID: 106568] [glusterd-svc-mgmt.c:253:glusterd_svc_stop]
0-management: quotad service is stopped
[2020-09-14 12:20:42.168942] I [rpc-clnt.c:1005:rpc_clnt_connection_init] 0-bitd: setting
frame-timeout to 600
[2020-09-14 12:20:42.169151] I [MSGID: 106131]
[glusterd-proc-mgmt.c:86:glusterd_proc_stop] 0-management: bitd already stopped
[2020-09-14 12:20:42.169164] I [MSGID: 106568] [glusterd-svc-mgmt.c:253:glusterd_svc_stop]
0-management: bitd service is stopped
[2020-09-14 12:20:42.169193] I [rpc-clnt.c:1005:rpc_clnt_connection_init] 0-scrub: setting
frame-timeout to 600
[2020-09-14 12:20:42.169377] I [MSGID: 106131]
[glusterd-proc-mgmt.c:86:glusterd_proc_stop] 0-management: scrub already stopped
[2020-09-14 12:20:42.169388] I [MSGID: 106568] [glusterd-svc-mgmt.c:253:glusterd_svc_stop]
0-management: scrub service is stopped
[2020-09-14 12:20:42.169420] I [rpc-clnt.c:1005:rpc_clnt_connection_init] 0-snapd: setting
frame-timeout to 600
[2020-09-14 12:20:42.169506] I [rpc-clnt.c:1005:rpc_clnt_connection_init] 0-snapd: setting
frame-timeout to 600
[2020-09-14 12:20:42.169586] I [rpc-clnt.c:1005:rpc_clnt_connection_init] 0-snapd: setting
frame-timeout to 600
[2020-09-14 12:20:42.169670] I [rpc-clnt.c:1005:rpc_clnt_connection_init] 0-gfproxyd:
setting frame-timeout to 600
[2020-09-14 12:20:42.169797] I [rpc-clnt.c:1005:rpc_clnt_connection_init] 0-gfproxyd:
setting frame-timeout to 600
[2020-09-14 12:20:42.169928] I [rpc-clnt.c:1005:rpc_clnt_connection_init] 0-gfproxyd:
setting frame-timeout to 600
[2020-09-14 12:20:42.170136] I [glusterd-utils.c:6225:glusterd_brick_start] 0-management:
discovered already-running brick /gluster_bricks/data/data
[2020-09-14 12:20:42.170151] I [MSGID: 106142] [glusterd-pmap.c:290:pmap_registry_bind]
0-pmap: adding brick /gluster_bricks/data/data on port 49152
[2020-09-14 12:20:42.182826] I [glusterd-utils.c:6225:glusterd_brick_start] 0-management:
discovered already-running brick /gluster_bricks/engine/engine
[2020-09-14 12:20:42.182844] I [MSGID: 106142] [glusterd-pmap.c:290:pmap_registry_bind]
0-pmap: adding brick /gluster_bricks/engine/engine on port 49153
[2020-09-14 12:20:42.207292] I [MSGID: 106131]
[glusterd-proc-mgmt.c:86:glusterd_proc_stop] 0-management: nfs already stopped
[2020-09-14 12:20:42.207314] I [MSGID: 106568] [glusterd-svc-mgmt.c:253:glusterd_svc_stop]
0-management: nfs service is stopped
[2020-09-14 12:20:42.207324] I [MSGID: 106599]
[glusterd-nfs-svc.c:81:glusterd_nfssvc_manager] 0-management: nfs/server.so xlator is not
installed
[2020-09-14 12:20:42.209056] I [MSGID: 106568]
[glusterd-proc-mgmt.c:92:glusterd_proc_stop] 0-management: Stopping glustershd daemon
running in pid: 33647
[2020-09-14 12:20:42.210310] I [MSGID: 106006]
[glusterd-svc-mgmt.c:356:glusterd_svc_common_rpc_notify] 0-management: glustershd has
disconnected from glusterd.
[2020-09-14 12:20:43.209219] I [MSGID: 106568] [glusterd-svc-mgmt.c:253:glusterd_svc_stop]
0-management: glustershd service is stopped
[2020-09-14 12:20:43.209621] I [MSGID: 106567]
[glusterd-svc-mgmt.c:220:glusterd_svc_start] 0-management: Starting glustershd service
[2020-09-14 12:20:44.212440] I [MSGID: 106131]
[glusterd-proc-mgmt.c:86:glusterd_proc_stop] 0-management: quotad already stopped
[2020-09-14 12:20:44.212490] I [MSGID: 106568] [glusterd-svc-mgmt.c:253:glusterd_svc_stop]
0-management: quotad service is stopped
[2020-09-14 12:20:44.212691] I [MSGID: 106131]
[glusterd-proc-mgmt.c:86:glusterd_proc_stop] 0-management: bitd already stopped
[2020-09-14 12:20:44.212712] I [MSGID: 106568] [glusterd-svc-mgmt.c:253:glusterd_svc_stop]
0-management: bitd service is stopped
[2020-09-14 12:20:44.212897] I [MSGID: 106131]
[glusterd-proc-mgmt.c:86:glusterd_proc_stop] 0-management: scrub already stopped
[2020-09-14 12:20:44.212911] I [MSGID: 106568] [glusterd-svc-mgmt.c:253:glusterd_svc_stop]
0-management: scrub service is stopped
[2020-09-14 12:20:44.213829] I [MSGID: 106499]
[glusterd-handler.c:4429:__glusterd_handle_status_volume] 0-management: Received status
volume req for volume engine
[2020-09-14 12:20:46.779903] I [MSGID: 106488]
[glusterd-handler.c:1577:__glusterd_handle_cli_get_volume] 0-management: Received get vol
req
The message "I [MSGID: 106488]
[glusterd-handler.c:1577:__glusterd_handle_cli_get_volume] 0-management: Received get vol
req" repeated 5 times between [2020-09-14 12:20:46.779903] and [2020-09-14
12:20:48.081053]
[2020-09-14 12:20:54.612834] I [MSGID: 106533]
[glusterd-volume-ops.c:982:__glusterd_handle_cli_heal_volume] 0-management: Received heal
vol req for volume engine
[2020-09-14 12:21:01.461754] I [MSGID: 106533]
[glusterd-volume-ops.c:982:__glusterd_handle_cli_heal_volume] 0-management: Received heal
vol req for volume data
[2020-09-14 12:21:37.658372] I [MSGID: 106488]
[glusterd-handler.c:1577:__glusterd_handle_cli_get_volume] 0-management: Received get vol
req
The message "I [MSGID: 106488]
[glusterd-handler.c:1577:__glusterd_handle_cli_get_volume] 0-management: Received get vol
req" repeated 8 times between [2020-09-14 12:21:37.658372] and [2020-09-14
12:22:28.492880]
[2020-09-14 12:22:49.552197] I [MSGID: 106488]
[glusterd-handler.c:1577:__glusterd_handle_cli_get_volume] 0-management: Received get vol
req
[2020-09-14 12:23:39.208008] I [MSGID: 106533]
[glusterd-volume-ops.c:982:__glusterd_handle_cli_heal_volume] 0-management: Received heal
vol req for volume data
[2020-09-14 13:23:42.678379] I [MSGID: 106488]
[glusterd-handler.c:1577:__glusterd_handle_cli_get_volume] 0-management: Received get vol
req
[2020-09-14 13:23:43.038221] I [MSGID: 106488]
[glusterd-handler.c:1577:__glusterd_handle_cli_get_volume] 0-management: Received get vol
req
glustersh.log
-----------------
[2020-09-14 08:40:09.990714] I [socket.c:811:__socket_shutdown] 0-data-client-0:
intentional socket shutdown(5)
[2020-09-14 08:40:15.005498] I [socket.c:811:__socket_shutdown] 0-engine-client-0:
intentional socket shutdown(5)
[2020-09-14 08:40:27.068884] W [socket.c:721:__socket_rwv] 0-engine-client-2: readv on
192.168.0.103:49153 failed (No data available)
[2020-09-14 08:40:27.068922] I [MSGID: 114018] [client.c:2331:client_rpc_notify]
0-engine-client-2: disconnected from engine-client-2. Client process will keep trying to
connect to glusterd until brick's port is available
[2020-09-14 08:40:27.068954] W [MSGID: 108001] [afr-common.c:5613:afr_notify]
0-engine-replicate-0: Client-quorum is not met
[2020-09-14 08:40:37.082276] E [MSGID: 114058]
[client-handshake.c:1449:client_query_portmap_cbk] 0-engine-client-2: failed to get the
port number for remote subvolume. Please run 'gluster volume status' on server to
see if brick process is running.
[2020-09-14 08:40:39.648157] W [socket.c:721:__socket_rwv] 0-data-client-2: readv on
192.168.0.103:49152 failed (No data available)
[2020-09-14 08:40:39.648190] I [MSGID: 114018] [client.c:2331:client_rpc_notify]
0-data-client-2: disconnected from data-client-2. Client process will keep trying to
connect to glusterd until brick's port is available
[2020-09-14 08:40:39.648217] W [MSGID: 108001] [afr-common.c:5613:afr_notify]
0-data-replicate-0: Client-quorum is not met
[2020-09-14 08:40:50.133996] E [MSGID: 114058]
[client-handshake.c:1449:client_query_portmap_cbk] 0-data-client-2: failed to get the port
number for remote subvolume. Please run 'gluster volume status' on server to see
if brick process is running.
[2020-09-14 08:40:50.134055] I [MSGID: 114018] [client.c:2331:client_rpc_notify]
0-data-client-2: disconnected from data-client-2. Client process will keep trying to
connect to glusterd until brick's port is available
[2020-09-14 08:41:05.100136] W [socket.c:721:__socket_rwv] 0-engine-client-1: readv on
192.168.0.102:49153 failed (No data available)
[2020-09-14 08:41:05.100176] I [MSGID: 114018] [client.c:2331:client_rpc_notify]
0-engine-client-1: disconnected from engine-client-1. Client process will keep trying to
connect to glusterd until brick's port is available
[2020-09-14 08:41:05.100193] E [MSGID: 108006]
[afr-common.c:5323:__afr_handle_child_down_event] 0-engine-replicate-0: All subvolumes are
down. Going offline until at least one of them comes back up.
[2020-09-14 08:41:05.147135] W [socket.c:721:__socket_rwv] 0-data-client-1: readv on
192.168.0.102:49152 failed (No data available)
[2020-09-14 08:41:05.147165] I [MSGID: 114018] [client.c:2331:client_rpc_notify]
0-data-client-1: disconnected from data-client-1. Client process will keep trying to
connect to glusterd until brick's port is available
[2020-09-14 08:41:05.147179] E [MSGID: 108006]
[afr-common.c:5323:__afr_handle_child_down_event] 0-data-replicate-0: All subvolumes are
down. Going offline until at least one of them comes back up.
[2020-09-14 08:41:15.279570] E [MSGID: 114058]
[client-handshake.c:1449:client_query_portmap_cbk] 0-engine-client-1: failed to get the
port number for remote subvolume. Please run 'gluster volume status' on server to
see if brick process is running.
[2020-09-14 08:41:15.279610] I [MSGID: 114018] [client.c:2331:client_rpc_notify]
0-engine-client-1: disconnected from engine-client-1. Client process will keep trying to
connect to glusterd until brick's port is available
[2020-09-14 08:41:15.281851] E [MSGID: 114058]
[client-handshake.c:1449:client_query_portmap_cbk] 0-data-client-1: failed to get the port
number for remote subvolume. Please run 'gluster volume status' on server to see
if brick process is running.
[2020-09-14 08:41:15.281891] I [MSGID: 114018] [client.c:2331:client_rpc_notify]
0-data-client-1: disconnected from data-client-1. Client process will keep trying to
connect to glusterd until brick's port is available
[2020-09-14 08:40:37.082339] I [MSGID: 114018] [client.c:2331:client_rpc_notify]
0-engine-client-2: disconnected from engine-client-2. Client process will keep trying to
connect to glusterd until brick's port is available
[2020-09-14 08:42:16.707442] I [socket.c:811:__socket_shutdown] 0-data-client-0:
intentional socket shutdown(5)
[2020-09-14 08:42:21.745058] I [socket.c:811:__socket_shutdown] 0-engine-client-0:
intentional socket shutdown(5)
[2020-09-14 08:42:40.886787] I [socket.c:811:__socket_shutdown] 0-engine-client-2:
intentional socket shutdown(5)
...
...
[2020-09-14 12:20:29.094083] W [socket.c:721:__socket_rwv] 0-glusterfs: readv on
127.0.0.1:24007 failed (No data available)
[2020-09-14 12:20:29.094148] I [glusterfsd-mgmt.c:2443:mgmt_rpc_notify] 0-glusterfsd-mgmt:
disconnected from remote-host: localhost
[2020-09-14 12:20:34.927760] I [socket.c:811:__socket_shutdown] 0-engine-client-2:
intentional socket shutdown(5)
[2020-09-14 12:20:39.959928] I [glusterfsd-mgmt.c:2019:mgmt_getspec_cbk] 0-glusterfs: No
change in volfile,continuing
[2020-09-14 12:20:40.166029] W [glusterfsd.c:1570:cleanup_and_exit]
(-->/lib64/libpthread.so.0(+0x7e65) [0x7f5f21f62e65]
-->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xe5) [0x557934e2e1f5]
-->/usr/sbin/glusterfs(cleanup_and_exit+0x6b) [0x557934e2e05b] ) 0-: received signum
(15), shutting down
[2020-09-14 12:20:41.186725] I [MSGID: 100030] [glusterfsd.c:2847:main]
0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version 6.6 (args:
/usr/sbin/glusterfs -s localhost --volfile-id gluster/glustershd -p
/var/run/gluster/glustershd/glustershd.pid -l /var/log/glusterfs/glustershd.log -S
/var/run/gluster/9713eea6455553ca.socket --xlator-option
*replicate*.node-uuid=6fe3d6e3-ab45-4004-af59-93c2fc3afc93 --process-name glustershd
--client-pid=-6)
[2020-09-14 12:20:41.187235] I [glusterfsd.c:2556:daemonize] 0-glusterfs: Pid of current
running process is 33647
[2020-09-14 12:20:41.191605] I [socket.c:904:__socket_server_bind] 0-socket.glusterfsd:
closing (AF_UNIX) reuse check socket 9
[2020-09-14 12:20:41.196867] I [MSGID: 101190]
[event-epoll.c:680:event_dispatch_epoll_worker] 0-epoll: Started thread with index 1
[2020-09-14 12:20:41.196870] I [MSGID: 101190]
[event-epoll.c:680:event_dispatch_epoll_worker] 0-epoll: Started thread with index 0
[2020-09-14 12:20:42.209282] W [glusterfsd.c:1570:cleanup_and_exit]
(-->/lib64/libpthread.so.0(+0x7e65) [0x7f995868ee65]
-->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xe5) [0x55963354e1f5]
-->/usr/sbin/glusterfs(cleanup_and_exit+0x6b) [0x55963354e05b] ) 0-: received signum
(15), shutting down
[2020-09-14 12:20:43.236232] I [MSGID: 100030] [glusterfsd.c:2847:main]
0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version 6.6 (args:
/usr/sbin/glusterfs -s localhost --volfile-id gluster/glustershd -p
/var/run/gluster/glustershd/glustershd.pid -l /var/log/glusterfs/glustershd.log -S
/var/run/gluster/9713eea6455553ca.socket --xlator-option
*replicate*.node-uuid=6fe3d6e3-ab45-4004-af59-93c2fc3afc93 --process-name glustershd
--client-pid=-6)
[2020-09-14 12:20:43.236667] I [glusterfsd.c:2556:daemonize] 0-glusterfs: Pid of current
running process is 33705
[2020-09-14 12:20:43.240743] I [socket.c:904:__socket_server_bind] 0-socket.glusterfsd:
closing (AF_UNIX) reuse check socket 9
[2020-09-14 12:20:43.246262] I [MSGID: 101190]
[event-epoll.c:680:event_dispatch_epoll_worker] 0-epoll: Started thread with index 0
[2020-09-14 12:20:43.246283] I [MSGID: 101190]
[event-epoll.c:680:event_dispatch_epoll_worker] 0-epoll: Started thread with index 1
[2020-09-14 12:20:44.217875] I [MSGID: 101190]
[event-epoll.c:680:event_dispatch_epoll_worker] 0-epoll: Started thread with index 2
[2020-09-14 12:20:44.217962] I [MSGID: 101190]
[event-epoll.c:680:event_dispatch_epoll_worker] 0-epoll: Started thread with index 3
[2020-09-14 12:20:44.221374] I [MSGID: 114020] [client.c:2401:notify] 0-data-client-0:
parent translators are ready, attempting connect on transport
[2020-09-14 12:20:44.224361] I [MSGID: 114020] [client.c:2401:notify] 0-data-client-1:
parent translators are ready, attempting connect on transport
[2020-09-14 12:20:44.224637] I [rpc-clnt.c:2028:rpc_clnt_reconfig] 0-data-client-0:
changing port to 49152 (from 0)
[2020-09-14 12:20:44.224673] I [socket.c:811:__socket_shutdown] 0-data-client-0:
intentional socket shutdown(12)
[2020-09-14 12:20:44.227134] I [MSGID: 114020] [client.c:2401:notify] 0-data-client-2:
parent translators are ready, attempting connect on transport
[2020-09-14 12:20:44.227728] I [rpc-clnt.c:2028:rpc_clnt_reconfig] 0-data-client-1:
changing port to 49152 (from 0)
[2020-09-14 12:20:44.227758] I [socket.c:811:__socket_shutdown] 0-data-client-1:
intentional socket shutdown(13)
[2020-09-14 12:20:44.230217] I [MSGID: 114046]
[client-handshake.c:1106:client_setvolume_cbk] 0-data-client-0: Connected to
data-client-0, attached to remote volume '/gluster_bricks/data/data'.
[2020-09-14 12:20:44.230242] I [MSGID: 108005]
[afr-common.c:5245:__afr_handle_child_up_event] 0-data-replicate-0: Subvolume
'data-client-0' came back up; going online.
[2020-09-14 12:20:44.231537] I [MSGID: 114020] [client.c:2401:notify] 0-engine-client-0:
parent translators are ready, attempting connect on transport
[2020-09-14 12:20:44.232072] I [rpc-clnt.c:2028:rpc_clnt_reconfig] 0-data-client-2:
changing port to 49152 (from 0)
[2020-09-14 12:20:44.232112] I [socket.c:811:__socket_shutdown] 0-data-client-2:
intentional socket shutdown(15)
[2020-09-14 12:20:44.234981] I [MSGID: 114046]
[client-handshake.c:1106:client_setvolume_cbk] 0-data-client-1: Connected to
data-client-1, attached to remote volume '/gluster_bricks/data/data'.
[2020-09-14 12:20:44.235005] I [MSGID: 108002] [afr-common.c:5607:afr_notify]
0-data-replicate-0: Client-quorum is met
[2020-09-14 12:20:44.235988] I [MSGID: 114020] [client.c:2401:notify] 0-engine-client-1:
parent translators are ready, attempting connect on transport
[2020-09-14 12:20:44.236176] I [rpc-clnt.c:2028:rpc_clnt_reconfig] 0-engine-client-0:
changing port to 49153 (from 0)
[2020-09-14 12:20:44.236217] I [socket.c:811:__socket_shutdown] 0-engine-client-0:
intentional socket shutdown(12)
[2020-09-14 12:20:44.239425] I [MSGID: 114046]
[client-handshake.c:1106:client_setvolume_cbk] 0-data-client-2: Connected to
data-client-2, attached to remote volume '/gluster_bricks/data/data'.
[2020-09-14 12:20:44.240456] I [MSGID: 114020] [client.c:2401:notify] 0-engine-client-2:
parent translators are ready, attempting connect on transport
[2020-09-14 12:20:44.240960] I [rpc-clnt.c:2028:rpc_clnt_reconfig] 0-engine-client-1:
changing port to 49153 (from 0)
[2020-09-14 12:20:44.240987] I [socket.c:811:__socket_shutdown] 0-engine-client-1:
intentional socket shutdown(17)
[2020-09-14 12:20:44.243360] I [MSGID: 114046]
[client-handshake.c:1106:client_setvolume_cbk] 0-engine-client-0: Connected to
engine-client-0, attached to remote volume '/gluster_bricks/engine/engine'.
[2020-09-14 12:20:44.243385] I [MSGID: 108005]
[afr-common.c:5245:__afr_handle_child_up_event] 0-engine-replicate-0: Subvolume
'engine-client-0' came back up; going online.
[2020-09-14 12:20:44.245314] I [rpc-clnt.c:2028:rpc_clnt_reconfig] 0-engine-client-2:
changing port to 49153 (from 0)
[2020-09-14 12:20:44.245344] I [socket.c:811:__socket_shutdown] 0-engine-client-2:
intentional socket shutdown(15)
[2020-09-14 12:20:44.248543] I [MSGID: 114046]
[client-handshake.c:1106:client_setvolume_cbk] 0-engine-client-1: Connected to
engine-client-1, attached to remote volume '/gluster_bricks/engine/engine'.
[2020-09-14 12:20:44.248564] I [MSGID: 108002] [afr-common.c:5607:afr_notify]
0-engine-replicate-0: Client-quorum is met
[2020-09-14 12:20:44.253661] I [MSGID: 114046]
[client-handshake.c:1106:client_setvolume_cbk] 0-engine-client-2: Connected to
engine-client-2, attached to remote volume '/gluster_bricks/engine/engine'.
[2020-09-14 12:20:44.486491] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on b72e302c-478d-4f90-bdd7-5d542928bcc1. sources=[1] 2 sinks=0
[2020-09-14 12:20:47.661708] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on b0ba8fbc-750e-4a2d-a218-d49500d95d26. sources=[1] 2 sinks=0
[2020-09-14 12:20:53.451995] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on bbfa9185-5383-44f8-a32d-9b1aaa6e43a7. sources=[1] 2 sinks=0
[2020-09-14 12:20:54.148249] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 1f2ee964-e4bc-432e-be43-5e95783cace1. sources=[1] 2 sinks=0
[2020-09-14 12:20:54.148386] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 937a7d82-5b99-487b-9614-3aad7ef2dc97. sources=[1] 2 sinks=0
[2020-09-14 12:20:54.148420] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on b9a47d47-111b-403b-9d05-b741c2878ff6. sources=[1] 2 sinks=0
[2020-09-14 12:20:54.443785] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on abe59023-a60b-495e-a1a3-2698dbf747b5. sources=[1] 2 sinks=0
[2020-09-14 12:20:54.647683] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 7d1de36f-d6ff-4a34-91f3-30d69e28cadb. sources=[1] 2 sinks=0
[2020-09-14 12:20:54.796212] I [MSGID: 108026]
[afr-self-heal-entry.c:898:afr_selfheal_entry_do] 0-engine-replicate-0: performing entry
selfheal on 65fc9ad6-69bb-4d7a-87de-8f286450ea5b
[2020-09-14 12:20:54.805989] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-engine-replicate-0: Completed entry
selfheal on 65fc9ad6-69bb-4d7a-87de-8f286450ea5b. sources=[0] 1 sinks=2
[2020-09-14 12:20:55.295470] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 7bae68bc-7629-4fa3-8fe4-b13768a98fcc. sources=[1] 2 sinks=0
[2020-09-14 12:21:03.753629] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 7432e014-3990-44fe-9409-9eba4090f547. sources=[1] 2 sinks=0
[2020-09-14 12:21:03.773390] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on cf3abe71-ba89-490a-bb33-5e900cb78967. sources=[1] 2 sinks=0
[2020-09-14 12:21:04.421092] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on f623d6ec-1541-4db2-bc21-35a450d142de. sources=[1] 2 sinks=0
[2020-09-14 12:21:04.503141] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 06874587-6885-4ab3-9cc4-fabb2db99d25. sources=[1] 2 sinks=0
[2020-09-14 12:21:04.505516] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 46c3a7b2-7d5a-432e-857a-dbf332908c89. sources=[1] 2 sinks=0
[2020-09-14 12:21:04.506978] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on f93818dc-0279-4280-856c-151c4eb00f8f. sources=[1] 2 sinks=0
[2020-09-14 12:21:04.507327] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 2fcf8b6a-79a2-4328-bde8-e5e8021c0256. sources=[1] 2 sinks=0
[2020-09-14 12:21:04.507510] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 7f20e494-afc6-4d60-822e-33aa30fe52f7. sources=[1] 2 sinks=0
[2020-09-14 12:21:11.025834] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 4719908b-2613-4976-bdc9-ed10847ea24e. sources=[1] 2 sinks=0
[2020-09-14 12:21:11.692657] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on db070d51-7c16-44f5-8b58-cb923edfc72a. sources=[1] 2 sinks=0
[2020-09-14 12:21:19.500323] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 1f87aefd-684f-4dd3-ad77-a3e47006ef14. sources=[1] 2 sinks=0
[2020-09-14 12:21:19.556305] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 22639cf7-62f5-4198-912f-2b446062ee9b. sources=[1] 2 sinks=0
[2020-09-14 12:21:20.147475] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 0a109124-0e71-4912-9475-47be55805c7f. sources=[1] 2 sinks=0
[2020-09-14 12:21:20.369201] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 028b1032-61ab-49e8-a427-b982d22ede87. sources=[1] 2 sinks=0
[2020-09-14 12:21:20.498106] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 7d99c05f-45b7-4c5e-a7a0-29a6eb6d5353. sources=[1] 2 sinks=0
[2020-09-14 12:21:20.627626] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 731016ec-16be-456f-9f94-d545e83fb730. sources=[1] 2 sinks=0
[2020-09-14 12:21:20.932841] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 54a4dd15-f2ed-4cee-b52c-d76c3a14ee1d. sources=[1] 2 sinks=0
[2020-09-14 12:21:22.950551] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on d9f8db65-bbf6-43d2-a58d-02062e68b094. sources=[1] 2 sinks=0
[2020-09-14 12:21:27.591944] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 2a369725-e692-4985-b1e6-c2ca919977f7. sources=[1] 2 sinks=0
[2020-09-14 12:21:27.811709] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 220f5b3f-7f1f-4afa-8e02-7039dfc46c71. sources=[1] 2 sinks=0
[2020-09-14 12:21:27.847105] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on cf06956a-4726-4a62-9800-310b860ad8c0. sources=[1] 2 sinks=0
[2020-09-14 12:21:27.850830] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on c25c8901-a0f6-40fe-92a2-c616e9bad32e. sources=[1] 2 sinks=0
[2020-09-14 12:21:28.051319] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on ad151cc0-f703-4ba8-8d88-9f21b9b808af. sources=[1] 2 sinks=0
[2020-09-14 12:21:28.254506] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 24a52f25-a440-431a-9ed2-9b7bfa640507. sources=[1] 2 sinks=0
[2020-09-14 12:21:28.436180] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 7f56b5c9-632c-4112-9c9e-bf62b8af09ff. sources=[1] 2 sinks=0
[2020-09-14 12:21:30.127830] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 0f96020d-b860-45ab-9e1d-4ca1839e19f6. sources=[1] 2 sinks=0
[2020-09-14 12:21:31.986013] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on b3b7dc3b-8429-4cb6-8f80-4fe7c23715ac. sources=[1] 2 sinks=0
[2020-09-14 12:21:32.928726] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on d0c4461a-0fa2-439c-ab81-8b97c16a9a8b. sources=[1] 2 sinks=0
[2020-09-14 12:21:33.096508] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 5dff0bbe-cfc6-458f-871a-47cfebfa018a. sources=[1] 2 sinks=0
[2020-09-14 12:21:33.394394] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on f9705bf7-4213-45ca-a17a-f1ebbf11f9ce. sources=[1] 2 sinks=0
[2020-09-14 12:21:38.089910] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on ebe1cce3-3abd-42d3-9504-9e7893721e4b. sources=[1] 2 sinks=0
[2020-09-14 12:21:43.452533] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on ee204b66-5890-469b-a796-7d877b4bf1ab. sources=[1] 2 sinks=0
[2020-09-14 12:21:43.454253] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on ea047187-b01d-49e2-8d02-c38dbf9e77b5. sources=[1] 2 sinks=0
[2020-09-14 12:21:43.710631] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on ce6556b5-a644-4634-91fa-bc70765dbc5a. sources=[1] 2 sinks=0
[2020-09-14 12:21:44.641002] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 96691acb-b98f-4913-a261-c06e15ca18d1. sources=[1] 2 sinks=0
[2020-09-14 12:21:45.332554] I [MSGID: 108026]
[afr-self-heal-common.c:1742:afr_log_selfheal] 0-data-replicate-0: Completed data selfheal
on 8a84ab8b-3890-4bf6-bd49-ebd8d9b96ce2. sources=[1] 2 sinks=0