
Hello, I ,ve setup a test lab with 3 nodes installed with centos 7 I configured manualy gluster fs. Glusterfs is up and running [root@kvm380 ~]# gluster peer status Number of Peers: 2 Hostname: kvm320.durchhalten.intern Uuid: dac066db-55f7-4770-900d-4830c740ffbf State: Peer in Cluster (Connected) Hostname: kvm360.durchhalten.intern Uuid: 4291be40-f77f-4f41-98f6-dc48fd993842 State: Peer in Cluster (Connected) [root@kvm380 ~]# gluster volume info Volume Name: data Type: Replicate Volume ID: 3586de82-e504-4c62-972b-448abead13d3 Status: Started Snapshot Count: 0 Number of Bricks: 1 x 3 = 3 Transport-type: tcp Bricks: Brick1: kvm380.durchhalten.intern:/gluster/data Brick2: kvm360.durchhalten.intern:/gluster/data Brick3: kvm320.durchhalten.intern:/gluster/data Options Reconfigured: storage.owner-uid: 36 storage.owner-gid: 36 features.shard: on performance.low-prio-threads: 32 performance.strict-o-direct: on network.ping-timeout: 30 user.cifs: off network.remote-dio: off performance.quick-read: off performance.read-ahead: off performance.io-cache: off cluster.eager-lock: enable transport.address-family: inet nfs.disable: on performance.client-io-threads: off Volume Name: engine Type: Replicate Volume ID: dcfbd322-5dd0-4bfe-a775-99ecc79e1416 Status: Started Snapshot Count: 0 Number of Bricks: 1 x 3 = 3 Transport-type: tcp Bricks: Brick1: kvm380.durchhalten.intern:/gluster/engine Brick2: kvm360.durchhalten.intern:/gluster/engine Brick3: kvm320.durchhalten.intern:/gluster/engine Options Reconfigured: storage.owner-uid: 36 storage.owner-gid: 36 features.shard: on performance.low-prio-threads: 32 performance.strict-o-direct: on network.remote-dio: off network.ping-timeout: 30 user.cifs: off performance.quick-read: off performance.read-ahead: off performance.io-cache: off cluster.eager-lock: enable transport.address-family: inet nfs.disable: on performance.client-io-threads: off After that I deployed a selfhosted engine And add the two other hosts, at the beginning it looks good, but without changing anything I got following error by two hosts ! 20.12.2018 11:35:05 Failed to connect Host kvm320.durchhalten.intern to Storage Pool Default ! 20.12.2018 11:35:05 Host kvm320.durchhalten.intern cannot access the Storage Domain(s) hosted_storage attached to the Data Center Default. Setting Host state to Non-Operational. X 20.12.2018 11:35:05 Host kvm320.durchhalten.intern reports about one of the Active Storage Domains as Problematic. ! 20.12.2018 11:35:05 Kdump integration is enabled for host kvm320.durchhalten.intern, but kdump is not configured properly on host. ! 20.12.2018 11:35:04 Failed to connect Host kvm360.durchhalten.intern to Storage Pool Default ! 20.12.2018 11:35:04 Host kvm360.durchhalten.intern cannot access the Storage Domain(s) hosted_storage attached to the Data Center Default. Setting Host state to Non-Operational. X 20.12.2018 11:35:04 Host kvm360.durchhalten.intern reports about one of the Active Storage Domains as Problematic. Before glusterfs I had a setup with nfs on 4. Server Where is the problem? thx