I am trying to remove a brick, from a server which is no longer part of the gluster pool, but I keep running into errors for which I cannot find answers on google.
[root@virt2 ~]# gluster peer status
Number of Peers: 3
Hostname: srv1
Uuid: 2bed7e51-430f-49f5-afbc-06f8cec9baeb
State: Peer in Cluster (Disconnected)
Hostname: srv3
Uuid: 0e78793c-deca-4e3b-a36f-2333c8f91825
State: Peer in Cluster (Connected)
Hostname: srv4
Uuid: 1a6eedc6-59eb-4329-b091-2b9bc6f0834f
State: Peer in Cluster (Connected)
[root@virt2 ~]#
[root@virt2 ~]# gluster volume info data
Volume Name: data
Type: Replicate
Volume ID: d09e4534-8bc0-4b30-be89-bc1ec2b439c7
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: srv1:/gluster/data/brick1
Brick2: srv2:/gluster/data/brick1
Brick3: srv3:/gluster/data/brick1
Options Reconfigured:
nfs.disable: on
transport.address-family: inet
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.low-prio-threads: 32
network.remote-dio: enable
cluster.eager-lock: enable
cluster.quorum-type: auto
cluster.server-quorum-type: server
cluster.data-self-heal-algorithm: full
cluster.locking-scheme: granular
cluster.shd-max-threads: 8
cluster.shd-wait-qlength: 10000
features.shard: on
user.cifs: off
storage.owner-uid: 36
storage.owner-gid: 36
features.shard-block-size: 512MB
[root@virt2 ~]# gluster volume remove-brick data replica 2 srv1:/gluster/data/brick1 start
volume remove-brick start: failed: Migration of data is not needed when reducing replica count. Use the 'force' option
[root@virt2 ~]# gluster volume remove-brick data replica 2 srv1:/gluster/data/brick1 commit
Removing brick(s) can result in data loss. Do you want to Continue? (y/n) y
volume remove-brick commit: failed: Brick srv1:/gluster/data/brick1 is not decommissioned. Use start or force option
The server virt1 is not part of the cluster anymore.