Hi Satheesaran,

gluster volume info engine
 
Volume Name: engine
Type: Replicate
Volume ID: 3caae601-74dd-40d1-8629-9a61072bec0f
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x (2 + 1) = 3
Transport-type: tcp
Bricks:
Brick1: gluster0:/gluster/engine/brick
Brick2: gluster1:/gluster/engine/brick
Brick3: gluster2:/gluster/engine/brick (arbiter)
Options Reconfigured:
nfs.disable: on
performance.readdir-ahead: on
transport.address-family: inet
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.stat-prefetch: off
performance.low-prio-threads: 32
network.remote-dio: off
cluster.eager-lock: enable
cluster.quorum-type: auto
cluster.server-quorum-type: server
cluster.data-self-heal-algorithm: full
cluster.locking-scheme: granular
cluster.shd-max-threads: 8
cluster.shd-wait-qlength: 10000
features.shard: on
user.cifs: off
storage.owner-uid: 36
storage.owner-gid: 36
network.ping-timeout: 30
performance.strict-o-direct: on
cluster.granular-entry-heal: enable
nfs.export-volumes: on

As per my previous, i have resolved this by following the steps described.


On Tue, Jun 27, 2017 at 1:42 PM, Satheesaran Sundaramoorthi <sasundar@redhat.com> wrote:
On Sat, Jun 24, 2017 at 3:17 PM, Abi Askushi <rightkicktech@gmail.com> wrote:
Hi all,

For the records, I had to remove manually the conflicting directory and ts respective gfid from the arbiter volume:

 getfattr -m . -d -e hex e1c80750-b880-495e-9609-b8bc7760d101/ha_agent

That gave me the gfid: 0x277c9caa9dce4a17a2a93775357befd5

Then cd .glusterfs/27/7c

rm -rf 277c9caa-9dce-4a17-a2a9-3775357befd5 (or move it out of there)

Triggerred heal: gluster volume heal engine

Then all ok:

gluster volume heal engine info
Brick gluster0:/gluster/engine/brick
Status: Connected
Number of entries: 0

Brick gluster1:/gluster/engine/brick
Status: Connected
Number of entries: 0

Brick gluster2:/gluster/engine/brick
Status: Connected
Number of entries: 0

Thanx. 

​Hi Abi,

What is the volume type of 'engine' volume ?
Could you also provide the output of 'gluster volume info engine' to get to the
closer look at the problem

​-- sas​