On Fri, Feb 19, 2016 at 11:54 PM, Cameron Christensen
<cameron.christensen(a)uk2group.com> wrote:
I cannot put the gluster domain into maintenance. I'm believe
this is
because the data center has a status of Non responsive (because a host
cannot connect to storage or start SPM). The only option available on the
gluster storage is activate. I have put all the hosts into maintenance. Is
this enough to continue with the initialize lockspace step?
Yes, if all hosts are in maintenance, no host will access the gluster
storage domain, and you can repair is safely.
On Fri, 2016-02-19 at 23:34 +0200, Nir Soffer wrote:
On Fri, Feb 19, 2016 at 10:58 PM, Cameron Christensen
<cameron.christensen(a)uk2group.com> wrote:
Hello,
I am using glusterfs storage and ran into a split-brain issue. One of the
file affected by split-brain was dom_md/ids. In attempts to fix the
split-brain issue I deleted the dom_md/ids file. Is there a method to
recreate or reconstruct this file?
You can do this:
1. Put the gluster domain to maintenance (via engine)
No host should access it while you reconstruct the ids file
2. Mount the gluster volume manually
mkdir repair
mount -t glusterfs <server>:/<path> repair/
3. Create the file:
touch repair/<sd_uuid>/dom_md/ids
4. Initialize the lockspace
sanlock direct init -s <sd_uuid>:0:repair/<sd_uuid>/dom_md/ids:0
5. Unmount the gluster volume
umount repair
6. Activate the gluster domain (via engine)
The domain should become active after a while.
David: can you confirm this is the best way to reconstruct the ids file?
Nir