Re: Single instance scaleup.

Hi Leo, As you do not have a distributed volume , you can easily switch to replica 2 arbiter 1 or replica 3 volumes. You can use the following for adding the bricks: https://access.redhat.com/documentation/en-US/Red_Hat_Storage/2.1/html/Admin... Best Regards, Strahil NikolivOn May 26, 2019 10:54, Leo David <leoalex@gmail.com> wrote:
Hi Stahil, Thank you so much for yout input !
gluster volume info
Volume Name: engine Type: Distribute Volume ID: d7449fc2-cc35-4f80-a776-68e4a3dbd7e1 Status: Started Snapshot Count: 0 Number of Bricks: 1 Transport-type: tcp Bricks: Brick1: 192.168.80.191:/gluster_bricks/engine/engine Options Reconfigured: nfs.disable: on transport.address-family: inet storage.owner-uid: 36 storage.owner-gid: 36 features.shard: on performance.low-prio-threads: 32 performance.strict-o-direct: off network.remote-dio: off network.ping-timeout: 30 user.cifs: off performance.quick-read: off performance.read-ahead: off performance.io-cache: off cluster.eager-lock: enable Volume Name: ssd-samsung Type: Distribute Volume ID: 76576cc6-220b-4651-952d-99846178a19e Status: Started Snapshot Count: 0 Number of Bricks: 1 Transport-type: tcp Bricks: Brick1: 192.168.80.191:/gluster_bricks/sdc/data Options Reconfigured: cluster.eager-lock: enable performance.io-cache: off performance.read-ahead: off performance.quick-read: off user.cifs: off network.ping-timeout: 30 network.remote-dio: off performance.strict-o-direct: on performance.low-prio-threads: 32 features.shard: on storage.owner-gid: 36 storage.owner-uid: 36 transport.address-family: inet nfs.disable: on
The other two hosts will be 192.168.80.192/193 - this is gluster dedicated network over 10GB sfp+ switch. - host 2 wil have identical harware configuration with host 1 ( each disk is actually a raid0 array ) - host 3 has: - 1 ssd for OS - 1 ssd - for adding to engine volume in a full replica 3 - 2 ssd's in a raid 1 array to be added as arbiter for the data volume ( ssd-samsung ) So the plan is to have "engine" scaled in a full replica 3, and "ssd-samsung" scalled in a replica 3 arbitrated.
On Sun, May 26, 2019 at 10:34 AM Strahil <hunter86_bg@yahoo.com> wrote:
Hi Leo,
Gluster is quite smart, but in order to provide any hints , can you provide output of 'gluster volume info <glustervol>'. If you have 2 more systems , keep in mind that it is best to mirror the storage on the second replica (2 disks on 1 machine -> 2 disks on the new machine), while for the arbiter this is not neccessary.
What is your network and NICs ? Based on my experience , I can recommend at least 10 gbit/s interfase(s).
Best Regards, Strahil Nikolov
On May 26, 2019 07:52, Leo David <leoalex@gmail.com> wrote:
Hello Everyone, Can someone help me to clarify this ? I have a single-node 4.2.8 installation ( only two gluster storage domains - distributed single drive volumes ). Now I just got two identintical servers and I would like to go for a 3 nodes bundle. Is it possible ( after joining the new nodes to the cluster ) to expand the existing volumes across the new nodes and change them to replica 3 arbitrated ? If so, could you share with me what would it be the procedure ? Thank you very much !
Leo
-- Best regards, Leo David

Thank you Strahil, The engine and ssd-samsung are distributed... So these are the ones that I need to have replicated accross new nodes. I am not very sure about the procedure to accomplish this. Thanks, Leo On Sun, May 26, 2019, 13:04 Strahil <hunter86_bg@yahoo.com> wrote:
Hi Leo, As you do not have a distributed volume , you can easily switch to replica 2 arbiter 1 or replica 3 volumes.
You can use the following for adding the bricks:
https://access.redhat.com/documentation/en-US/Red_Hat_Storage/2.1/html/Admin...
Best Regards, Strahil Nikoliv On May 26, 2019 10:54, Leo David <leoalex@gmail.com> wrote:
Hi Stahil, Thank you so much for yout input !
gluster volume info
Volume Name: engine Type: Distribute Volume ID: d7449fc2-cc35-4f80-a776-68e4a3dbd7e1 Status: Started Snapshot Count: 0 Number of Bricks: 1 Transport-type: tcp Bricks: Brick1: 192.168.80.191:/gluster_bricks/engine/engine Options Reconfigured: nfs.disable: on transport.address-family: inet storage.owner-uid: 36 storage.owner-gid: 36 features.shard: on performance.low-prio-threads: 32 performance.strict-o-direct: off network.remote-dio: off network.ping-timeout: 30 user.cifs: off performance.quick-read: off performance.read-ahead: off performance.io-cache: off cluster.eager-lock: enable Volume Name: ssd-samsung Type: Distribute Volume ID: 76576cc6-220b-4651-952d-99846178a19e Status: Started Snapshot Count: 0 Number of Bricks: 1 Transport-type: tcp Bricks: Brick1: 192.168.80.191:/gluster_bricks/sdc/data Options Reconfigured: cluster.eager-lock: enable performance.io-cache: off performance.read-ahead: off performance.quick-read: off user.cifs: off network.ping-timeout: 30 network.remote-dio: off performance.strict-o-direct: on performance.low-prio-threads: 32 features.shard: on storage.owner-gid: 36 storage.owner-uid: 36 transport.address-family: inet nfs.disable: on
The other two hosts will be 192.168.80.192/193 - this is gluster dedicated network over 10GB sfp+ switch. - host 2 wil have identical harware configuration with host 1 ( each disk is actually a raid0 array ) - host 3 has: - 1 ssd for OS - 1 ssd - for adding to engine volume in a full replica 3 - 2 ssd's in a raid 1 array to be added as arbiter for the data volume ( ssd-samsung ) So the plan is to have "engine" scaled in a full replica 3, and "ssd-samsung" scalled in a replica 3 arbitrated.
On Sun, May 26, 2019 at 10:34 AM Strahil <hunter86_bg@yahoo.com> wrote:
Hi Leo,
Gluster is quite smart, but in order to provide any hints , can you provide output of 'gluster volume info <glustervol>'. If you have 2 more systems , keep in mind that it is best to mirror the storage on the second replica (2 disks on 1 machine -> 2 disks on the new machine), while for the arbiter this is not neccessary.
What is your network and NICs ? Based on my experience , I can recommend at least 10 gbit/s interfase(s).
Best Regards, Strahil Nikolov On May 26, 2019 07:52, Leo David <leoalex@gmail.com> wrote:
Hello Everyone, Can someone help me to clarify this ? I have a single-node 4.2.8 installation ( only two gluster storage domains - distributed single drive volumes ). Now I just got two identintical servers and I would like to go for a 3 nodes bundle. Is it possible ( after joining the new nodes to the cluster ) to expand the existing volumes across the new nodes and change them to replica 3 arbitrated ? If so, could you share with me what would it be the procedure ? Thank you very much !
Leo
-- Best regards, Leo David

Yeah,it seems different from the docs.I'm adding the gluster users list ,as they are more experienced into that. @Gluster-users, can you provide some hint how to add aditional replicas to the below volumes , so they become 'replica 2 arbiter 1' or 'replica 3' type volumes ? Best Regards,Strahil Nikolov В неделя, 26 май 2019 г., 15:16:18 ч. Гринуич+3, Leo David <leoalex@gmail.com> написа: Thank you Strahil,The engine and ssd-samsung are distributed...So these are the ones that I need to have replicated accross new nodes.I am not very sure about the procedure to accomplish this.Thanks, Leo On Sun, May 26, 2019, 13:04 Strahil <hunter86_bg@yahoo.com> wrote: Hi Leo, As you do not have a distributed volume , you can easily switch to replica 2 arbiter 1 or replica 3 volumes. You can use the following for adding the bricks: https://access.redhat.com/documentation/en-US/Red_Hat_Storage/2.1/html/Admin... Best Regards, Strahil Nikoliv On May 26, 2019 10:54, Leo David <leoalex@gmail.com> wrote: Hi Stahil,Thank you so much for yout input ! gluster volume info Volume Name: engine Type: Distribute Volume ID: d7449fc2-cc35-4f80-a776-68e4a3dbd7e1 Status: Started Snapshot Count: 0 Number of Bricks: 1 Transport-type: tcp Bricks: Brick1: 192.168.80.191:/gluster_bricks/engine/engine Options Reconfigured: nfs.disable: on transport.address-family: inet storage.owner-uid: 36 storage.owner-gid: 36 features.shard: on performance.low-prio-threads: 32 performance.strict-o-direct: off network.remote-dio: off network.ping-timeout: 30 user.cifs: off performance.quick-read: off performance.read-ahead: off performance.io-cache: off cluster.eager-lock: enableVolume Name: ssd-samsung Type: Distribute Volume ID: 76576cc6-220b-4651-952d-99846178a19e Status: Started Snapshot Count: 0 Number of Bricks: 1 Transport-type: tcp Bricks: Brick1: 192.168.80.191:/gluster_bricks/sdc/data Options Reconfigured: cluster.eager-lock: enable performance.io-cache: off performance.read-ahead: off performance.quick-read: off user.cifs: off network.ping-timeout: 30 network.remote-dio: off performance.strict-o-direct: on performance.low-prio-threads: 32 features.shard: on storage.owner-gid: 36 storage.owner-uid: 36 transport.address-family: inet nfs.disable: on The other two hosts will be 192.168.80.192/193 - this is gluster dedicated network over 10GB sfp+ switch.- host 2 wil have identical harware configuration with host 1 ( each disk is actually a raid0 array )- host 3 has: - 1 ssd for OS - 1 ssd - for adding to engine volume in a full replica 3 - 2 ssd's in a raid 1 array to be added as arbiter for the data volume ( ssd-samsung )So the plan is to have "engine" scaled in a full replica 3, and "ssd-samsung" scalled in a replica 3 arbitrated. On Sun, May 26, 2019 at 10:34 AM Strahil <hunter86_bg@yahoo.com> wrote: Hi Leo, Gluster is quite smart, but in order to provide any hints , can you provide output of 'gluster volume info <glustervol>'. If you have 2 more systems , keep in mind that it is best to mirror the storage on the second replica (2 disks on 1 machine -> 2 disks on the new machine), while for the arbiter this is not neccessary. What is your network and NICs ? Based on my experience , I can recommend at least 10 gbit/s interfase(s). Best Regards, Strahil Nikolov On May 26, 2019 07:52, Leo David <leoalex@gmail.com> wrote: Hello Everyone,Can someone help me to clarify this ?I have a single-node 4.2.8 installation ( only two gluster storage domains - distributed single drive volumes ). Now I just got two identintical servers and I would like to go for a 3 nodes bundle.Is it possible ( after joining the new nodes to the cluster ) to expand the existing volumes across the new nodes and change them to replica 3 arbitrated ?If so, could you share with me what would it be the procedure ?Thank you very much ! Leo -- Best regards, Leo David

Hi, Any suggestions ? Thank you very much ! Leo On Sun, May 26, 2019 at 4:38 PM Strahil Nikolov <hunter86_bg@yahoo.com> wrote:
Yeah, it seems different from the docs. I'm adding the gluster users list ,as they are more experienced into that.
@Gluster-users,
can you provide some hint how to add aditional replicas to the below volumes , so they become 'replica 2 arbiter 1' or 'replica 3' type volumes ?
Best Regards, Strahil Nikolov
В неделя, 26 май 2019 г., 15:16:18 ч. Гринуич+3, Leo David < leoalex@gmail.com> написа:
Thank you Strahil, The engine and ssd-samsung are distributed... So these are the ones that I need to have replicated accross new nodes. I am not very sure about the procedure to accomplish this. Thanks,
Leo
On Sun, May 26, 2019, 13:04 Strahil <hunter86_bg@yahoo.com> wrote:
Hi Leo, As you do not have a distributed volume , you can easily switch to replica 2 arbiter 1 or replica 3 volumes.
You can use the following for adding the bricks:
https://access.redhat.com/documentation/en-US/Red_Hat_Storage/2.1/html/Admin...
Best Regards, Strahil Nikoliv On May 26, 2019 10:54, Leo David <leoalex@gmail.com> wrote:
Hi Stahil, Thank you so much for yout input !
gluster volume info
Volume Name: engine Type: Distribute Volume ID: d7449fc2-cc35-4f80-a776-68e4a3dbd7e1 Status: Started Snapshot Count: 0 Number of Bricks: 1 Transport-type: tcp Bricks: Brick1: 192.168.80.191:/gluster_bricks/engine/engine Options Reconfigured: nfs.disable: on transport.address-family: inet storage.owner-uid: 36 storage.owner-gid: 36 features.shard: on performance.low-prio-threads: 32 performance.strict-o-direct: off network.remote-dio: off network.ping-timeout: 30 user.cifs: off performance.quick-read: off performance.read-ahead: off performance.io-cache: off cluster.eager-lock: enable Volume Name: ssd-samsung Type: Distribute Volume ID: 76576cc6-220b-4651-952d-99846178a19e Status: Started Snapshot Count: 0 Number of Bricks: 1 Transport-type: tcp Bricks: Brick1: 192.168.80.191:/gluster_bricks/sdc/data Options Reconfigured: cluster.eager-lock: enable performance.io-cache: off performance.read-ahead: off performance.quick-read: off user.cifs: off network.ping-timeout: 30 network.remote-dio: off performance.strict-o-direct: on performance.low-prio-threads: 32 features.shard: on storage.owner-gid: 36 storage.owner-uid: 36 transport.address-family: inet nfs.disable: on
The other two hosts will be 192.168.80.192/193 - this is gluster dedicated network over 10GB sfp+ switch. - host 2 wil have identical harware configuration with host 1 ( each disk is actually a raid0 array ) - host 3 has: - 1 ssd for OS - 1 ssd - for adding to engine volume in a full replica 3 - 2 ssd's in a raid 1 array to be added as arbiter for the data volume ( ssd-samsung ) So the plan is to have "engine" scaled in a full replica 3, and "ssd-samsung" scalled in a replica 3 arbitrated.
On Sun, May 26, 2019 at 10:34 AM Strahil <hunter86_bg@yahoo.com> wrote:
Hi Leo,
Gluster is quite smart, but in order to provide any hints , can you provide output of 'gluster volume info <glustervol>'. If you have 2 more systems , keep in mind that it is best to mirror the storage on the second replica (2 disks on 1 machine -> 2 disks on the new machine), while for the arbiter this is not neccessary.
What is your network and NICs ? Based on my experience , I can recommend at least 10 gbit/s interfase(s).
Best Regards, Strahil Nikolov On May 26, 2019 07:52, Leo David <leoalex@gmail.com> wrote:
Hello Everyone, Can someone help me to clarify this ? I have a single-node 4.2.8 installation ( only two gluster storage domains - distributed single drive volumes ). Now I just got two identintical servers and I would like to go for a 3 nodes bundle. Is it possible ( after joining the new nodes to the cluster ) to expand the existing volumes across the new nodes and change them to replica 3 arbitrated ? If so, could you share with me what would it be the procedure ? Thank you very much !
Leo
-- Best regards, Leo David
-- Best regards, Leo David

Hi, Looks like the only way arround would be to create a brand-new volume as replicated on other disks, and start moving the vms all around the place between volumes ? Cheers, Leo On Mon, May 27, 2019 at 1:53 PM Leo David <leoalex@gmail.com> wrote:
Hi, Any suggestions ? Thank you very much !
Leo
On Sun, May 26, 2019 at 4:38 PM Strahil Nikolov <hunter86_bg@yahoo.com> wrote:
Yeah, it seems different from the docs. I'm adding the gluster users list ,as they are more experienced into that.
@Gluster-users,
can you provide some hint how to add aditional replicas to the below volumes , so they become 'replica 2 arbiter 1' or 'replica 3' type volumes ?
Best Regards, Strahil Nikolov
В неделя, 26 май 2019 г., 15:16:18 ч. Гринуич+3, Leo David < leoalex@gmail.com> написа:
Thank you Strahil, The engine and ssd-samsung are distributed... So these are the ones that I need to have replicated accross new nodes. I am not very sure about the procedure to accomplish this. Thanks,
Leo
On Sun, May 26, 2019, 13:04 Strahil <hunter86_bg@yahoo.com> wrote:
Hi Leo, As you do not have a distributed volume , you can easily switch to replica 2 arbiter 1 or replica 3 volumes.
You can use the following for adding the bricks:
https://access.redhat.com/documentation/en-US/Red_Hat_Storage/2.1/html/Admin...
Best Regards, Strahil Nikoliv On May 26, 2019 10:54, Leo David <leoalex@gmail.com> wrote:
Hi Stahil, Thank you so much for yout input !
gluster volume info
Volume Name: engine Type: Distribute Volume ID: d7449fc2-cc35-4f80-a776-68e4a3dbd7e1 Status: Started Snapshot Count: 0 Number of Bricks: 1 Transport-type: tcp Bricks: Brick1: 192.168.80.191:/gluster_bricks/engine/engine Options Reconfigured: nfs.disable: on transport.address-family: inet storage.owner-uid: 36 storage.owner-gid: 36 features.shard: on performance.low-prio-threads: 32 performance.strict-o-direct: off network.remote-dio: off network.ping-timeout: 30 user.cifs: off performance.quick-read: off performance.read-ahead: off performance.io-cache: off cluster.eager-lock: enable Volume Name: ssd-samsung Type: Distribute Volume ID: 76576cc6-220b-4651-952d-99846178a19e Status: Started Snapshot Count: 0 Number of Bricks: 1 Transport-type: tcp Bricks: Brick1: 192.168.80.191:/gluster_bricks/sdc/data Options Reconfigured: cluster.eager-lock: enable performance.io-cache: off performance.read-ahead: off performance.quick-read: off user.cifs: off network.ping-timeout: 30 network.remote-dio: off performance.strict-o-direct: on performance.low-prio-threads: 32 features.shard: on storage.owner-gid: 36 storage.owner-uid: 36 transport.address-family: inet nfs.disable: on
The other two hosts will be 192.168.80.192/193 - this is gluster dedicated network over 10GB sfp+ switch. - host 2 wil have identical harware configuration with host 1 ( each disk is actually a raid0 array ) - host 3 has: - 1 ssd for OS - 1 ssd - for adding to engine volume in a full replica 3 - 2 ssd's in a raid 1 array to be added as arbiter for the data volume ( ssd-samsung ) So the plan is to have "engine" scaled in a full replica 3, and "ssd-samsung" scalled in a replica 3 arbitrated.
On Sun, May 26, 2019 at 10:34 AM Strahil <hunter86_bg@yahoo.com> wrote:
Hi Leo,
Gluster is quite smart, but in order to provide any hints , can you provide output of 'gluster volume info <glustervol>'. If you have 2 more systems , keep in mind that it is best to mirror the storage on the second replica (2 disks on 1 machine -> 2 disks on the new machine), while for the arbiter this is not neccessary.
What is your network and NICs ? Based on my experience , I can recommend at least 10 gbit/s interfase(s).
Best Regards, Strahil Nikolov On May 26, 2019 07:52, Leo David <leoalex@gmail.com> wrote:
Hello Everyone, Can someone help me to clarify this ? I have a single-node 4.2.8 installation ( only two gluster storage domains - distributed single drive volumes ). Now I just got two identintical servers and I would like to go for a 3 nodes bundle. Is it possible ( after joining the new nodes to the cluster ) to expand the existing volumes across the new nodes and change them to replica 3 arbitrated ? If so, could you share with me what would it be the procedure ? Thank you very much !
Leo
-- Best regards, Leo David
-- Best regards, Leo David
-- Best regards, Leo David

Hi all. I acrtually did something like that myself. I started out with a single node HC cluster. I then added another node (and plan to add a third). This is what I did: 1) Set up the new node. Make sure that you have all dependencies. (In my case I started with a Centos 8 machine, and installed vdms-gluster and gluster-ansible) 2) Configure the bricks. For this I just copied over hc_wizard_inventory.yml over from the first node, edited it to fit the second node, and just ran the gluster.infra role. 3) Expand the volume. In this case with the following command: gluster volume add-brick engine replica 2 <host>:/gluster_bricks/engine/engine 4) now just add the host as a hypervisor using the management console. I plan on adding a third node. Then I want to have full replica on the engine, and replica 2 + arbiter on the vmstore volume. Expanding gluster volumes, migrating from distributed to replicated and even replacing bricks etc. is rather easy in Gluster once you know how it works. I have even replaced all the servers on a live gluster cluster, without service interruption… Krist On Jul 18, 2019, 09:58 +0200, Leo David <leoalex@gmail.com>, wrote:
Hi, Looks like the only way arround would be to create a brand-new volume as replicated on other disks, and start moving the vms all around the place between volumes ? Cheers,
Leo
On Mon, May 27, 2019 at 1:53 PM Leo David <leoalex@gmail.com> wrote:
Hi, Any suggestions ? Thank you very much !
Leo
On Sun, May 26, 2019 at 4:38 PM Strahil Nikolov <hunter86_bg@yahoo.com> wrote:
Yeah, it seems different from the docs. I'm adding the gluster users list ,as they are more experienced into that.
@Gluster-users,
can you provide some hint how to add aditional replicas to the below volumes , so they become 'replica 2 arbiter 1' or 'replica 3' type volumes ?
Best Regards, Strahil Nikolov
В неделя, 26 май 2019 г., 15:16:18 ч. Гринуич+3, Leo David <leoalex@gmail.com> написа:
Thank you Strahil, The engine and ssd-samsung are distributed... So these are the ones that I need to have replicated accross new nodes. I am not very sure about the procedure to accomplish this. Thanks,
Leo
On Sun, May 26, 2019, 13:04 Strahil <hunter86_bg@yahoo.com> wrote:
Hi Leo, As you do not have a distributed volume , you can easily switch to replica 2 arbiter 1 or replica 3 volumes. You can use the following for adding the bricks: https://access.redhat.com/documentation/en-US/Red_Hat_Storage/2.1/html/Admin... Best Regards, Strahil Nikoliv On May 26, 2019 10:54, Leo David <leoalex@gmail.com> wrote: > Hi Stahil, > Thank you so much for yout input ! > > gluster volume info > > > Volume Name: engine > Type: Distribute > Volume ID: d7449fc2-cc35-4f80-a776-68e4a3dbd7e1 > Status: Started > Snapshot Count: 0 > Number of Bricks: 1 > Transport-type: tcp > Bricks: > Brick1: 192.168.80.191:/gluster_bricks/engine/engine > Options Reconfigured: > nfs.disable: on > transport.address-family: inet > storage.owner-uid: 36 > storage.owner-gid: 36 > features.shard: on > performance.low-prio-threads: 32 > performance.strict-o-direct: off > network.remote-dio: off > network.ping-timeout: 30 > user.cifs: off > performance.quick-read: off > performance.read-ahead: off > performance.io-cache: off > cluster.eager-lock: enable > Volume Name: ssd-samsung > Type: Distribute > Volume ID: 76576cc6-220b-4651-952d-99846178a19e > Status: Started > Snapshot Count: 0 > Number of Bricks: 1 > Transport-type: tcp > Bricks: > Brick1: 192.168.80.191:/gluster_bricks/sdc/data > Options Reconfigured: > cluster.eager-lock: enable > performance.io-cache: off > performance.read-ahead: off > performance.quick-read: off > user.cifs: off > network.ping-timeout: 30 > network.remote-dio: off > performance.strict-o-direct: on > performance.low-prio-threads: 32 > features.shard: on > storage.owner-gid: 36 > storage.owner-uid: 36 > transport.address-family: inet > nfs.disable: on > > The other two hosts will be 192.168.80.192/193 - this is gluster dedicated network over 10GB sfp+ switch. > - host 2 wil have identical harware configuration with host 1 ( each disk is actually a raid0 array ) > - host 3 has: > - 1 ssd for OS > - 1 ssd - for adding to engine volume in a full replica 3 > - 2 ssd's in a raid 1 array to be added as arbiter for the data volume ( ssd-samsung ) > So the plan is to have "engine" scaled in a full replica 3, and "ssd-samsung" scalled in a replica 3 arbitrated. > > > > > On Sun, May 26, 2019 at 10:34 AM Strahil <hunter86_bg@yahoo.com> wrote: > > Hi Leo, > > Gluster is quite smart, but in order to provide any hints , can you provide output of 'gluster volume info <glustervol>'. > > If you have 2 more systems , keep in mind that it is best to mirror the storage on the second replica (2 disks on 1 machine -> 2 disks on the new machine), while for the arbiter this is not neccessary. > > What is your network and NICs ? Based on my experience , I can recommend at least 10 gbit/s interfase(s). > > Best Regards, > > Strahil Nikolov > > On May 26, 2019 07:52, Leo David <leoalex@gmail.com> wrote: > > > Hello Everyone, > > > Can someone help me to clarify this ? > > > I have a single-node 4.2.8 installation ( only two gluster storage domains - distributed single drive volumes ). Now I just got two identintical servers and I would like to go for a 3 nodes bundle. > > > Is it possible ( after joining the new nodes to the cluster ) to expand the existing volumes across the new nodes and change them to replica 3 arbitrated ? > > > If so, could you share with me what would it be the procedure ? > > > Thank you very much ! > > > > > > Leo > > > -- > Best regards, Leo David
-- Best regards, Leo David
-- Best regards, Leo David _______________________________________________ Gluster-users mailing list Gluster-users@gluster.org https://lists.gluster.org/mailman/listinfo/gluster-users
participants (4)
-
Krist van Besien
-
Leo David
-
Strahil
-
Strahil Nikolov