Hi Stahil,
Thank you so much for yout input !

 gluster volume info


Volume Name: engine
Type: Distribute
Volume ID: d7449fc2-cc35-4f80-a776-68e4a3dbd7e1
Status: Started
Snapshot Count: 0
Number of Bricks: 1
Transport-type: tcp
Bricks:
Brick1: 192.168.80.191:/gluster_bricks/engine/engine
Options Reconfigured:
nfs.disable: on
transport.address-family: inet
storage.owner-uid: 36
storage.owner-gid: 36
features.shard: on
performance.low-prio-threads: 32
performance.strict-o-direct: off
network.remote-dio: off
network.ping-timeout: 30
user.cifs: off
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
cluster.eager-lock: enable
Volume Name: ssd-samsung
Type: Distribute
Volume ID: 76576cc6-220b-4651-952d-99846178a19e
Status: Started
Snapshot Count: 0
Number of Bricks: 1
Transport-type: tcp
Bricks:
Brick1: 192.168.80.191:/gluster_bricks/sdc/data
Options Reconfigured:
cluster.eager-lock: enable
performance.io-cache: off
performance.read-ahead: off
performance.quick-read: off
user.cifs: off
network.ping-timeout: 30
network.remote-dio: off
performance.strict-o-direct: on
performance.low-prio-threads: 32
features.shard: on
storage.owner-gid: 36
storage.owner-uid: 36
transport.address-family: inet
nfs.disable: on

The other two hosts will be 192.168.80.192/193  - this is gluster dedicated
network over 10GB sfp+ switch.
- host 2 wil have identical harware configuration with host 1 ( each disk
is actually a raid0 array )
- host 3 has:
   -  1 ssd for OS
   -  1 ssd - for adding to engine volume in a full replica 3
   -  2 ssd's in a raid 1 array to be added as arbiter for the data volume
( ssd-samsung )
So the plan is to have "engine"  scaled in a full replica 3,  and
"ssd-samsung" scalled in a replica 3 arbitrated.




On Sun, May 26, 2019 at 10:34 AM Strahil <[email protected]> wrote:

> Hi Leo,
>
> Gluster is quite smart, but in order to provide any hints , can you
> provide output of 'gluster volume info <glustervol>'.
> If you have 2 more systems , keep in mind that it is best to mirror the
> storage on the second replica (2 disks on 1 machine -> 2 disks on the new
> machine), while for the arbiter this is not neccessary.
>
> What is your network and NICs ? Based on my experience , I can recommend
> at least 10 gbit/s  interfase(s).
>
> Best Regards,
> Strahil Nikolov
> On May 26, 2019 07:52, Leo David <[email protected]> wrote:
>
> Hello Everyone,
> Can someone help me to clarify this ?
> I have a single-node 4.2.8 installation ( only two gluster storage domains
> - distributed  single drive volumes ). Now I just got two identintical
> servers and I would like to go for a 3 nodes bundle.
> Is it possible ( after joining the new nodes to the cluster ) to expand
> the existing volumes across the new nodes and change them to replica 3
> arbitrated ?
> If so, could you share with me what would it be the procedure ?
> Thank you very much !
>
> Leo
>
>

-- 
Best regards, Leo David
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/PJ2OO6SNVG4VQZDLJEEEJPTGLPZVQMUV/

Reply via email to