Re: [Gluster-users] Brick Goes Offline After server reboot/Or Gluster Container is restarted, on which a gluster node is running

2020-03-01 Thread Sanju Rakonde
Hi Rifat,

I suspect you are hitting
https://bugzilla.redhat.com/show_bug.cgi?id=1773856. This bug has been
fixed in master and will be backported to release branches soon.

Although the bug says, "with volume set operation when a node is down we
see the issue", according to the RCA of the bug it can be hit when there is
any kind of configuration change in volumes.

Hope this helps.

On Fri, Feb 28, 2020 at 11:52 PM Strahil Nikolov 
wrote:

> On February 28, 2020 4:49:45 PM GMT+02:00, Rifat Ucal 
> wrote:
> >Hi Gluster Team,
> >
> >
> >I am trying to implment gluster fs in podman containers, which is
> >running except the problems described below.
> >
> >
> >My observations:
> >
> >- The bricks on the server is gong offline when one of the podman
> >container is restarted or the appropriate server is rebooted.
> >
> >- Althoug the status of the bricks are offline, the replication seems
> >to be working, as data will be replicated.
> >
> >- I see that the replicated data will also be replicated on arbiter
> >node, where I was expecting to see only meta data.
> >
> >
> >
> >My configuration.
> >
> >I created glusterfs for replication in 3 nodes on centos7 but in podman
> >containers
> >
> >The containers in the first and second nodes should be normal
> >replication and 3rd node arbiter node.
> >
> >After creation replication and enabling heal processes I can see also
> >that 3rd node is marked as arbiter node.
> >
> >According to description of arbiter, the arbiter node should store only
> >metadata but in my configuration the replicated data will be stored in
> >all bricks including arbiter node.
> >
> >
> >
> >Questions:
> >
> >When rebooting one of the server or restarting one of the glusterfs
> >container the restarted container is not going online until gluster
> >volume is stopped and started again. Is it a solution inbetween to
> >resolve this problem?
> >
> >-why Arbiter node stores all the data, allthough it should only have
> >some metadata to restore the replicated data on other nodes. I would
> >not have problem that replication is done in all three nodes. I just
> >need to know
> >
> >- Can you give me feedback, whether some one experience or similar
> >porblems with glusterfs implemented in podman containers?
> >
> >
> >Here are my configurations:
> >
> >on all containers I have :CentOS Linux release 7.7.1908 glusterfs
> >version 7.3 and systemctl is enabled for glusterd service
> >
> >
> >My gluster volume creation:
> >
> >gluster volume create cgvol1 replica 2 arbiter 1 transport tcp
> >avm1:/cbricks/brick1/data avm2:/cbricks/brick1/data
> >dvm1:/cbricks/brick1/data force
> >
> >
> >gluster peer status excuted on avm2:
> >Number of Peers: 2
> >
> >Hostname: avm1
> >Uuid: 5d1dc6a7-8f34-45a3-a7c9-c69c442b66dc
> >State: Peer in Cluster (Connected)
> >
> >Hostname: dvm1
> >Uuid: 310ffd58-28ab-43f1-88d3-1e381bd46ab3
> >State: Peer in Cluster (Connected)
> >
> >
> >gluster volume info
> >
> >Volume Name: cgvol1
> >Type: Replicate
> >Volume ID: da975178-b68f-410c-884c-a7f635e4381a
> >Status: Started
> >Snapshot Count: 0
> >Number of Bricks: 1 x (2 + 1) = 3
> >Transport-type: tcp
> >Bricks:
> >Brick1: arvm1:/cbricks/brick1/data
> >Brick2: avm2:/cbricks/brick1/data
> >Brick3: devm1:/cbricks/brick1/data (arbiter)
> >Options Reconfigured:
> >cluster.self-heal-daemon: on
> >cluster.entry-self-heal: on
> >cluster.metadata-self-heal: on
> >cluster.data-self-heal: on
> >transport.address-family: inet
> >storage.fips-mode-rchecksum: on
> >nfs.disable: on
> >performance.client-io-threads: off
> >
> >
> >gluster volume status
> >Status of volume: cgvol1
> >Gluster process TCP Port RDMA Port Online Pid
> >
>
> >--
> >Brick avm1:/cbricks/brick1/data 49152 0 Y 516
> >Brick avm2:/cbricks/brick1/data 49152 0 Y 353
> >Brick dvm1:/cbricks/brick1/data 49152 0 Y 572
> >Self-heal Daemon on localhost N/A N/A Y 537
> >Self-heal Daemon on dvm1 N/A N/A Y 593
> >Self-heal Daemon on avm2 N/A N/A Y 374
> >
> >Task Status of Volume cgvol1
>
> >--
> >There are no active volume tasks
> >
> >gluster volume heal cgvol1 info
> >Brick avm1:/cbricks/brick1/data
> >Status: Connected
> >Number of entries: 0
> >
> >Brick avm2:/cbricks/brick1/data
> >Status: Connected
> >Number of entries: 0
> >
> >Brick dvm1:/cbricks/brick1/data
> >Status: Connected
> >Number of entries: 0
> >
> >
> >
> >Best Regards,
> >
> >Rifat Ucal
> >
> >
> >
> >> Jorick Astrego  hat am 14. Februar 2020 um 10:10
> >geschrieben:
> >>
> >>
> >> Hi,
> >>
> >> It looks like you have a two node setup?
> >>
> >> Then it's expected as with two nodes you don't have quorum and
> >this can lead to split brains.
> >>
> >> To have HA, add another node or an arbiter node.
> >>
> >>
> >
> https://docs.gluster.org/en/latest/Administrator%20Guide/arbiter-volumes-and-quorum/
> >>
> >> You can also modify the quoru

Re: [Gluster-users] Brick Goes Offline After server reboot/Or Gluster Container is restarted, on which a gluster node is running

2020-02-28 Thread Strahil Nikolov
On February 28, 2020 4:49:45 PM GMT+02:00, Rifat Ucal  wrote:
>Hi Gluster Team,
>
>
>I am trying to implment gluster fs in podman containers, which is
>running except the problems described below.
>
>
>My observations:
>
>- The bricks on the server is gong offline when one of the podman
>container is restarted or the appropriate server is rebooted.
>
>- Althoug the status of the bricks are offline, the replication seems
>to be working, as data will be replicated.
>
>- I see that the replicated data will also be replicated on arbiter
>node, where I was expecting to see only meta data.
>
> 
>
>My configuration.
>
>I created glusterfs for replication in 3 nodes on centos7 but in podman
>containers
>
>The containers in the first and second nodes should be normal
>replication and 3rd node arbiter node.
>
>After creation replication and enabling heal processes I can see also
>that 3rd node is marked as arbiter node.
>
>According to description of arbiter, the arbiter node should store only
>metadata but in my configuration the replicated data will be stored in
>all bricks including arbiter node.
>
> 
>
>Questions:
>
>When rebooting one of the server or restarting one of the glusterfs
>container the restarted container is not going online until gluster
>volume is stopped and started again. Is it a solution inbetween to
>resolve this problem?
>
>-why Arbiter node stores all the data, allthough it should only have
>some metadata to restore the replicated data on other nodes. I would
>not have problem that replication is done in all three nodes. I just
>need to know 
>
>- Can you give me feedback, whether some one experience or similar
>porblems with glusterfs implemented in podman containers?
>
>
>Here are my configurations:
>
>on all containers I have :CentOS Linux release 7.7.1908 glusterfs
>version 7.3 and systemctl is enabled for glusterd service
>
>
>My gluster volume creation:
>
>gluster volume create cgvol1 replica 2 arbiter 1 transport tcp
>avm1:/cbricks/brick1/data avm2:/cbricks/brick1/data
>dvm1:/cbricks/brick1/data force
>
>
>gluster peer status excuted on avm2:
>Number of Peers: 2
>
>Hostname: avm1
>Uuid: 5d1dc6a7-8f34-45a3-a7c9-c69c442b66dc
>State: Peer in Cluster (Connected)
>
>Hostname: dvm1
>Uuid: 310ffd58-28ab-43f1-88d3-1e381bd46ab3
>State: Peer in Cluster (Connected)
>
>
>gluster volume info
>
>Volume Name: cgvol1
>Type: Replicate
>Volume ID: da975178-b68f-410c-884c-a7f635e4381a
>Status: Started
>Snapshot Count: 0
>Number of Bricks: 1 x (2 + 1) = 3
>Transport-type: tcp
>Bricks:
>Brick1: arvm1:/cbricks/brick1/data
>Brick2: avm2:/cbricks/brick1/data
>Brick3: devm1:/cbricks/brick1/data (arbiter)
>Options Reconfigured:
>cluster.self-heal-daemon: on
>cluster.entry-self-heal: on
>cluster.metadata-self-heal: on
>cluster.data-self-heal: on
>transport.address-family: inet
>storage.fips-mode-rchecksum: on
>nfs.disable: on
>performance.client-io-threads: off
>
>
>gluster volume status
>Status of volume: cgvol1
>Gluster process TCP Port RDMA Port Online Pid
>
>--
>Brick avm1:/cbricks/brick1/data 49152 0 Y 516
>Brick avm2:/cbricks/brick1/data 49152 0 Y 353
>Brick dvm1:/cbricks/brick1/data 49152 0 Y 572
>Self-heal Daemon on localhost N/A N/A Y 537
>Self-heal Daemon on dvm1 N/A N/A Y 593
>Self-heal Daemon on avm2 N/A N/A Y 374
>
>Task Status of Volume cgvol1
>--
>There are no active volume tasks
>
>gluster volume heal cgvol1 info
>Brick avm1:/cbricks/brick1/data
>Status: Connected
>Number of entries: 0
>
>Brick avm2:/cbricks/brick1/data
>Status: Connected
>Number of entries: 0
>
>Brick dvm1:/cbricks/brick1/data
>Status: Connected
>Number of entries: 0
>
>
>
>Best Regards,
>
>Rifat Ucal
>
>
>
>> Jorick Astrego  hat am 14. Februar 2020 um 10:10
>geschrieben:
>> 
>> 
>> Hi,
>> 
>> It looks like you have a two node setup?
>> 
>> Then it's expected as with two nodes you don't have quorum and
>this can lead to split brains.
>> 
>> To have HA, add another node or an arbiter node.
>> 
>>
>https://docs.gluster.org/en/latest/Administrator%20Guide/arbiter-volumes-and-quorum/
>> 
>> You can also modify the quorum but then you shouldn't be too
>attachted to the data you have on it.
>> 
>> Regards, Jorick
>> 
>> On 2/14/20 9:27 AM, Cloud Udupi wrote:
>> 
>> > > Hi,
>> > 
>> > I am new to glusterfs. I have used this guide on
>CentOS-7.6. 
>> >
>https://microdevsys.com/wp/glusterfs-configuration-and-setup-w-nfs-ganesha-for-an-ha-nfs-cluster/
>> > 
>> > glusterfs —version
>> > glusterfs 7.2
>> > 
>> > Firewall is disabled. Self heal is enabled.
>> > Everything works fine until I reboot one of the servers.
>When the server reboots the brick doesn't come online.
>> > 
>> > gluster volume status
>> > 
>> > Status of volume: gv01
>> > Gluster process

[Gluster-users] Brick Goes Offline After server reboot/Or Gluster Container is restarted, on which a gluster node is running

2020-02-28 Thread Rifat Ucal
Hi Gluster Team,


I am trying to implment gluster fs in podman containers, which is running 
except the problems described below.


My observations:

- The bricks on the server is gong offline when one of the podman container is 
restarted or the appropriate server is rebooted.

- Althoug the status of the bricks are offline, the replication seems to be 
working, as data will be replicated.

- I see that the replicated data will also be replicated on arbiter node, where 
I was expecting to see only meta data.

 

My configuration.

I created glusterfs for replication in 3 nodes on centos7 but in podman 
containers

The containers in the first and second nodes should be normal replication and 
3rd node arbiter node.

After creation replication and enabling heal processes I can see also that 3rd 
node is marked as arbiter node.

According to description of arbiter, the arbiter node should store only 
metadata but in my configuration the replicated data will be stored in all 
bricks including arbiter node.

 

Questions:

When rebooting one of the server or restarting one of the glusterfs container 
the restarted container is not going online until gluster volume is stopped and 
started again. Is it a solution inbetween to resolve this problem?

-why Arbiter node stores all the data, allthough it should only have some 
metadata to restore the replicated data on other nodes. I would not have 
problem that replication is done in all three nodes. I just need to know 

- Can you give me feedback, whether some one experience or similar porblems 
with glusterfs implemented in podman containers?


Here are my configurations:

on all containers I have :CentOS Linux release 7.7.1908 glusterfs version 7.3 
and systemctl is enabled for glusterd service


My gluster volume creation:

gluster volume create cgvol1 replica 2 arbiter 1 transport tcp 
avm1:/cbricks/brick1/data avm2:/cbricks/brick1/data dvm1:/cbricks/brick1/data 
force


gluster peer status excuted on avm2:
Number of Peers: 2

Hostname: avm1
Uuid: 5d1dc6a7-8f34-45a3-a7c9-c69c442b66dc
State: Peer in Cluster (Connected)

Hostname: dvm1
Uuid: 310ffd58-28ab-43f1-88d3-1e381bd46ab3
State: Peer in Cluster (Connected)


gluster volume info

Volume Name: cgvol1
Type: Replicate
Volume ID: da975178-b68f-410c-884c-a7f635e4381a
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x (2 + 1) = 3
Transport-type: tcp
Bricks:
Brick1: arvm1:/cbricks/brick1/data
Brick2: avm2:/cbricks/brick1/data
Brick3: devm1:/cbricks/brick1/data (arbiter)
Options Reconfigured:
cluster.self-heal-daemon: on
cluster.entry-self-heal: on
cluster.metadata-self-heal: on
cluster.data-self-heal: on
transport.address-family: inet
storage.fips-mode-rchecksum: on
nfs.disable: on
performance.client-io-threads: off


gluster volume status
Status of volume: cgvol1
Gluster process TCP Port RDMA Port Online Pid

--
Brick avm1:/cbricks/brick1/data 49152 0 Y 516
Brick avm2:/cbricks/brick1/data 49152 0 Y 353
Brick dvm1:/cbricks/brick1/data 49152 0 Y 572
Self-heal Daemon on localhost N/A N/A Y 537
Self-heal Daemon on dvm1 N/A N/A Y 593
Self-heal Daemon on avm2 N/A N/A Y 374

Task Status of Volume cgvol1
--
There are no active volume tasks

gluster volume heal cgvol1 info
Brick avm1:/cbricks/brick1/data
Status: Connected
Number of entries: 0

Brick avm2:/cbricks/brick1/data
Status: Connected
Number of entries: 0

Brick dvm1:/cbricks/brick1/data
Status: Connected
Number of entries: 0



Best Regards,

Rifat Ucal



> Jorick Astrego  hat am 14. Februar 2020 um 10:10 
> geschrieben:
> 
> 
> Hi,
> 
> It looks like you have a two node setup?
> 
> Then it's expected as with two nodes you don't have quorum and this can 
> lead to split brains.
> 
> To have HA, add another node or an arbiter node.
> 
> 
> https://docs.gluster.org/en/latest/Administrator%20Guide/arbiter-volumes-and-quorum/
> 
> You can also modify the quorum but then you shouldn't be too attachted to 
> the data you have on it.
> 
> Regards, Jorick
> 
> On 2/14/20 9:27 AM, Cloud Udupi wrote:
> 
> > > Hi,
> > 
> > I am new to glusterfs. I have used this guide on CentOS-7.6. 
> > 
> > https://microdevsys.com/wp/glusterfs-configuration-and-setup-w-nfs-ganesha-for-an-ha-nfs-cluster/
> > 
> > glusterfs —version
> > glusterfs 7.2
> > 
> > Firewall is disabled. Self heal is enabled.
> > Everything works fine until I reboot one of the servers. When the 
> > server reboots the brick doesn't come online.
> > 
> > gluster volume status
> > 
> > Status of volume: gv01
> > Gluster process TCP Port  RDMA Port  
> > Online  Pid
> > 
> > --
> > Brick server1:/bricks/0/gv0 

Re: [Gluster-users] Brick Goes Offline After server reboot.

2020-02-14 Thread Jorick Astrego
Hi,

It looks like you have a two node setup?

Then it's expected as with two nodes you don't have quorum and this can
lead to split brains.

To have HA, add another node or an arbiter node.

https://docs.gluster.org/en/latest/Administrator%20Guide/arbiter-volumes-and-quorum/

You can also modify the quorum but then you shouldn't be too attachted
to the data you have on it.

Regards, Jorick

On 2/14/20 9:27 AM, Cloud Udupi wrote:
> Hi,
>
> I am new to glusterfs. I have used this guide on CentOS-7.6. 
> https://microdevsys.com/wp/glusterfs-configuration-and-setup-w-nfs-ganesha-for-an-ha-nfs-cluster/
>
>
> glusterfs —version
>
> glusterfs 7.2
>
>
> Firewall is disabled. Self heal is enabled.
> Everything works fine until I reboot one of the servers. When the
> server reboots the brick doesn't come online.
>
> /*gluster volume status*/
>
> /
> /
>
> /Status of volume: gv01/
>
> /Gluster process                             TCP Port  RDMA Port 
> Online  Pid/
>
> /--/
>
> /Brick server1:/bricks/0/gv0                 N/A       N/A        N  
>     N/A  /
>
> /Brick server2:/bricks/0/gv0                 49152     0          Y  
>     99870/
>
> /Self-heal Daemon on localhost               N/A       N/A        Y  
>     109802/
>
> /Self-heal Daemon on server1                 N/A       N/A        Y  
>     2142 /
>
> / /
>
> /Task Status of Volume gv01/
>
> /--/
>
> /There are no active volume tasks/
>
>
> */gluster volume heal gv01/*
>
> */
> /*
>
> /Launching heal operation to perform index self heal on volume gv01
> has been unsuccessful:/
>
> //
>
> /Glusterd Syncop Mgmt brick op 'Heal' failed. Please check glustershd
> log file for details/.
>
>
> */gluster volume heal gv01 info/*
>
> */
> /*
>
> / gluster volume heal gv01 info/
>
> /Brick server1:/bricks/0/gv0/
>
> /Status: Transport endpoint is not connected/
>
> //
>
> /Number of entries: -/
>
> /
> /
>
> *When I do"gluster volume start gv01 force" brick starts.*
>
> *
> *
>
> I want the brick to come online automatically after the reboot. I have
> attached log file.
>
> Please help.
>
>
> Regards,
>
> Mark.
>
>
> 
>
> Community Meeting Calendar:
>
> APAC Schedule -
> Every 2nd and 4th Tuesday at 11:30 AM IST
> Bridge: https://bluejeans.com/441850968
>
> NA/EMEA Schedule -
> Every 1st and 3rd Tuesday at 01:00 PM EDT
> Bridge: https://bluejeans.com/441850968
>
> Gluster-users mailing list
> Gluster-users@gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users




Met vriendelijke groet, With kind regards,

Jorick Astrego

Netbulae Virtualization Experts 



Tel: 053 20 30 270  i...@netbulae.euStaalsteden 4-3A
KvK 08198180
Fax: 053 20 30 271  www.netbulae.eu 7547 TA Enschede
BTW NL821234584B01





Community Meeting Calendar:

APAC Schedule -
Every 2nd and 4th Tuesday at 11:30 AM IST
Bridge: https://bluejeans.com/441850968

NA/EMEA Schedule -
Every 1st and 3rd Tuesday at 01:00 PM EDT
Bridge: https://bluejeans.com/441850968

Gluster-users mailing list
Gluster-users@gluster.org
https://lists.gluster.org/mailman/listinfo/gluster-users


[Gluster-users] Brick Goes Offline After server reboot.

2020-02-14 Thread Cloud Udupi
Hi,

I am new to glusterfs. I have used this guide on CentOS-7.6.
https://microdevsys.com/wp/glusterfs-configuration-and-setup-w-nfs-ganesha-for-an-ha-nfs-cluster/


glusterfs —version

glusterfs 7.2


Firewall is disabled. Self heal is enabled.
Everything works fine until I reboot one of the servers. When the server
reboots the brick doesn't come online.

*gluster volume status*


*Status of volume: gv01*

*Gluster process TCP Port  RDMA Port  Online
Pid*

*--*

*Brick server1:/bricks/0/gv0 N/A   N/AN
N/A  *

*Brick server2:/bricks/0/gv0 49152 0  Y
99870*

*Self-heal Daemon on localhost   N/A   N/AY
109802*

*Self-heal Daemon on server1 N/A   N/AY
2142 *



*Task Status of Volume gv01*

*--*

*There are no active volume tasks*


*gluster volume heal gv01*


*Launching heal operation to perform index self heal on volume gv01 has
been unsuccessful:*

*Glusterd Syncop Mgmt brick op 'Heal' failed. Please check glustershd log
file for details*.

*gluster volume heal gv01 info*


* gluster volume heal gv01 info*

*Brick server1:/bricks/0/gv0*

*Status: Transport endpoint is not connected*

*Number of entries: -*


*When I do "gluster volume start gv01 force" brick starts.*


I want the brick to come online automatically after the reboot. I have
attached log file.

Please help.


Regards,

Mark.


logs.pdf
Description: Adobe PDF document


Community Meeting Calendar:

APAC Schedule -
Every 2nd and 4th Tuesday at 11:30 AM IST
Bridge: https://bluejeans.com/441850968

NA/EMEA Schedule -
Every 1st and 3rd Tuesday at 01:00 PM EDT
Bridge: https://bluejeans.com/441850968

Gluster-users mailing list
Gluster-users@gluster.org
https://lists.gluster.org/mailman/listinfo/gluster-users