Additionally the brick log file of the same brick would be required. Please look for if brick process went down or crashed. Doing a volume start force should resolve the issue.
On Wed, 13 Sep 2017 at 16:28, Gaurav Yadav <gya...@redhat.com> wrote: > Please send me the logs as well i.e glusterd.logs and cmd_history.log. > > > On Wed, Sep 13, 2017 at 1:45 PM, lejeczek <pelj...@yahoo.co.uk> wrote: > >> >> >> On 13/09/17 06:21, Gaurav Yadav wrote: >> > Please provide the output of gluster volume info, gluster volume status >>> and gluster peer status. >>> >>> Apart from above info, please provide glusterd logs, cmd_history.log. >>> >>> Thanks >>> Gaurav >>> >>> On Tue, Sep 12, 2017 at 2:22 PM, lejeczek <pelj...@yahoo.co.uk <mailto: >>> pelj...@yahoo.co.uk>> wrote: >>> >>> hi everyone >>> >>> I have 3-peer cluster with all vols in replica mode, 9 >>> vols. >>> What I see, unfortunately, is one brick fails in one >>> vol, when it happens it's always the same vol on the >>> same brick. >>> Command: gluster vol status $vol - would show brick >>> not online. >>> Restarting glusterd with systemclt does not help, only >>> system reboot seem to help, until it happens, next time. >>> >>> How to troubleshoot this weird misbehaviour? >>> many thanks, L. >>> >>> . >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users@gluster.org >>> >> <mailto:Gluster-users@gluster.org> >>> http://lists.gluster.org/mailman/listinfo/gluster-users >>> <http://lists.gluster.org/mailman/listinfo/gluster-users> >>> >>> >>> >> hi, here: >> >> $ gluster vol info C-DATA >> >> Volume Name: C-DATA >> Type: Replicate >> Volume ID: 18ffba73-532e-4a4d-84da-fceea52f8c2e >> Status: Started >> Snapshot Count: 0 >> Number of Bricks: 1 x 3 = 3 >> Transport-type: tcp >> Bricks: >> Brick1: 10.5.6.49:/__.aLocalStorages/0/0-GLUSTERs/0GLUSTER-C-DATA >> Brick2: 10.5.6.100:/__.aLocalStorages/0/0-GLUSTERs/0GLUSTER-C-DATA >> Brick3: 10.5.6.32:/__.aLocalStorages/0/0-GLUSTERs/0GLUSTER-C-DATA >> Options Reconfigured: >> performance.md-cache-timeout: 600 >> performance.cache-invalidation: on >> performance.stat-prefetch: on >> features.cache-invalidation-timeout: 600 >> features.cache-invalidation: on >> performance.io-thread-count: 64 >> performance.cache-size: 128MB >> cluster.self-heal-daemon: enable >> features.quota-deem-statfs: on >> changelog.changelog: on >> geo-replication.ignore-pid-check: on >> geo-replication.indexing: on >> features.inode-quota: on >> features.quota: on >> performance.readdir-ahead: on >> nfs.disable: on >> transport.address-family: inet >> performance.cache-samba-metadata: on >> >> >> $ gluster vol status C-DATA >> Status of volume: C-DATA >> Gluster process TCP Port RDMA Port Online >> Pid >> >> ------------------------------------------------------------------------------ >> Brick 10.5.6.49:/__.aLocalStorages/0/0-GLUS >> TERs/0GLUSTER-C-DATA N/A N/A N N/A >> Brick 10.5.6.100:/__.aLocalStorages/0/0-GLU >> STERs/0GLUSTER-C-DATA 49152 0 Y 9376 >> Brick 10.5.6.32:/__.aLocalStorages/0/0-GLUS >> TERs/0GLUSTER-C-DATA 49152 0 Y 8638 >> Self-heal Daemon on localhost N/A N/A Y 387879 >> Quota Daemon on localhost N/A N/A Y 387891 >> Self-heal Daemon on rider.private.ccnr.ceb. >> private.cam.ac.uk N/A N/A Y 16439 >> Quota Daemon on rider.private.ccnr.ceb.priv >> ate.cam.ac.uk N/A N/A Y 16451 >> Self-heal Daemon on 10.5.6.32 N/A N/A Y 7708 >> Quota Daemon on 10.5.6.32 N/A N/A Y 8623 >> Self-heal Daemon on 10.5.6.17 N/A N/A Y 20549 >> Quota Daemon on 10.5.6.17 N/A N/A Y 9337 >> >> Task Status of Volume C-DATA >> >> ------------------------------------------------------------------------------ >> There are no active volume tasks > > >> >> >> >> . >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users@gluster.org >> http://lists.gluster.org/mailman/listinfo/gluster-users >> > _______________________________________________ > Gluster-users mailing list > Gluster-users@gluster.org > http://lists.gluster.org/mailman/listinfo/gluster-users -- --Atin
_______________________________________________ Gluster-users mailing list Gluster-users@gluster.org http://lists.gluster.org/mailman/listinfo/gluster-users