We have been using zabbix for tracking gluster but that works because
we are using zabbix for the rest of our monitoring of things like
network and disk IO.
One thing to track that is not part of the usual suspects is the heal
counts.
They should always be 0 unless you have a problem
sorry, I didn't notice you had already looked at gstatus.
Nonetheless with its JSON output you certainly cover the issues you
described
i.e. "When Brick went down (crash, failure, shutdown), node failure,
peering issue, on-going healing"
which is how we use it.
-wk
On 10/27/2020 9:33 AM,
https://github.com/gluster/gstatus
we run this from an ansible driven cronjob and check for the healthy
signal in status, as well as looking for healing files that seem to persist.
We have a number of gluster clusters and we have found its warnings both
useful and timely.
-wk
On
Hello
How do you keep track of the health status of your Gluster volumes? When
Brick went down (crash, failure, shutdown), node failure, peering issue,
on-going healing?
Gluster Tendrl is complex and sometimes it's broken, Prometheus exporter
still lacking, gstatus is basic.
Currently, to
Completed
--
- Configuration format is now changed to `toml` format for ease of
use. Usage is upadated in README
PR: https://github.com/gluster/gluster-prometheus/pull/16
- Enabled Travis tests to validate incoming PRs and added build status
in README
PRs:
## Quick start:
```
cd $GOPATH/src/github.com/gluster
git clone https://github.com/gluster/gluster-prometheus.git
cd gluster-prometheus
PREFIX=/usr make
PREFIX=/usr make install
# Enable and start using,
systemctl enable gluster-exporter
systemctl start gluster-exporter
```
Note: By default
Hi,
I installed PCP(http://www.pcp.io/man/man1/pmdagluster.1.html) on gluster
nodes and enabled to collect data using Install from
/var/lib/pcp/pmdas/gluster/.
I don't see any data for bricks using pminfo command as below.
# pminfo -f gluster.brick
gluster.brick.latency.fgetxattr.count