Re: [ceph-users] Inconsistent PGs every few days

2018-08-07 Thread Konstantin Shalygin

Hi,

I run a cluster with 7 OSD. The cluster has no much traffic on it. But
every few days, I get a HEALTH_ERR, because of inconsistent PGs:

root at Sam   ~ # ceph 
status
cluster:
  id: c4bfc288-8ba8-4c3a-b3a6-ed95503f50b7
  health: HEALTH_ERR
  3 scrub errors
  Possible data damage: 3 pgs inconsistent

services:
  mon: 1 daemons, quorum mon1
  mgr: ceph-osd1(active)
  mds: FS-1/1/1 up  {0=ceph-osd1=up:active}
  osd: 11 osds: 8 up, 7 in
  rgw: 1 daemon active

data:
  pools:   6 pools, 168 pgs
  objects: 901.8 k objects, 2.6 TiB
  usage: 7.9 TiB used, 7.4 TiB / 15 TiB avail
  pgs: 165 active+clean
   3   active+clean+inconsistent

io:
  client:   641 KiB/s wr, 0 op/s rd, 3 op/s wr


root at Sam   ~ # ceph 
health detail
HEALTH_ERR 3 scrub errors; Possible data damage: 3 pgs inconsistent
OSD_SCRUB_ERRORS 3 scrub errors
PG_DAMAGED Possible data damage: 3 pgs inconsistent
  pg 5.1d is active+clean+inconsistent, acting [6,8,3]
  pg 5.20 is active+clean+inconsistent, acting [3,9,0]
  pg 5.4a is active+clean+inconsistent, acting [6,3,7]

What's the reason for this problem? How can I analyse it?



Dimitri, what is your hardware? Your RAM is ECC?



k

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Inconsistent PGs every few days

2018-08-03 Thread Dimitri Roschkowski

Hi,

I run a cluster with 7 OSD. The cluster has no much traffic on it. But 
every few days, I get a HEALTH_ERR, because of inconsistent PGs:


root@Sam ~ # ceph status
  cluster:
id: c4bfc288-8ba8-4c3a-b3a6-ed95503f50b7
health: HEALTH_ERR
3 scrub errors
Possible data damage: 3 pgs inconsistent

  services:
mon: 1 daemons, quorum mon1
mgr: ceph-osd1(active)
mds: FS-1/1/1 up  {0=ceph-osd1=up:active}
osd: 11 osds: 8 up, 7 in
rgw: 1 daemon active

  data:
pools:   6 pools, 168 pgs
objects: 901.8 k objects, 2.6 TiB
usage: 7.9 TiB used, 7.4 TiB / 15 TiB avail
pgs: 165 active+clean
 3   active+clean+inconsistent

  io:
client:   641 KiB/s wr, 0 op/s rd, 3 op/s wr


root@Sam ~ # ceph health detail
HEALTH_ERR 3 scrub errors; Possible data damage: 3 pgs inconsistent
OSD_SCRUB_ERRORS 3 scrub errors
PG_DAMAGED Possible data damage: 3 pgs inconsistent
pg 5.1d is active+clean+inconsistent, acting [6,8,3]
pg 5.20 is active+clean+inconsistent, acting [3,9,0]
pg 5.4a is active+clean+inconsistent, acting [6,3,7]

What's the reason for this problem? How can I analyse it?

Cheers, Dimitri
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com