> Op 29 september 2016 om 1:57 schreef Tyler Bishop 
> <[email protected]>:
> 
> 
> S1148 is down but the cluster does not mark it as such. 
> 

A host will never be marked as down, but the output shows that all OSDs are 
marked as down however.

Wido

> cluster 3aac8ab8-1011-43d6-b281-d16e7a61b2bd 
> health HEALTH_WARN 
> 3888 pgs backfill 
> 196 pgs backfilling 
> 6418 pgs degraded 
> 52 pgs down 
> 52 pgs peering 
> 1 pgs recovery_wait 
> 3653 pgs stuck degraded 
> 52 pgs stuck inactive 
> 6088 pgs stuck unclean 
> 3653 pgs stuck undersized 
> 6417 pgs undersized 
> 186 requests are blocked > 32 sec 
> recovery 42096983/185765821 objects degraded (22.661%) 
> recovery 49940341/185765821 objects misplaced (26.883%) 
> 16/330 in osds are down 
> monmap e1: 3 mons at 
> {ceph0-mon0=10.1.8.40:6789/0,ceph0-mon1=10.1.8.41:6789/0,ceph0-mon2=10.1.8.42:6789/0}
>  
> election epoch 13550, quorum 0,1,2 ceph0-mon0,ceph0-mon1,ceph0-mon2 
> osdmap e236889: 370 osds: 314 up, 330 in; 4096 remapped pgs 
> pgmap v47890297: 20920 pgs, 19 pools, 316 TB data, 85208 kobjects 
> 530 TB used, 594 TB / 1125 TB avail 
> 42096983/185765821 objects degraded (22.661%) 
> 49940341/185765821 objects misplaced (26.883%) 
> 14390 active+clean 
> 3846 active+undersized+degraded+remapped+wait_backfill 
> 2375 active+undersized+degraded 
> 196 active+undersized+degraded+remapped+backfilling 
> 52 down+peering 
> 42 active+remapped+wait_backfill 
> 11 active+remapped 
> 7 active+clean+scrubbing+deep 
> 1 active+recovery_wait+degraded+remapped 
> recovery io 2408 MB/s, 623 objects/s 
> 
> 
> -43 304.63928 host ceph0-s1148 
> 303 5.43999 osd.303 down 0 1.00000 
> 304 5.43999 osd.304 down 0 1.00000 
> 305 5.43999 osd.305 down 0 1.00000 
> 306 5.43999 osd.306 down 0 1.00000 
> 307 5.43999 osd.307 down 0 1.00000 
> 308 5.43999 osd.308 down 0 1.00000 
> 309 5.43999 osd.309 down 0 1.00000 
> 310 5.43999 osd.310 down 0 1.00000 
> 311 5.43999 osd.311 down 0 1.00000 
> 312 5.43999 osd.312 down 0 1.00000 
> 313 5.43999 osd.313 down 0 1.00000 
> 314 5.43999 osd.314 down 0 1.00000 
> 315 5.43999 osd.315 down 0 1.00000 
> 316 5.43999 osd.316 down 0 1.00000 
> 317 5.43999 osd.317 down 0 1.00000 
> 318 5.43999 osd.318 down 0 1.00000 
> 319 5.43999 osd.319 down 0 1.00000 
> 320 5.43999 osd.320 down 0 1.00000 
> 321 5.43999 osd.321 down 0 1.00000 
> 322 5.43999 osd.322 down 0 1.00000 
> 323 5.43999 osd.323 down 0 1.00000 
> 324 5.43999 osd.324 down 0 1.00000 
> 325 5.43999 osd.325 down 0 1.00000 
> 326 5.43999 osd.326 down 0 1.00000 
> 327 5.43999 osd.327 down 0 1.00000 
> 328 5.43999 osd.328 down 0 1.00000 
> 329 5.43999 osd.329 down 0 1.00000 
> 330 5.43999 osd.330 down 0 1.00000 
> 331 5.43999 osd.331 down 0 1.00000 
> 332 5.43999 osd.332 down 1.00000 1.00000 
> 333 5.43999 osd.333 down 1.00000 1.00000 
> 334 5.43999 osd.334 down 1.00000 1.00000 
> 335 5.43999 osd.335 down 0 1.00000 
> 337 5.43999 osd.337 down 1.00000 1.00000 
> 338 5.43999 osd.338 down 0 1.00000 
> 339 5.43999 osd.339 down 1.00000 1.00000 
> 340 5.43999 osd.340 down 0 1.00000 
> 341 5.43999 osd.341 down 0 1.00000 
> 342 5.43999 osd.342 down 0 1.00000 
> 343 5.43999 osd.343 down 0 1.00000 
> 344 5.43999 osd.344 down 0 1.00000 
> 345 5.43999 osd.345 down 0 1.00000 
> 346 5.43999 osd.346 down 0 1.00000 
> 347 5.43999 osd.347 down 1.00000 1.00000 
> 348 5.43999 osd.348 down 1.00000 1.00000 
> 349 5.43999 osd.349 down 0 1.00000 
> 350 5.43999 osd.350 down 1.00000 1.00000 
> 351 5.43999 osd.351 down 1.00000 1.00000 
> 352 5.43999 osd.352 down 1.00000 1.00000 
> 353 5.43999 osd.353 down 1.00000 1.00000 
> 354 5.43999 osd.354 down 1.00000 1.00000 
> 355 5.43999 osd.355 down 1.00000 1.00000 
> 356 5.43999 osd.356 down 1.00000 1.00000 
> 357 5.43999 osd.357 down 1.00000 1.00000 
> 358 5.43999 osd.358 down 0 1.00000 
> 369 5.43999 osd.369 down 1.00000 1.00000 
> 
> 
> 
> 
> 
>       
> 
> Tyler Bishop 
> Chief Technical Officer 
> 513-299-7108 x10 
> 
> 
> 
> [email protected] 
> 
> 
> If you are not the intended recipient of this transmission you are notified 
> that disclosing, copying, distributing or taking any action in reliance on 
> the contents of this information is strictly prohibited. 
> 
> 
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to