try setting: 

mon_osd_min_down_reporters = 1 

On 2018-01-31 20:46, Steven Vacaroaia wrote:

> Hi, 
> 
> Why is ceph osd tree reports that osd.4 is up when the server on which osd.4 
> is running is actually down ?? 
> 
> Any help will be appreciated  
> 
> [root@osd01 ~]# ping -c 2 osd02 
> PING osd02 (10.10.30.182) 56(84) bytes of data. 
> From osd01 (10.10.30.181) icmp_seq=1 Destination Host Unreachable 
> From osd01 (10.10.30.181) icmp_seq=2 Destination Host Unreachable 
> 
> [root@osd01 ~]# ceph osd tree 
> ID  CLASS WEIGHT  TYPE NAME          STATUS REWEIGHT PRI-AFF 
> -9             0 root ssds 
> -10             0     host osd01-ssd 
> -11             0     host osd02-ssd 
> -12             0     host osd04-ssd 
> -1       4.22031 root default 
> -3       1.67967     host osd01 
> 0   hdd 0.55989         osd.0        down        0 1.00000 
> 3   hdd 0.55989         osd.3        down        0 1.00000 
> 6   hdd 0.55989         osd.6          up  1.00000 1.00000 
> -5       1.67967     host osd02 
> 1   hdd 0.55989         osd.1        down  1.00000 1.00000 
> 4   hdd 0.55989         osd.4          up  1.00000 1.00000 
> 7   hdd 0.55989         osd.7        down  1.00000 1.00000 
> -7       0.86096     host osd04 
> 2   hdd 0.28699         osd.2        down        0 1.00000 
> 5   hdd 0.28699         osd.5        down  1.00000 1.00000 
> 8   hdd 0.28699         osd.8        down  1.00000 1.00000 
> [root@osd01 ~]# ceph tell osd.4 bench 
> ^CError EINTR: problem getting command descriptions from osd.4 
> [root@osd01 ~]# ceph osd df 
> ID CLASS WEIGHT  REWEIGHT SIZE  USE    AVAIL %USE VAR  PGS 
> 0   hdd 0.55989        0     0      0     0    0    0   0 
> 3   hdd 0.55989        0     0      0     0    0    0   0 
> 6   hdd 0.55989  1.00000  573G 16474M  557G 2.81 0.84   0 
> 1   hdd 0.55989  1.00000  573G 16516M  557G 2.81 0.84   0 
> 4   hdd 0.55989  1.00000  573G 16465M  557G 2.80 0.84   0 
> 7   hdd 0.55989  1.00000  573G 16473M  557G 2.81 0.84   0 
> 2   hdd 0.28699        0     0      0     0    0    0   0 
> 5   hdd 0.28699  1.00000  293G 16466M  277G 5.47 1.63   0 
> 8   hdd 0.28699  1.00000  293G 16461M  277G 5.47 1.63   0 
> TOTAL 2881G 98857M 2784G 3.35 
> MIN/MAX VAR: 0.84/1.63  STDDEV: 1.30 
> [root@osd01 ~]# ceph osd df tree 
> ID  CLASS WEIGHT  REWEIGHT SIZE  USE    AVAIL %USE VAR  PGS TYPE NAME 
> -9             0        -     0      0     0    0    0   - root ssds 
> -10             0        -     0      0     0    0    0   -     host 
> osd01-ssd 
> -11             0        -     0      0     0    0    0   -     host 
> osd02-ssd 
> -12             0        -     0      0     0    0    0   -     host 
> osd04-ssd 
> -1       4.22031        - 2881G 98857M 2784G 3.35 1.00   - root default 
> -3       1.67967        -  573G 16474M  557G 2.81 0.84   -     host osd01 
> 0   hdd 0.55989        0     0      0     0    0    0   0         osd.0 
> 3   hdd 0.55989        0     0      0     0    0    0   0         osd.3 
> 6   hdd 0.55989  1.00000  573G 16474M  557G 2.81 0.84   0         osd.6 
> -5       1.67967        - 1720G 49454M 1671G 2.81 0.84   -     host osd02 
> 1   hdd 0.55989  1.00000  573G 16516M  557G 2.81 0.84   0         osd.1 
> 4   hdd 0.55989  1.00000  573G 16465M  557G 2.80 0.84   0         osd.4 
> 7   hdd 0.55989  1.00000  573G 16473M  557G 2.81 0.84   0         osd.7 
> -7       0.86096        -  587G 32928M  555G 5.47 1.63   -     host osd04 
> 2   hdd 0.28699        0     0      0     0    0    0   0         osd.2 
> 5   hdd 0.28699  1.00000  293G 16466M  277G 5.47 1.63   0         osd.5 
> 8   hdd 0.28699  1.00000  293G 16461M  277G 5.47 1.63   0         osd.8 
> TOTAL 2881G 98857M 2784G 3.35 
> MIN/MAX VAR: 0.84/1.63  STDDEV: 1.30 
> 
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to