On Wed, Feb 1, 2017 at 3:38 AM, Gregory Farnum <gfar...@redhat.com> wrote:
> On Tue, Jan 31, 2017 at 9:06 AM, Muthusamy Muthiah
> <muthiah.muthus...@gmail.com> wrote:
>> Hi Greg,
>>
>> the problem is in kraken,  when a pool is created with EC profile , min_size
>> equals erasure size.
>>
>> For 3+1 profile , following is the pool status ,
>> pool 2 'cdvr_ec' erasure size 4 min_size 4 crush_ruleset 1 object_hash
>> rjenkins pg_num 1024 pgp_num 1024 last_change 234 flags hashpspool
>> stripe_width 4128
>>
>> For 4+1 profile:
>> pool 5 'cdvr_ec' erasure size 5 min_size 5 crush_ruleset 1 object_hash
>> rjenkins pg_num 4096 pgp_num 4096
>>
>> For 3+2 profile :
>> pool 3 'cdvr_ec' erasure size 5 min_size 4 crush_ruleset 1 object_hash
>> rjenkins pg_num 1024 pgp_num 1024 last_change 412 flags hashpspool
>> stripe_width 4128
>>
>> Where as on Jewel release for EC 4+1:
>> pool 30 'cdvr_ec' erasure size 5 min_size 4 crush_ruleset 1 object_hash
>> rjenkins pg_num 4096 pgp_num 4096
>>
>> Trying to modify min_size and verify the status.
>>
>> Is there any reason behind this change in ceph kraken  or a bug.
>
> The change was made on purpose because running with k replicas on a
> k+m pool is a bad idea. However, it definitely should have recovered
> the missing shard and then gone active, which doesn't appear to have

Yeah, that might be true.

> happened in this case.
>
> It looks like we just screwed up and don't let EC pools do recovery on
> min size. You can restore the old behavior by setting min_size equal
> to k and we'll be fixing this for the next release. (In general, k+1

If it's true, we should stop users to set up this ratio of data /
coding chunks. Or there should be any reasonable warning.

Should it be feature request?

> pools are not a good idea, which is why we didn't catch this in
> testing.)
> -Greg
>
>>
>> Thanks,
>> Muthu
>>
>>
>>
>>
>> On 31 January 2017 at 18:17, Muthusamy Muthiah <muthiah.muthus...@gmail.com>
>> wrote:
>>>
>>> Hi Greg,
>>>
>>> Following are the test outcomes on EC profile ( n = k + m)
>>>
>>>
>>>
>>> 1.       Kraken filestore and bluetore with m=1 , recovery does not start
>>> .
>>>
>>> 2.       Jewel filestore and bluestore with m=1 , recovery happens .
>>>
>>> 3.       Kraken bluestore all default configuration and m=1, no recovery.
>>>
>>> 4.       Kraken bluestore with m=2 , recovery happens when one OSD is down
>>> and for 2 OSD fails.
>>>
>>>
>>>
>>> So, the issue seems to be on ceph-kraken release. Your views…
>>>
>>>
>>>
>>> Thanks,
>>>
>>> Muthu
>>>
>>>
>>>
>>>
>>> On 31 January 2017 at 14:18, Muthusamy Muthiah
>>> <muthiah.muthus...@gmail.com> wrote:
>>>>
>>>> Hi Greg,
>>>>
>>>> Now we could see the same problem exists for kraken-filestore also.
>>>> Attached the requested osdmap and crushmap.
>>>>
>>>> OSD.1 was stopped in this following procedure and OSD map for a PG is
>>>> displayed.
>>>>
>>>> ceph osd dump | grep cdvr_ec
>>>> 2017-01-31 08:39:44.827079 7f323d66c700 -1 WARNING: the following
>>>> dangerous and experimental features are enabled: bluestore,rocksdb
>>>> 2017-01-31 08:39:44.848901 7f323d66c700 -1 WARNING: the following
>>>> dangerous and experimental features are enabled: bluestore,rocksdb
>>>> pool 2 'cdvr_ec' erasure size 4 min_size 4 crush_ruleset 1 object_hash
>>>> rjenkins pg_num 1024 pgp_num 1024 last_change 234 flags hashpspool
>>>> stripe_width 4128
>>>>
>>>> [root@ca-cn2 ~]# ceph osd getmap -o /tmp/osdmap
>>>>
>>>>
>>>> [root@ca-cn2 ~]# osdmaptool --pool 2 --test-map-object object1
>>>> /tmp/osdmap
>>>> osdmaptool: osdmap file '/tmp/osdmap'
>>>>  object 'object1' -> 2.2bc -> [20,47,1,36]
>>>>
>>>> [root@ca-cn2 ~]# ceph osd map cdvr_ec object1
>>>> osdmap e402 pool 'cdvr_ec' (2) object 'object1' -> pg 2.bac5debc (2.2bc)
>>>> -> up ([20,47,1,36], p20) acting ([20,47,1,36], p20)
>>>>
>>>> [root@ca-cn2 ~]# systemctl stop ceph-osd@1.service
>>>>
>>>> [root@ca-cn2 ~]# ceph osd getmap -o /tmp/osdmap1
>>>>
>>>>
>>>> [root@ca-cn2 ~]# osdmaptool --pool 2 --test-map-object object1
>>>> /tmp/osdmap1
>>>> osdmaptool: osdmap file '/tmp/osdmap1'
>>>>  object 'object1' -> 2.2bc -> [20,47,2147483647,36]
>>>>
>>>>
>>>> [root@ca-cn2 ~]# ceph osd map cdvr_ec object1
>>>> osdmap e406 pool 'cdvr_ec' (2) object 'object1' -> pg 2.bac5debc (2.2bc)
>>>> -> up ([20,47,39,36], p20) acting ([20,47,NONE,36], p20)
>>>>
>>>>
>>>> [root@ca-cn2 ~]# ceph osd tree
>>>> 2017-01-31 08:42:19.606876 7f4ed856a700 -1 WARNING: the following
>>>> dangerous and experimental features are enabled: bluestore,rocksdb
>>>> 2017-01-31 08:42:19.628358 7f4ed856a700 -1 WARNING: the following
>>>> dangerous and experimental features are enabled: bluestore,rocksdb
>>>> ID WEIGHT    TYPE NAME       UP/DOWN REWEIGHT PRIMARY-AFFINITY
>>>> -1 327.47314 root default
>>>> -2  65.49463     host ca-cn4
>>>>  3   5.45789         osd.3        up  1.00000          1.00000
>>>>  5   5.45789         osd.5        up  1.00000          1.00000
>>>> 10   5.45789         osd.10       up  1.00000          1.00000
>>>> 16   5.45789         osd.16       up  1.00000          1.00000
>>>> 21   5.45789         osd.21       up  1.00000          1.00000
>>>> 27   5.45789         osd.27       up  1.00000          1.00000
>>>> 30   5.45789         osd.30       up  1.00000          1.00000
>>>> 35   5.45789         osd.35       up  1.00000          1.00000
>>>> 42   5.45789         osd.42       up  1.00000          1.00000
>>>> 47   5.45789         osd.47       up  1.00000          1.00000
>>>> 51   5.45789         osd.51       up  1.00000          1.00000
>>>> 53   5.45789         osd.53       up  1.00000          1.00000
>>>> -3  65.49463     host ca-cn3
>>>>  2   5.45789         osd.2        up  1.00000          1.00000
>>>>  6   5.45789         osd.6        up  1.00000          1.00000
>>>> 11   5.45789         osd.11       up  1.00000          1.00000
>>>> 15   5.45789         osd.15       up  1.00000          1.00000
>>>> 20   5.45789         osd.20       up  1.00000          1.00000
>>>> 25   5.45789         osd.25       up  1.00000          1.00000
>>>> 29   5.45789         osd.29       up  1.00000          1.00000
>>>> 33   5.45789         osd.33       up  1.00000          1.00000
>>>> 38   5.45789         osd.38       up  1.00000          1.00000
>>>> 40   5.45789         osd.40       up  1.00000          1.00000
>>>> 45   5.45789         osd.45       up  1.00000          1.00000
>>>> 49   5.45789         osd.49       up  1.00000          1.00000
>>>> -4  65.49463     host ca-cn5
>>>>  0   5.45789         osd.0        up  1.00000          1.00000
>>>>  7   5.45789         osd.7        up  1.00000          1.00000
>>>> 12   5.45789         osd.12       up  1.00000          1.00000
>>>> 17   5.45789         osd.17       up  1.00000          1.00000
>>>> 23   5.45789         osd.23       up  1.00000          1.00000
>>>> 26   5.45789         osd.26       up  1.00000          1.00000
>>>> 32   5.45789         osd.32       up  1.00000          1.00000
>>>> 34   5.45789         osd.34       up  1.00000          1.00000
>>>> 41   5.45789         osd.41       up  1.00000          1.00000
>>>> 46   5.45789         osd.46       up  1.00000          1.00000
>>>> 52   5.45789         osd.52       up  1.00000          1.00000
>>>> 56   5.45789         osd.56       up  1.00000          1.00000
>>>> -5  65.49463     host ca-cn1
>>>>  4   5.45789         osd.4        up  1.00000          1.00000
>>>>  9   5.45789         osd.9        up  1.00000          1.00000
>>>> 14   5.45789         osd.14       up  1.00000          1.00000
>>>> 19   5.45789         osd.19       up  1.00000          1.00000
>>>> 24   5.45789         osd.24       up  1.00000          1.00000
>>>> 36   5.45789         osd.36       up  1.00000          1.00000
>>>> 43   5.45789         osd.43       up  1.00000          1.00000
>>>> 50   5.45789         osd.50       up  1.00000          1.00000
>>>> 55   5.45789         osd.55       up  1.00000          1.00000
>>>> 57   5.45789         osd.57       up  1.00000          1.00000
>>>> 58   5.45789         osd.58       up  1.00000          1.00000
>>>> 59   5.45789         osd.59       up  1.00000          1.00000
>>>> -6  65.49463     host ca-cn2
>>>>  1   5.45789         osd.1      down        0          1.00000
>>>>  8   5.45789         osd.8        up  1.00000          1.00000
>>>> 13   5.45789         osd.13       up  1.00000          1.00000
>>>> 18   5.45789         osd.18       up  1.00000          1.00000
>>>> 22   5.45789         osd.22       up  1.00000          1.00000
>>>> 28   5.45789         osd.28       up  1.00000          1.00000
>>>> 31   5.45789         osd.31       up  1.00000          1.00000
>>>> 37   5.45789         osd.37       up  1.00000          1.00000
>>>> 39   5.45789         osd.39       up  1.00000          1.00000
>>>> 44   5.45789         osd.44       up  1.00000          1.00000
>>>> 48   5.45789         osd.48       up  1.00000          1.00000
>>>> 54   5.45789         osd.54       up  1.00000          1.00000
>>>>
>>>> health HEALTH_ERR
>>>>             69 pgs are stuck inactive for more than 300 seconds
>>>>             69 pgs incomplete
>>>>             69 pgs stuck inactive
>>>>             69 pgs stuck unclean
>>>>             512 requests are blocked > 32 sec
>>>>      monmap e2: 5 mons at
>>>> {ca-cn1=10.50.5.117:6789/0,ca-cn2=10.50.5.118:6789/0,ca-cn3=10.50.5.119:6789/0,ca-cn4=10.50.5.120:6789/0,ca-cn5=10.50.5.121:6789/0}
>>>>             election epoch 8, quorum 0,1,2,3,4
>>>> ca-cn1,ca-cn2,ca-cn3,ca-cn4,ca-cn5
>>>>         mgr active: ca-cn4 standbys: ca-cn2, ca-cn5, ca-cn3, ca-cn1
>>>>      osdmap e406: 60 osds: 59 up, 59 in; 69 remapped pgs
>>>>             flags sortbitwise,require_jewel_osds,require_kraken_osds
>>>>       pgmap v23018: 1024 pgs, 1 pools, 3892 GB data, 7910 kobjects
>>>>             6074 GB used, 316 TB / 322 TB avail
>>>>                  955 active+clean
>>>>                   69 remapped+incomplete
>>>>
>>>> Thanks,
>>>> Muthu
>>>>
>>>>
>>>> On 31 January 2017 at 02:54, Gregory Farnum <gfar...@redhat.com> wrote:
>>>>>
>>>>> You might also check out "ceph osd tree" and crush dump and make sure
>>>>> they look the way you expect.
>>>>>
>>>>> On Mon, Jan 30, 2017 at 1:23 PM, Gregory Farnum <gfar...@redhat.com>
>>>>> wrote:
>>>>> > On Sun, Jan 29, 2017 at 6:40 AM, Muthusamy Muthiah
>>>>> > <muthiah.muthus...@gmail.com> wrote:
>>>>> >> Hi All,
>>>>> >>
>>>>> >> Also tried EC profile 3+1 on 5 node cluster with bluestore enabled  .
>>>>> >> When
>>>>> >> an OSD is down the cluster goes to ERROR state even when the cluster
>>>>> >> is n+1
>>>>> >> . No recovery happening.
>>>>> >>
>>>>> >> health HEALTH_ERR
>>>>> >>             75 pgs are stuck inactive for more than 300 seconds
>>>>> >>             75 pgs incomplete
>>>>> >>             75 pgs stuck inactive
>>>>> >>             75 pgs stuck unclean
>>>>> >>      monmap e2: 5 mons at
>>>>> >>
>>>>> >> {ca-cn1=10.50.5.117:6789/0,ca-cn2=10.50.5.118:6789/0,ca-cn3=10.50.5.119:6789/0,ca-cn4=10.50.5.120:6789/0,ca-cn5=10.50.5.121:6789/0}
>>>>> >>             election epoch 10, quorum 0,1,2,3,4
>>>>> >> ca-cn1,ca-cn2,ca-cn3,ca-cn4,ca-cn5
>>>>> >>         mgr active: ca-cn1 standbys: ca-cn4, ca-cn3, ca-cn5, ca-cn2
>>>>> >>      osdmap e264: 60 osds: 59 up, 59 in; 75 remapped pgs
>>>>> >>             flags sortbitwise,require_jewel_osds,require_kraken_osds
>>>>> >>       pgmap v119402: 1024 pgs, 1 pools, 28519 GB data, 21548 kobjects
>>>>> >>             39976 GB used, 282 TB / 322 TB avail
>>>>> >>                  941 active+clean
>>>>> >>                   75 remapped+incomplete
>>>>> >>                    8 active+clean+scrubbing
>>>>> >>
>>>>> >> this seems to be an issue with bluestore , recovery not happening
>>>>> >> properly
>>>>> >> with EC .
>>>>> >
>>>>> > It's possible but it seems a lot more likely this is some kind of
>>>>> > config issue. Can you share your osd map ("ceph osd getmap")?
>>>>> > -Greg
>>>>> >
>>>>> >>
>>>>> >> Thanks,
>>>>> >> Muthu
>>>>> >>
>>>>> >> On 24 January 2017 at 12:57, Muthusamy Muthiah
>>>>> >> <muthiah.muthus...@gmail.com>
>>>>> >> wrote:
>>>>> >>>
>>>>> >>> Hi Greg,
>>>>> >>>
>>>>> >>> We use EC:4+1 on 5 node cluster in production deployments with
>>>>> >>> filestore
>>>>> >>> and it does recovery and peering when one OSD goes down. After few
>>>>> >>> mins ,
>>>>> >>> other OSD from a node where the fault OSD exists will take over the
>>>>> >>> PGs
>>>>> >>> temporarily and all PGs goes to active + clean state . Cluster also
>>>>> >>> does not
>>>>> >>> goes down during this recovery process.
>>>>> >>>
>>>>> >>> Only on bluestore we see cluster going to error state when one OSD
>>>>> >>> is
>>>>> >>> down.
>>>>> >>> We are still validating this and let you know additional findings.
>>>>> >>>
>>>>> >>> Thanks,
>>>>> >>> Muthu
>>>>> >>>
>>>>> >>> On 21 January 2017 at 02:06, Shinobu Kinjo <ski...@redhat.com>
>>>>> >>> wrote:
>>>>> >>>>
>>>>> >>>> `ceph pg dump` should show you something like:
>>>>> >>>>
>>>>> >>>>  * active+undersized+degraded ... [NONE,3,2,4,1]    3
>>>>> >>>> [NONE,3,2,4,1]
>>>>> >>>>
>>>>> >>>> Sam,
>>>>> >>>>
>>>>> >>>> Am I wrong? Or is it up to something else?
>>>>> >>>>
>>>>> >>>>
>>>>> >>>> On Sat, Jan 21, 2017 at 4:22 AM, Gregory Farnum
>>>>> >>>> <gfar...@redhat.com>
>>>>> >>>> wrote:
>>>>> >>>> > I'm pretty sure the default configs won't let an EC PG go active
>>>>> >>>> > with
>>>>> >>>> > only "k" OSDs in its PG; it needs at least k+1 (or possibly more?
>>>>> >>>> > Not
>>>>> >>>> > certain). Running an "n+1" EC config is just not a good idea.
>>>>> >>>> > For testing you could probably adjust this with the equivalent of
>>>>> >>>> > min_size for EC pools, but I don't know the parameters off the
>>>>> >>>> > top of
>>>>> >>>> > my head.
>>>>> >>>> > -Greg
>>>>> >>>> >
>>>>> >>>> > On Fri, Jan 20, 2017 at 2:15 AM, Muthusamy Muthiah
>>>>> >>>> > <muthiah.muthus...@gmail.com> wrote:
>>>>> >>>> >> Hi ,
>>>>> >>>> >>
>>>>> >>>> >> We are validating kraken 11.2.0 with bluestore  on 5 node
>>>>> >>>> >> cluster with
>>>>> >>>> >> EC
>>>>> >>>> >> 4+1.
>>>>> >>>> >>
>>>>> >>>> >> When an OSD is down , the peering is not happening and ceph
>>>>> >>>> >> health
>>>>> >>>> >> status
>>>>> >>>> >> moved to ERR state after few mins. This was working in previous
>>>>> >>>> >> development
>>>>> >>>> >> releases. Any additional configuration required in v11.2.0
>>>>> >>>> >>
>>>>> >>>> >> Following is our ceph configuration:
>>>>> >>>> >>
>>>>> >>>> >> mon_osd_down_out_interval = 30
>>>>> >>>> >> mon_osd_report_timeout = 30
>>>>> >>>> >> mon_osd_down_out_subtree_limit = host
>>>>> >>>> >> mon_osd_reporter_subtree_level = host
>>>>> >>>> >>
>>>>> >>>> >> and the recovery parameters set to default.
>>>>> >>>> >>
>>>>> >>>> >> [root@ca-cn1 ceph]# ceph osd crush show-tunables
>>>>> >>>> >>
>>>>> >>>> >> {
>>>>> >>>> >>     "choose_local_tries": 0,
>>>>> >>>> >>     "choose_local_fallback_tries": 0,
>>>>> >>>> >>     "choose_total_tries": 50,
>>>>> >>>> >>     "chooseleaf_descend_once": 1,
>>>>> >>>> >>     "chooseleaf_vary_r": 1,
>>>>> >>>> >>     "chooseleaf_stable": 1,
>>>>> >>>> >>     "straw_calc_version": 1,
>>>>> >>>> >>     "allowed_bucket_algs": 54,
>>>>> >>>> >>     "profile": "jewel",
>>>>> >>>> >>     "optimal_tunables": 1,
>>>>> >>>> >>     "legacy_tunables": 0,
>>>>> >>>> >>     "minimum_required_version": "jewel",
>>>>> >>>> >>     "require_feature_tunables": 1,
>>>>> >>>> >>     "require_feature_tunables2": 1,
>>>>> >>>> >>     "has_v2_rules": 1,
>>>>> >>>> >>     "require_feature_tunables3": 1,
>>>>> >>>> >>     "has_v3_rules": 0,
>>>>> >>>> >>     "has_v4_buckets": 0,
>>>>> >>>> >>     "require_feature_tunables5": 1,
>>>>> >>>> >>     "has_v5_rules": 0
>>>>> >>>> >> }
>>>>> >>>> >>
>>>>> >>>> >> ceph status:
>>>>> >>>> >>
>>>>> >>>> >>      health HEALTH_ERR
>>>>> >>>> >>             173 pgs are stuck inactive for more than 300 seconds
>>>>> >>>> >>             173 pgs incomplete
>>>>> >>>> >>             173 pgs stuck inactive
>>>>> >>>> >>             173 pgs stuck unclean
>>>>> >>>> >>      monmap e2: 5 mons at
>>>>> >>>> >>
>>>>> >>>> >>
>>>>> >>>> >> {ca-cn1=10.50.5.117:6789/0,ca-cn2=10.50.5.118:6789/0,ca-cn3=10.50.5.119:6789/0,ca-cn4=10.50.5.120:6789/0,ca-cn5=10.50.5.121:6789/0}
>>>>> >>>> >>             election epoch 106, quorum 0,1,2,3,4
>>>>> >>>> >> ca-cn1,ca-cn2,ca-cn3,ca-cn4,ca-cn5
>>>>> >>>> >>         mgr active: ca-cn1 standbys: ca-cn2, ca-cn4, ca-cn5,
>>>>> >>>> >> ca-cn3
>>>>> >>>> >>      osdmap e1128: 60 osds: 59 up, 59 in; 173 remapped pgs
>>>>> >>>> >>             flags
>>>>> >>>> >> sortbitwise,require_jewel_osds,require_kraken_osds
>>>>> >>>> >>       pgmap v782747: 2048 pgs, 1 pools, 63133 GB data, 46293
>>>>> >>>> >> kobjects
>>>>> >>>> >>             85199 GB used, 238 TB / 322 TB avail
>>>>> >>>> >>                 1868 active+clean
>>>>> >>>> >>                  173 remapped+incomplete
>>>>> >>>> >>                    7 active+clean+scrubbing
>>>>> >>>> >>
>>>>> >>>> >> MON log:
>>>>> >>>> >>
>>>>> >>>> >> 2017-01-20 09:25:54.715684 7f55bcafb700  0 log_channel(cluster)
>>>>> >>>> >> log
>>>>> >>>> >> [INF] :
>>>>> >>>> >> osd.54 out (down for 31.703786)
>>>>> >>>> >> 2017-01-20 09:25:54.725688 7f55bf4d5700  0
>>>>> >>>> >> mon.ca-cn1@0(leader).osd
>>>>> >>>> >> e1120
>>>>> >>>> >> crush map has features 288250512065953792, adjusting msgr
>>>>> >>>> >> requires
>>>>> >>>> >> 2017-01-20 09:25:54.729019 7f55bf4d5700  0 log_channel(cluster)
>>>>> >>>> >> log
>>>>> >>>> >> [INF] :
>>>>> >>>> >> osdmap e1120: 60 osds: 59 up, 59 in
>>>>> >>>> >> 2017-01-20 09:25:54.735987 7f55bf4d5700  0 log_channel(cluster)
>>>>> >>>> >> log
>>>>> >>>> >> [INF] :
>>>>> >>>> >> pgmap v781993: 2048 pgs: 1869 active+clean, 173 incomplete, 6
>>>>> >>>> >> active+clean+scrubbing; 63159 GB data, 85201 GB used, 238 TB /
>>>>> >>>> >> 322 TB
>>>>> >>>> >> avail;
>>>>> >>>> >> 21825 B/s rd, 163 MB/s wr, 2046 op/s
>>>>> >>>> >> 2017-01-20 09:25:55.737749 7f55bf4d5700  0
>>>>> >>>> >> mon.ca-cn1@0(leader).osd
>>>>> >>>> >> e1121
>>>>> >>>> >> crush map has features 288250512065953792, adjusting msgr
>>>>> >>>> >> requires
>>>>> >>>> >> 2017-01-20 09:25:55.744338 7f55bf4d5700  0 log_channel(cluster)
>>>>> >>>> >> log
>>>>> >>>> >> [INF] :
>>>>> >>>> >> osdmap e1121: 60 osds: 59 up, 59 in
>>>>> >>>> >> 2017-01-20 09:25:55.749616 7f55bf4d5700  0 log_channel(cluster)
>>>>> >>>> >> log
>>>>> >>>> >> [INF] :
>>>>> >>>> >> pgmap v781994: 2048 pgs: 29 remapped+incomplete, 1869
>>>>> >>>> >> active+clean,
>>>>> >>>> >> 144
>>>>> >>>> >> incomplete, 6 active+clean+scrubbing; 63159 GB data, 85201 GB
>>>>> >>>> >> used,
>>>>> >>>> >> 238 TB /
>>>>> >>>> >> 322 TB avail; 44503 B/s rd, 45681 kB/s wr, 518 op/s
>>>>> >>>> >> 2017-01-20 09:25:56.768721 7f55bf4d5700  0 log_channel(cluster)
>>>>> >>>> >> log
>>>>> >>>> >> [INF] :
>>>>> >>>> >> pgmap v781995: 2048 pgs: 47 remapped+incomplete, 1869
>>>>> >>>> >> active+clean,
>>>>> >>>> >> 126
>>>>> >>>> >> incomplete, 6 active+clean+scrubbing; 63159 GB data, 85201 GB
>>>>> >>>> >> used,
>>>>> >>>> >> 238 TB /
>>>>> >>>> >> 322 TB avail; 20275 B/s rd, 72742 kB/s wr, 665 op/s
>>>>> >>>> >>
>>>>> >>>> >> Thanks,
>>>>> >>>> >> Muthu
>>>>> >>>> >>
>>>>> >>>> >>
>>>>> >>>> >> _______________________________________________
>>>>> >>>> >> ceph-users mailing list
>>>>> >>>> >> ceph-users@lists.ceph.com
>>>>> >>>> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>> >>>> >>
>>>>> >>>> > _______________________________________________
>>>>> >>>> > ceph-users mailing list
>>>>> >>>> > ceph-users@lists.ceph.com
>>>>> >>>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>> >>>
>>>>> >>>
>>>>> >>
>>>>> >>
>>>>> >> _______________________________________________
>>>>> >> ceph-users mailing list
>>>>> >> ceph-users@lists.ceph.com
>>>>> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>> >>
>>>>
>>>>
>>>
>>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to