Hmm that does seem odd. How are you looking at those sizes?

On Wed, Dec 12, 2018 at 4:38 AM Sergey Dolgov <palz...@gmail.com> wrote:

> Greq, for example for our cluster ~1000 osd:
>
> size osdmap.1357881__0_F7FE779D__none = 363KB (crush_version 9860,
> modified 2018-12-12 04:00:17.661731)
> size osdmap.1357882__0_F7FE772D__none = 363KB
> size osdmap.1357883__0_F7FE74FD__none = 363KB (crush_version 9861,
> modified 2018-12-12 04:00:27.385702)
> size inc_osdmap.1357882__0_B783A4EA__none = 1.2MB
>
> difference between epoch 1357881 and 1357883: crush weight one osd was
> increased by 0.01 so we get 5 new pg_temp in osdmap.1357883 but size
> inc_osdmap so huge
>
> чт, 6 дек. 2018 г. в 06:20, Gregory Farnum <gfar...@redhat.com>:
> >
> > On Wed, Dec 5, 2018 at 3:32 PM Sergey Dolgov <palz...@gmail.com> wrote:
> >>
> >> Hi guys
> >>
> >> I faced strange behavior of crushmap change. When I change crush
> >> weight osd I sometimes get  increment osdmap(1.2MB) which size is
> >> significantly bigger than size of osdmap(0.4MB)
> >
> >
> > This is probably because when CRUSH changes, the new primary OSDs for a
> PG will tend to set a "pg temp" value (in the OSDMap) that temporarily
> reassigns it to the old acting set, so the data can be accessed while the
> new OSDs get backfilled. Depending on the size of your cluster, the number
> of PGs on it, and the size of the CRUSH change, this can easily be larger
> than the rest of the map because it is data with size linear in the number
> of PGs affected, instead of being more normally proportional to the number
> of OSDs.
> > -Greg
> >
> >>
> >> I use luminois 12.2.8. Cluster was installed a long ago, I suppose
> >> that initially it was firefly
> >> How can I view content of increment osdmap or can you give me opinion
> >> on this problem. I think that spikes of traffic tight after change of
> >> crushmap relates to this crushmap behavior
> >> _______________________________________________
> >> ceph-users mailing list
> >> ceph-users@lists.ceph.com
> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
> --
> Best regards, Sergey Dolgov
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to