Hello Greg,

Output of 'ceph osd tree':

# id    weight  type name       up/down reweight
-1      27.3    root default
-2      9.1             host stor1
0       3.64                    osd.0   up      1
1       3.64                    osd.1   up      1
2       1.82                    osd.2   up      1
-3      9.1             host stor2
3       3.64                    osd.3   up      1
4       1.82                    osd.4   up      1
6       3.64                    osd.6   up      1
-4      9.1             host stor3
7       3.64                    osd.7   up      1
8       3.64                    osd.8   up      1
9       1.82                    osd.9   up      1

(missing of osd.5 comes from previous test when I remove HDD from a working
cluster, but I think this is not relevant now)

root@stor3:~# ceph osd pool get .rgw.buckets pg_num
pg_num: 250
root@stor3:~# ceph osd pool get .rgw.buckets pgp_num
pgp_num: 250

pgmap v129814: 514 pgs: 514 active; 818 GB data, 1682 GB used

Thank you,
Mihaly

2013/9/16 Gregory Farnum <[email protected]>

> What is your PG count and what's the output of "ceph osd tree"? It's
> possible that you've just got a slightly off distribution since there
> still isn't much data in the cluster (probabilistic placement and all
> that), but let's cover the basics first.
> -Greg
> Software Engineer #42 @ http://inktank.com | http://ceph.com
>
>
> On Mon, Sep 16, 2013 at 2:08 AM, Mihály Árva-Tóth
> <[email protected]> wrote:
> > Hello,
> >
> > I made some tests on 3 node Ceph cluster: upload 3 million 50 KiB object
> to
> > single container. Speed and performance were okay. But data does not
> > distributed correctly. Every node has got 2 pcs. 4 TB and 1 pc. 2 TB HDD.
> >
> > osd.0 41 GB (4 TB)
> > osd.1 47 GB (4 TB)
> > osd.3 16 GB (2 TB)
> > osd.4 40 GB (4 TB)
> > osd.5 49 GB (4 TB)
> > osd.6 17 GB (2 TB)
> > osd.7 48 GB (4 TB)
> > osd.8 42 GB (4 TB)
> > osd.9 18 GB (2 TB)
> >
> > Every 4 TB and 2 TB HDDs are from same vendor and same type. (WD RE SATA)
> >
> > I monitored iops with Zabbix under test, you can see here:
> > http://ctrlv.in/237368
> > (sda and sdb are system HDDs) This graph are same on every three nodes.
> >
> > Is there any idea what's wrong or what should I see?
> >
> > I'm using ceph-0.67.3 on Ubuntu 12.04.3 x86_64.
> >
> > Thank you,
> > Mihaly
> >
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to