hi ceph,
just for testing (on emperor 0.72.1) I created two OSD’s on a single server,
resized the pool to a replication factor of one, and created 200 PG’s for that
pool:
# ceph osd dump
...
pool 4 'rbd' rep size 1 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num
200 pgp_num 200 last_change 64 owner 18446744073709551615
crush_ruleset 0 is - IMHO - stating that data should be distributed on the OSD
level (step chooseleaf firstn 0 type osd):
# rules
rule data {
ruleset 0
type replicated
min_size 1
max_size 10
step take default
step chooseleaf firstn 0 type osd
step emit
}
Now I do an rbd import of an RBD Image (which is 1G in size), and I would
expect that RBD image to stripe across the two OSD’s. Well, this is just not
happening, everything sits on OSD2 (osd1 and osd0 have been removed in the mean
time, they have been part of the first test):
# df -h
...
/dev/vdc1 2.0G 905M 1.2G 45% /var/lib/ceph/osd/ceph-2
/dev/vdb1 2.0G 37M 2.0G 2% /var/lib/ceph/osd/ceph-3
# ceph -w
cluster 6db7c956-cfbb-437a-88b6-78e1c9e68c80
health HEALTH_OK
monmap e1: 1 mons at {ceph-node1=XXX:6789/0}, election epoch 1, quorum 0
ceph-node1
osdmap e65: 2 osds: 2 up, 2 in
pgmap v187: 200 pgs, 5 pools, 868 MB data, 220 objects
941 MB used, 3132 MB / 4073 MB avail
200 active+clean
2013-12-05 09:46:43.210312 mon.0 [INF] pgmap v187: 200 pgs: 200 active+clean;
868 MB data, 941 MB used, 3132 MB / 4073 MB avail
Any hints are more than welcome, this is for sure not a real life scenario, but
it really confuses my understanding of ceph.
Wolfgang
--
http://www.wogri.at
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com