I am a novice ceph user creating a simple 4 OSD default cluster (initially)
and experimenting with RADOS BENCH to understand basic HDD (OSD)
performance. Each interation of rados bench -p data I want the cluster OSDs
in initial state i.e. 0 objects . I assumed the easiest way was to remove
and re-create the data pool each time.
While this appears to work , when I run ceph -s it shows me the pool count
is incrementing each time:
matt@redstar9:~$ sudo ceph -s
cluster c677f4c3-46a5-4ae1-b8aa-b070326c3b24
health HEALTH_WARN clock skew detected on mon.redstar10, mon.redstar11
monmap e1: 3 mons at
{redstar10=192.168.5.40:6789/0,redstar11=192.168.5.41:6789/0,redstar9=192.168.5.39:6789/0},
election epoch 6, quorum 0,1,2 redstar10,redstar11,redstar9
osdmap e52: 4 osds: 4 up, 4 in
pgmap v5240: 136 pgs, 14 pools, 768 MB data, 194 objects
1697 MB used, 14875 GB / 14876 GB avail
136 active+clean
even though lspools still only shows me the 3 default pools (metadata, rbd,
data)
Is this a bug, AND/OR, is there a better way to zero my cluster for these
experiments?
Thanks,
Matt Latter
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com