What you are seeing is expected behavior.  Pool numbers do not get reused; they 
increment up.  Pool names can be reused once they are deleted.  One note, 
though, if you delete and recreate the data pool, and want to use cephfs, 
you'll need to run 'ceph mds newfs <metadata pool #> <data pool #> 
--yes-i-really-mean-it' before mounting it.

Brad

-----Original Message-----
From: [email protected] 
[mailto:[email protected]] On Behalf Of [email protected]
Sent: Tuesday, March 18, 2014 11:53 AM
To: [email protected]
Subject: [ceph-users] Pool Count incrementing on each create even though I 
removed the pool each time


I am a novice ceph user creating a simple 4 OSD default cluster (initially) and 
experimenting with RADOS BENCH to understand basic HDD (OSD) performance. Each 
interation of rados bench -p data I want the cluster OSDs in initial state  
i.e. 0 objects . I assumed the easiest way was to remove and re-create the data 
pool each time.

While this appears to work , when I run ceph -s it shows me the pool count is 
incrementing each time:

matt@redstar9:~$ sudo ceph -s
    cluster c677f4c3-46a5-4ae1-b8aa-b070326c3b24
     health HEALTH_WARN clock skew detected on mon.redstar10, mon.redstar11
     monmap e1: 3 mons at
{redstar10=192.168.5.40:6789/0,redstar11=192.168.5.41:6789/0,redstar9=192.168.5.39:6789/0},
 election epoch 6, quorum 0,1,2 redstar10,redstar11,redstar9
     osdmap e52: 4 osds: 4 up, 4 in
      pgmap v5240: 136 pgs, 14 pools, 768 MB data, 194 objects
            1697 MB used, 14875 GB / 14876 GB avail
                 136 active+clean


even though lspools still only shows me the 3 default pools (metadata, rbd,
data)

Is this a bug, AND/OR, is there a better way to zero my cluster for these 
experiments?

Thanks,

Matt Latter

_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to