There have been bugs here in the recent past which have been fixed for
hammer, at least...it's possible we didn't backport it for the giant
point release. :(

But for users going forward that procedure should be good!
-Greg

On Thu, Mar 26, 2015 at 11:26 AM, Kyle Hutson <kylehut...@ksu.edu> wrote:
> For what it's worth, I don't think  "being patient" was the answer. I was
> having the same problem a couple of weeks ago, and I waited from before 5pm
> one day until after 8am the next, and still got the same errors. I ended up
> adding a "new" cephfs pool with a newly-created small pool, but was never
> able to actually remove cephfs altogether.
>
> On Thu, Mar 26, 2015 at 12:45 PM, Jake Grimmett <j...@mrc-lmb.cam.ac.uk>
> wrote:
>>
>> On 03/25/2015 05:44 PM, Gregory Farnum wrote:
>>>
>>> On Wed, Mar 25, 2015 at 10:36 AM, Jake Grimmett <j...@mrc-lmb.cam.ac.uk>
>>> wrote:
>>>>
>>>> Dear All,
>>>>
>>>> Please forgive this post if it's naive, I'm trying to familiarise myself
>>>> with cephfs!
>>>>
>>>> I'm using Scientific Linux 6.6. with Ceph 0.87.1
>>>>
>>>> My first steps with cephfs using a replicated pool worked OK.
>>>>
>>>> Now trying now to test cephfs via a replicated caching tier on top of an
>>>> erasure pool. I've created an erasure pool, cannot put it under the
>>>> existing
>>>> replicated pool.
>>>>
>>>> My thoughts were to delete the existing cephfs, and start again, however
>>>> I
>>>> cannot delete the existing cephfs:
>>>>
>>>> errors are as follows:
>>>>
>>>> [root@ceph1 ~]# ceph fs rm cephfs2
>>>> Error EINVAL: all MDS daemons must be inactive before removing
>>>> filesystem
>>>>
>>>> I've tried killing the ceph-mds process, but this does not prevent the
>>>> above
>>>> error.
>>>>
>>>> I've also tried this, which also errors:
>>>>
>>>> [root@ceph1 ~]# ceph mds stop 0
>>>> Error EBUSY: must decrease max_mds or else MDS will immediately
>>>> reactivate
>>>
>>>
>>> Right, so did you run "ceph mds set_max_mds 0" and then repeating the
>>> stop command? :)
>>>
>>>>
>>>> This also fail...
>>>>
>>>> [root@ceph1 ~]# ceph-deploy mds destroy
>>>> [ceph_deploy.conf][DEBUG ] found configuration file at:
>>>> /root/.cephdeploy.conf
>>>> [ceph_deploy.cli][INFO  ] Invoked (1.5.21): /usr/bin/ceph-deploy mds
>>>> destroy
>>>> [ceph_deploy.mds][ERROR ] subcommand destroy not implemented
>>>>
>>>> Am I doing the right thing in trying to wipe the original cephfs config
>>>> before attempting to use an erasure cold tier? Or can I just redefine
>>>> the
>>>> cephfs?
>>>
>>>
>>> Yeah, unfortunately you need to recreate it if you want to try and use
>>> an EC pool with cache tiering, because CephFS knows what pools it
>>> expects data to belong to. Things are unlikely to behave correctly if
>>> you try and stick an EC pool under an existing one. :(
>>>
>>> Sounds like this is all just testing, which is good because the
>>> suitability of EC+cache is very dependent on how much hot data you
>>> have, etc...good luck!
>>> -Greg
>>>
>>>>
>>>> many thanks,
>>>>
>>>> Jake Grimmett
>>>> _______________________________________________
>>>> ceph-users mailing list
>>>> ceph-users@lists.ceph.com
>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>> Thanks for your help - much appreciated.
>>
>> The "set_max_mds 0" command worked, but only after I rebooted the server,
>> and restarted ceph twice. Before this I still got an
>> "mds active" error, and so was unable to destroy the cephfs.
>>
>> Possibly I was being impatient, and needed to let mds go inactive? there
>> were ~1 million files on the system.
>>
>> [root@ceph1 ~]# ceph mds set_max_mds 0
>> max_mds = 0
>>
>> [root@ceph1 ~]# ceph mds stop 0
>> telling mds.0 10.1.0.86:6811/3249 to deactivate
>>
>> [root@ceph1 ~]# ceph mds stop 0
>> Error EEXIST: mds.0 not active (up:stopping)
>>
>> [root@ceph1 ~]# ceph fs rm cephfs2
>> Error EINVAL: all MDS daemons must be inactive before removing filesystem
>>
>> There shouldn't be any other mds servers running..
>> [root@ceph1 ~]# ceph mds stop 1
>> Error EEXIST: mds.1 not active (down:dne)
>>
>> At this point I rebooted the server, did a "service ceph restart" twice.
>> Shutdown ceph, then restarted ceph before this command worked:
>>
>> [root@ceph1 ~]# ceph fs rm cephfs2 --yes-i-really-mean-it
>>
>> Anyhow, I've now been able to create an erasure coded pool, with a
>> replicated tier which cephfs is running on :)
>>
>> *Lots* of testing to go!
>>
>> Again, many thanks
>>
>> Jake
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to