Hey Eugen,

Pleasure. Glad you problem resolved itself.

Regards,
Tom

On Wed, Jan 24, 2018 at 5:29 PM, Eugen Block <[email protected]> wrote:

> Hi Tom,
>
> thanks for the detailed steps.
>
> I think our problem literally vanished. A couple of days after my email I
> noticed that the web interface suddenly  listed only one cephFS. Also the
> command "ceph fs status" doesn't return an error anymore but shows the
> corret output.
> I guess Ceph is indeed a self-healing storage solution! :-)
>
> Regards,
> Eugen
>
>
> Zitat von Thomas Bennett <[email protected]>:
>
> Hi Eugen,
>>
>> From my experiences, to truely delete and recreate  the Ceph FS *cephfs*
>>
>> file system I've done the following:
>>
>> 1. Remove the file system:
>> ceph fs rm cephfs --yes-i-really-mean-it
>> ceph fs rm_data_pool cephfs_data
>> ceph fs rm_data_pool cephfs cephfs_data
>>
>> 2. Remove the associated pools:
>> ceph osd pool delete cephfs_data cephfs_data --yes-i-really-really-mean-it
>> ceph osd pool delete cephfs_metadata
>> cephfs_metadata --yes-i-really-really-mean-it
>>
>> 3. Create a new default ceph file system:
>> ceph osd pool create cephfs_data <int[0-]> {<int[0-]>} {<ruleset>}
>> ceph osd pool create cephfs_metadata <int[0-]> {<int[0-]>} {<ruleset>}
>> ceph fs new cephfs cephfs_metadata cephfs_data
>> ceph fs set_default cephfs
>>
>> Not sure if this helps, as you may need to repeat the whole process from
>> the start.
>>
>> Regards,
>> Tom
>>
>> On Mon, Jan 8, 2018 at 2:19 PM, Eugen Block <[email protected]> wrote:
>>
>> Hi list,
>>>
>>> all this is on Ceph 12.2.2.
>>>
>>> An existing cephFS (named "cephfs") was backed up as a tar ball, then
>>> "removed" ("ceph fs rm cephfs --yes-i-really-mean-it"), a new one created
>>> ("ceph fs new cephfs cephfs-metadata cephfs-data") and the content
>>> restored
>>> from the tar ball. According to the output of "ceph fs rm",  the old
>>> cephFS
>>> has only been deactivated, not deleted.  Looking at the Ceph manager's
>>> web
>>> interface, it now lists two entries "cephfs", one with id 0 (the "old"
>>> FS)
>>> and id "1" (the currently active FS).
>>>
>>> When we try to run "ceph fs status", we get an error with a traceback:
>>>
>>> ---cut here---
>>> ceph3:~ # ceph fs status
>>> Error EINVAL: Traceback (most recent call last):
>>>   File "/usr/lib64/ceph/mgr/status/module.py", line 301, in
>>> handle_command
>>>     return self.handle_fs_status(cmd)
>>>   File "/usr/lib64/ceph/mgr/status/module.py", line 219, in
>>> handle_fs_status
>>>     stats = pool_stats[pool_id]
>>> KeyError: (29L,)
>>> ---cut here---
>>>
>>> while this works:
>>>
>>> ---cut here---
>>> ceph3:~ # ceph fs ls
>>> name: cephfs, metadata pool: cephfs-metadata, data pools: [cephfs-data ]
>>> ---cut here---
>>>
>>> We see the new id 1 when we run
>>>
>>> ---cut here---
>>> ceph3:~ #  ceph fs get cephfs
>>> Filesystem 'cephfs' (1)
>>> fs_name cephfs
>>> [...]
>>> data_pools      [35]
>>> metadata_pool   36
>>> inline_data     disabled
>>> balancer
>>> standby_count_wanted    1
>>> [...]
>>> ---cut here---
>>>
>>> The new FS seems to work properly and can be mounted from the clients,
>>> just like before removing and rebuilding it. I'm not sure which other
>>> commands would fail with this traceback, for now "ceph fs status" is the
>>> only one.
>>>
>>> So it seems that having one deactivated cephFS has an impact on some of
>>> the functions/commands. Is there any way to remove it properly? Most of
>>> the
>>> commands work with the name, not the id of the FS, so it's difficult to
>>> access the data from the old FS. Has anyone some insights on how to clean
>>> this up?
>>>
>>> Regards,
>>> Eugen
>>>
>>> --
>>> Eugen Block                             voice   : +49-40-559 51 75
>>> NDE Netzdesign und -entwicklung AG      fax     : +49-40-559 51 77
>>> Postfach 61 03 15
>>> D-22423 Hamburg                         e-mail  : [email protected]
>>>
>>>         Vorsitzende des Aufsichtsrates: Angelika Mozdzen
>>>           Sitz und Registergericht: Hamburg, HRB 90934
>>>                   Vorstand: Jens-U. Mozdzen
>>>                    USt-IdNr. DE 814 013 983
>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> [email protected]
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>>
>>
>>
>> --
>> Thomas Bennett
>>
>> SKA South Africa
>> Science Processing Team
>>
>> Office: +27 21 5067341
>> Mobile: +27 79 5237105
>>
>
>
>
> --
> Eugen Block                             voice   : +49-40-559 51 75
> NDE Netzdesign und -entwicklung AG      fax     : +49-40-559 51 77
> Postfach 61 03 15
> D-22423 Hamburg                         e-mail  : [email protected]
>
>         Vorsitzende des Aufsichtsrates: Angelika Mozdzen
>           Sitz und Registergericht: Hamburg, HRB 90934
>                   Vorstand: Jens-U. Mozdzen
>                    USt-IdNr. DE 814 013 983
>
>


-- 
Thomas Bennett

SKA South Africa
Science Processing Team

Office: +27 21 5067341
Mobile: +27 79 5237105
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to