[ceph-users] ceph 14.2.6 problem with default args to rbd (--name)

2020-01-20 Thread Rainer Krienke
Hello,

I am fighting with rbd and CEPH_ARGS in order to make typing easier on a
client. First I created a keyring on one of the ceph nodes:

# ceph auth add client.rainer  mon 'profile rbd' osd 'profile rbd'
added key for client.rainer

Then I added this keyring to /etc/ceph/ceph.keyring on a client machine.
# cat /etc/ceph/ceph.keyring:
[client.rainer]
key = xxyxyxyxyxyxyxyxyxyxyxyxyxyxy==

Now on the client I try:

# rbd ls
2020-01-20 14:15:49.124 7fa9191f0700 -1 monclient(hunting):
handle_auth_bad_method server allowed_methods [2] but i only support [2]
2020-01-20 14:15:49.124 7fa9189ef700 -1 monclient(hunting):
handle_auth_bad_method server allowed_methods [2] but i only support [2]
rbd: couldn't connect to the cluster!
rbd: listing images failed: (1) Operation not permitted

Using the --name  arg things work good:

# rbd ls --name client.rainer

...

Now to make typing easier I tried:
# export CEPH_ARGS="--name client.rainer --keyring=/etc/ceph/ceph.keyring"
# rbd ls
2020-01-20 14:18:33.367 7f691177c700 -1 monclient(hunting):
handle_auth_bad_method server allowed_methods [2] but i only support [2]
rbd: couldn't connect to the cluster!
rbd: listing images failed: (1) Operation not permitted

According to the documentation this should work, but it seems it
doesn't. Something I am doing wrong or is this a bug?

Thanks
Rainer
-- 
Rainer Krienke, Uni Koblenz, Rechenzentrum, A22, Universitaetsstrasse 1
56070 Koblenz, Tel: +49261287 1312 Fax +49261287 100 1312
Web: http://userpages.uni-koblenz.de/~krienke
PGP: http://userpages.uni-koblenz.de/~krienke/mypgp.html
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Strange CEPH_ARGS problems

2019-11-15 Thread Rainer Krienke
This is not my day :-)

Yes this flip beween client.rz and client.user was not intended. Its
another typo.  When trying to run rbd I used everywhere the same
client.rz user and the same keyring /etc/ceph/ceph.client.user.keyring.

Sorry
Rainer

Am 15.11.19 um 11:02 schrieb Janne Johansson:
> Is the flip between the client name "rz" and "user" also a mistype? It's
> hard to divinate if it is intentional or not since you are mixing it about.
> 
> 
> Den fre 15 nov. 2019 kl 10:57 skrev Rainer Krienke
> mailto:krie...@uni-koblenz.de>>:
> 
> I found a typo in my post:
> 
> Of course I tried
> 
> export CEPH_ARGS="-n client.rz --keyring="
> 
> and not
> 
> export CEPH_ARGS=="-n client.rz --keyring="
> 
> Thanks
> Rainer
> 
> Am 15.11.19 um 07:46 schrieb Rainer Krienke:
> > Hello,
> >
> > I try to use CEPH_ARGS in order to use eg rbd with a non client.admin
> > user and keyring without extra parameters. On a ceph-client with
> Ubuntu
> > 18.04.3 I get this:
> >
> > # unset CEPH_ARGS
> > # rbd --name=client.user
> --keyring=/etc/ceph/ceph.client.user.keyring ls
> > a
> > b
> > c
> >
> > # export CEPH_ARGS=="-n client.rz --keyring=/etc
> > /ceph/ceph.client.user.keyring"
> > # rbd ls
> > 
> > rbd: couldn't connect to the cluster!
> > rbd: listing images failed: (22) Invalid argument
> >
> > # export CEPH_ARGS=="--keyring=/etc/ceph/ceph.client.user.keyring"> 
> Thanks
> Rainer
> 
> Am 15.11.19 um 07:46 schrieb Rainer Krienke:
> > Hello,
> >
> > I try to use CEPH_ARGS in order to use eg rbd with a non client.admin
> > user and keyring without extra parameters. On a ceph-client with
> Ubuntu
> > 18.04.3 I get this:
> >
> > # unset CEPH_ARGS
> > # rbd --name=client.user
> --keyring=/etc/ceph/ceph.client.user.keyring ls
> > a
> > b
> > c
> >
> > # export CEPH_ARGS=="-n client.rz --keyring=/etc
> > /ceph/ceph.client.user.keyring"
> > # rbd ls
> > 
> > rbd: couldn't connect to the cluster!
> > rbd: listing images failed: (22) Invalid argument
> >
>     > # export CEPH_ARGS=="--keyring=/etc/ceph/ceph.client.user.keyring"
> > # rbd -n client.user ls
> > a
> > b
> > c
> >
> > Is this the desired behavior? I would like to set both user name and
> > keyring to be used, so that I can run rbd without any parameters.
> >
> > How do you do this?
> >
> > Thanks
> > Rainer
> >
> 
> 
> -- 
> Rainer Krienke, Uni Koblenz, Rechenzentrum, A22, Universitaetsstrasse 1
> 56070 Koblenz, Tel: +49261287 1312 Fax +49261287 100 1312
> Web: http://userpages.uni-koblenz.de/~krienke
> PGP: http://userpages.uni-koblenz.de/~krienke/mypgp.html
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 
> 
> 
> -- 
> May the most significant bit of your life be positive.
> > # rbd -n client.user ls
> > a
> > b
> > c
> >
> > Is this the desired behavior? I would like to set both user name and
> > keyring to be used, so that I can run rbd without any parameters.
> >
> > How do you do this?
> >
> > Thanks
> > Rainer
> >
> 
> 
> -- 
> Rainer Krienke, Uni Koblenz, Rechenzentrum, A22, Universitaetsstrasse 1
> 56070 Koblenz, Tel: +49261287 1312 Fax +49261287 100 1312
> Web: http://userpages.uni-koblenz.de/~krienke
> PGP: http://userpages.uni-koblenz.de/~krienke/mypgp.html
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 
> > Thanks
> Rainer
> 
> Am 15.11.19 um 07:46 schrieb Rainer Krienke:
> > Hello,
> >
> > I try to use CEPH_ARGS in order to use eg rbd with a non client.admin
> > user and keyring without extra parameters. On a ceph-client with
> Ubuntu
> > 18.04.3 I get this:
> >
> > # unset CEPH_ARGS
&

Re: [ceph-users] Strange CEPH_ARGS problems

2019-11-15 Thread Rainer Krienke
I found a typo in my post:

Of course I tried

export CEPH_ARGS="-n client.rz --keyring="

and not

export CEPH_ARGS=="-n client.rz --keyring="

Thanks
Rainer

Am 15.11.19 um 07:46 schrieb Rainer Krienke:
> Hello,
> 
> I try to use CEPH_ARGS in order to use eg rbd with a non client.admin
> user and keyring without extra parameters. On a ceph-client with Ubuntu
> 18.04.3 I get this:
> 
> # unset CEPH_ARGS
> # rbd --name=client.user --keyring=/etc/ceph/ceph.client.user.keyring ls
> a
> b
> c
> 
> # export CEPH_ARGS=="-n client.rz --keyring=/etc
> /ceph/ceph.client.user.keyring"
> # rbd ls
> 
> rbd: couldn't connect to the cluster!
> rbd: listing images failed: (22) Invalid argument
> 
> # export CEPH_ARGS=="--keyring=/etc/ceph/ceph.client.user.keyring"
> # rbd -n client.user ls
> a
> b
> c
> 
> Is this the desired behavior? I would like to set both user name and
> keyring to be used, so that I can run rbd without any parameters.
> 
> How do you do this?
> 
> Thanks
> Rainer
> 


-- 
Rainer Krienke, Uni Koblenz, Rechenzentrum, A22, Universitaetsstrasse 1
56070 Koblenz, Tel: +49261287 1312 Fax +49261287 100 1312
Web: http://userpages.uni-koblenz.de/~krienke
PGP: http://userpages.uni-koblenz.de/~krienke/mypgp.html
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Strange CEPH_ARGS problems

2019-11-14 Thread Rainer Krienke
Hello,

I try to use CEPH_ARGS in order to use eg rbd with a non client.admin
user and keyring without extra parameters. On a ceph-client with Ubuntu
18.04.3 I get this:

# unset CEPH_ARGS
# rbd --name=client.user --keyring=/etc/ceph/ceph.client.user.keyring ls
a
b
c

# export CEPH_ARGS=="-n client.rz --keyring=/etc
/ceph/ceph.client.user.keyring"
# rbd ls

rbd: couldn't connect to the cluster!
rbd: listing images failed: (22) Invalid argument

# export CEPH_ARGS=="--keyring=/etc/ceph/ceph.client.user.keyring"
# rbd -n client.user ls
a
b
c

Is this the desired behavior? I would like to set both user name and
keyring to be used, so that I can run rbd without any parameters.

How do you do this?

Thanks
Rainer

-- 
Rainer Krienke, Uni Koblenz, Rechenzentrum, A22, Universitaetsstrasse 1
56070 Koblenz, Tel: +49261287 1312 Fax +49261287 100 1312
Web: http://userpages.uni-koblenz.de/~krienke
PGP: http://userpages.uni-koblenz.de/~krienke/mypgp.html
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Two questions about ceph update/upgrade strategies

2019-06-04 Thread Rainer Krienke
I have a fresh ceph 14.2.1 cluster up and running based on Ubuntu 18.04.
It consists of 9 hosts (+1 admin host). The nine hosts have each 16
ceph-osd daemons running, three in these nine hosts also have
a ceph-mon and a ceph-mgr daemon running. So three hosts are running
osd, mon and also mgr daemons.

Now I am unsure about the right way to go for ceph upgrades and linux
system host updates.

Ceph-Upgrade:
Reading the ceph upgrade docs I ask myself how a future upgrade say to
14.2.2 should be performed correctly? The recommendation says to upgrade
first monitors, then osds etc...

So what is the correct way to go in a mixed setup like mine? Following
the rules strictly would mean not to use ceph-deploy install, but
instead to log into the mon(/osd) hosts and then upgrade only the
ceph-mon package and restart this mon, and then do the same with the
other monitors/osd hosts. After all mons have been successfully upgraded
I should then continue with upgrading OSDs (ceph-osd package) on one
host and restart all osds on this host one after another or reboot the
whole host. Then proceed to the next osd-host.

Is this the correct and best way to go?

Linux system updates:
The second point I would like to hear your opinions about is how you
handle linux system updates? Since even a non ceph linux system package
update might break ceph or even stop the whole linux host from booting,
care has to be taken. So how do you handle this problem? Do you run host
upgrades only manually in a fixed sequence eg first on a osd/mon host
and if the update is successful, then run the linux system package
updates on the other hosts?   Do you use another strategy?

Thanks
Rainer
-- 
Rainer Krienke, Uni Koblenz, Rechenzentrum, A22, Universitaetsstrasse 1
56070 Koblenz, Tel: +49261287 1312 Fax +49261287 100 1312
Web: http://userpages.uni-koblenz.de/~krienke
PGP: http://userpages.uni-koblenz.de/~krienke/mypgp.html
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Erasure code profiles and crush rules. Missing link...?

2019-05-23 Thread Rainer Krienke
Hello,

thanks for the hint. I opened a ticket with a feature request to include
the ec-profile information in the output of ceph osd pool ls detail.

http://tracker.ceph.com/issues/40009

Rainer

Am 22.05.19 um 17:04 schrieb Jan Fajerski:
> On Wed, May 22, 2019 at 03:38:27PM +0200, Rainer Krienke wrote:
>> Am 22.05.19 um 15:16 schrieb Dan van der Ster:
>>
>> Yes this is basically what I was looking for however I had expected that
>> its a little better visible in the output...
> Mind opening a tracker ticket on http://tracker.ceph.com/ so we can have
> this added to the non-json output of ceph osd pool ls detail?
>>

-- 
Rainer Krienke, Uni Koblenz, Rechenzentrum, A22, Universitaetsstrasse 1
56070 Koblenz, Tel: +49261287 1312 Fax +49261287 100 1312
Web: http://userpages.uni-koblenz.de/~krienke
PGP: http://userpages.uni-koblenz.de/~krienke/mypgp.html
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Erasure code profiles and crush rules. Missing link...?

2019-05-22 Thread Rainer Krienke
Am 22.05.19 um 15:16 schrieb Dan van der Ster:

Yes this is basically what I was looking for however I had expected that
its a little better visible in the output...

Rainer
> 
> Is this what you're looking for?
> 
> # ceph osd pool ls detail  -f json | jq .[0].erasure_code_profile
> "jera_4plus2"
> 
> -- Dan

-- 
Rainer Krienke, Uni Koblenz, Rechenzentrum, A22, Universitaetsstrasse 1
56070 Koblenz, Tel: +49261287 1312 Fax +49261287 100 1312
Web: http://userpages.uni-koblenz.de/~krienke
PGP: http://userpages.uni-koblenz.de/~krienke/mypgp.html
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Erasure code profiles and crush rules. Missing link...?

2019-05-22 Thread Rainer Krienke
Hello,

I created an erasure code profile named ecprofile-42 with the following
parameters:

$ ceph osd erasure-code-profile set ecprofile-42 plugin=jerasure k=4 m=2

Next I created a new pool using the ec profile from above:

$ ceph osd pool create my_erasure_pool 64 64  erasure ecprofile-42

The pool created then has an autogenerated crush rule with the contents
as shown at the end of this mail (see: ceph osd crush rule dump
my_erasure_pool).

What I am missing in the output of the crush rule dump below are the k,m
values used for this pool or a "link" from the crushrule to the erasure
code profile that contains these settings and was used creating the pool
and thus the ec crushrule.  If I had several ec profiles and pools
created with the different ec profiles how else could I see which k,m
values were used for the different pools?

For a replicated crush rule there is the size parameter which is part of
the crush-rule and indirectly tells you the number of replicas, but what
about erasure coded pools?

Probably there is somewhere the link I am looking for, but I din't find
it yet...

Thanks Rainer

#
# autogenerated crush rule my_erasure_pool:
#
$ ceph osd crush rule dump my_erasure_pool
{
"rule_id": 1,
"rule_name": "my_erasure_pool",
"ruleset": 1,
"type": 3,
"min_size": 3,
"max_size": 6,
"steps": [
{
"op": "set_chooseleaf_tries",
"num": 5
},
{
"op": "set_choose_tries",
"num": 100
},
{
"op": "take",
"item": -1,
"item_name": "default"
},
{
"op": "chooseleaf_indep",
"num": 0,
"type": "host"
},
{
"op": "emit"
}
]
}

-- 
Rainer Krienke, Uni Koblenz, Rechenzentrum, A22, Universitaetsstrasse 1
56070 Koblenz, Tel: +49261287 1312 Fax +49261287 100 1312
Web: http://userpages.uni-koblenz.de/~krienke
PGP: http://userpages.uni-koblenz.de/~krienke/mypgp.html
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] ceph nautilus namespaces for rbd and rbd image access problem

2019-05-20 Thread Rainer Krienke
Am 20.05.19 um 09:06 schrieb Jason Dillaman:

>> $ rbd --namespace=testnamespace map rbd/rbdtestns --name client.rainer
>> --keyring=/etc/ceph/ceph.keyring
>> rbd: sysfs write failed
>> rbd: error opening image rbdtestns: (1) Operation not permitted
>> In some cases useful info is found in syslog - try "dmesg | tail".
>> 2019-05-20 08:18:29.187 7f42ab7fe700 -1 librbd::image::RefreshRequest:
>> failed to retrieve pool metadata: (1) Operation not permitted
>> 2019-05-20 08:18:29.187 7f42aaffd700 -1 librbd::image::OpenRequest:
>> failed to refresh image: (1) Operation not permitted
>> 2019-05-20 08:18:29.187 7f42aaffd700 -1 librbd::ImageState:
>> 0x561792408860 failed to open image: (1) Operation not permitted
>> rbd: map failed: (22) Invalid argument
> 
> Hmm, it looks like we overlooked updating the 'rbd' profile when PR
> 27423 [1] was merged into v14.2.1. We'll get that fixed, but in the
> meantime, you can add a "class rbd metadata_list" cap on the base pool
> (w/o the namespace restriction) [2].
> 

Thanks for your answer. Well I still have Kernel 4.15 so namespaces
won't work for me at the moment.

Could you please explain what the magic behind "class rbd metadata_list"
is? Is it thought to "simply" allow access to the basepool (rbd in my
case), so I authorize access to the pool instead of a namespaces? And if
this is true then I do not understand the difference of your class cap
compared to a cap like  osd 'allow rw pool=rbd'?

-- 
Rainer Krienke, Uni Koblenz, Rechenzentrum, A22, Universitaetsstrasse 1
56070 Koblenz, Tel: +49261287 1312 Fax +49261287 100 1312
Web: http://userpages.uni-koblenz.de/~krienke
PGP: http://userpages.uni-koblenz.de/~krienke/mypgp.html
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] ceph nautilus namespaces for rbd and rbd image access problem

2019-05-20 Thread Rainer Krienke
Hello,

just saw this message on the client when trying and failing to map the
rbd image:

May 20 08:59:42 client kernel: libceph: bad option at
'_pool_ns=testnamespace'

Rainer

Am 20.05.19 um 08:56 schrieb Rainer Krienke:
> Hello,
> 
> on a ceph Nautilus cluster (14.2.1) running on Ubuntu 18.04 I try to set
> up rbd images with namespaces in order to allow different clients to
> access only their "own" rbd images in different namespaces in just one
> pool. The rbd image data are in an erasure encoded pool named "ecpool"
> and the metadata in the default "rbd" pool.
-- 
Rainer Krienke, Uni Koblenz, Rechenzentrum, A22, Universitaetsstrasse 1
56070 Koblenz, Tel: +49261287 1312 Fax +49261287 100 1312
Web: http://userpages.uni-koblenz.de/~krienke
PGP: http://userpages.uni-koblenz.de/~krienke/mypgp.html
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] ceph nautilus namespaces for rbd and rbd image access problem

2019-05-20 Thread Rainer Krienke
Hello,

on a ceph Nautilus cluster (14.2.1) running on Ubuntu 18.04 I try to set
up rbd images with namespaces in order to allow different clients to
access only their "own" rbd images in different namespaces in just one
pool. The rbd image data are in an erasure encoded pool named "ecpool"
and the metadata in the default "rbd" pool.

With this setup I am experiencing trouble when I try to access a rbd
image in a namespace from a (OpenSuSE Leap 15.0 with Ceph 14.2.1) client
and I do not understand what I am doing wrong. Hope someone can see the
problem and give me a hint:

# On one of the the ceph servers

$ rbd namespace create --namespace testnamespace
$ rbd namespace ls
NAME
testnamespace

$ ceph auth caps client.rainer mon 'profile rbd' osd 'profile rbd
pool=rbd namespace=testnamespace'

$ ceph auth get client.rainer
[client.rainer]
key = AQCcVt5cHC+WJhBBoRPKhErEYzxGuU8U/GA0xA++
caps mon = "profile rbd"
caps osd = "profile rbd pool=rbd namespace=testnamespace"

$ rbd create rbd/rbdtestns --namespace testnamespace --size 50G
--data-pool=rbd-ecpool

$ rbd --namespace testnamespace ls -l
NAME  SIZE   PARENT FMT PROT LOCK
rbdtestns 50 GiB  2

On the openSuSE Client:

$ rbd --namespace=testnamespace map rbd/rbdtestns --name client.rainer
--keyring=/etc/ceph/ceph.keyring
rbd: sysfs write failed
rbd: error opening image rbdtestns: (1) Operation not permitted
In some cases useful info is found in syslog - try "dmesg | tail".
2019-05-20 08:18:29.187 7f42ab7fe700 -1 librbd::image::RefreshRequest:
failed to retrieve pool metadata: (1) Operation not permitted
2019-05-20 08:18:29.187 7f42aaffd700 -1 librbd::image::OpenRequest:
failed to refresh image: (1) Operation not permitted
2019-05-20 08:18:29.187 7f42aaffd700 -1 librbd::ImageState:
0x561792408860 failed to open image: (1) Operation not permitted
rbd: map failed: (22) Invalid argument

Thanks for your help
Rainer
-- 
Rainer Krienke, Uni Koblenz, Rechenzentrum, A22, Universitaetsstrasse 1
56070 Koblenz, Tel: +49261287 1312 Fax +49261287 100 1312
Web: http://userpages.uni-koblenz.de/~krienke
PGP: http://userpages.uni-koblenz.de/~krienke/mypgp.html
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] ceph -s finds 4 pools but ceph osd lspools says no pool which is the expected answer

2019-05-16 Thread Rainer Krienke
Hello Greg,

thank you very much for your hint.
If I should see this problem again I will try to restart the ceph-mgr
daemon and see if this helps.

Rainer

> 
> I don't really see how this particular error can happen and be
> long-lived, but if you restart the ceph-mgr it will probably resolve
> itself.
> ("ceph osd lspools" looks directly at the OSDMap in the monitor,
> whereas the "ceph -s" data output is generated from the manager's
> pgmap, but there's a tight link where the pgmap gets updated and
> removes dead pools on every new OSDMap the manager sees and I can't
> see how that would go wrong.)
> -Greg
> 
> 
>> Thanks
>> Rainer
>> --
>> Rainer Krienke, Uni Koblenz, Rechenzentrum, A22, Universitaetsstrasse  1
>> 56070 Koblenz, Web: http://www.uni-koblenz.de/~krienke, Tel: +49261287 1312
>> PGP: http://www.uni-koblenz.de/~krienke/mypgp.html, Fax: +49261287
>> 1001312
>> ___
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


-- 
Rainer Krienke, Uni Koblenz, Rechenzentrum, A22, Universitaetsstrasse 1
56070 Koblenz, Tel: +49261287 1312 Fax +49261287 100 1312
Web: http://userpages.uni-koblenz.de/~krienke
PGP: http://userpages.uni-koblenz.de/~krienke/mypgp.html
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] ceph -s finds 4 pools but ceph osd lspools says no pool which is the expected answer

2019-05-15 Thread Rainer Krienke
Hello,

since I had no ideas by what the wrong pool number in the ceph -s output
could be caused I simply rebooted all machines of this cluster (it does
not yet contain any real data) which solved the problem.

So it seems that some caching problem might have caused this issue.

Thanks
Rainer

Am 14.05.19 um 20:03 schrieb Rainer Krienke:
> Hello,
> 
> for a fresh setup ceph cluster I see a strange difference in the number
> of existing pools in the output of ceph -s and what I know that should
> be there: no pools at all.
> 
> I set up a fresh Nautilus cluster with 144 OSDs on 9 hosts. Just to play
> around I created a pool named rbd with
> 



-- 
Rainer Krienke, Uni Koblenz, Rechenzentrum, A22, Universitaetsstrasse 1
56070 Koblenz, Tel: +49261287 1312 Fax +49261287 100 1312
Web: http://userpages.uni-koblenz.de/~krienke
PGP: http://userpages.uni-koblenz.de/~krienke/mypgp.html
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] ceph -s finds 4 pools but ceph osd lspools says no pool which is the expected answer

2019-05-14 Thread Rainer Krienke
Hello,

for a fresh setup ceph cluster I see a strange difference in the number
of existing pools in the output of ceph -s and what I know that should
be there: no pools at all.

I set up a fresh Nautilus cluster with 144 OSDs on 9 hosts. Just to play
around I created a pool named rbd with

$ ceph osd pool create rbd 512 512 replicated

In ceph -s I saw the pool but also saw a warning:

 cluster:
id: a-b-c-d-e
health: HEALTH_WARN
too few PGs per OSD (21 < min 30)

So I experimented around, removed the pool (ceph osd pool remove rbd)
and it was gone in ceph osd lspools, and created a new one with some
more PGs and repeated this a few times with larger PG nums. In the end
in the output of ceph -s I see that 4 pools do exist:

  cluster:
id: a-b-c-d-e
health: HEALTH_OK

  services:
mon: 3 daemons, quorum c2,c5,c8 (age 8h)
mgr: c2(active, since 8h)
osd: 144 osds: 144 up (since 8h), 144 in (since 8h)

  data:
pools:   4 pools, 0 pgs
objects: 0 objects, 0 B
usage:   155 GiB used, 524 TiB / 524 TiB avail
pgs:

but:

$ ceph osd lspools


Since I deleted each pool I created, 0 pools is the correct answer.
I could add another "ghost" pool by creating another pool named rbd with
only 512 PGs and then delete it again right away. ceph -s would then
show me 5 pools. This is the way I came from 3 to 4 "ghost pools".

This does not seem to happen if I use 2048 PGs for the new pool which I
do delete right afterwards. In this case the pool is created and ceph -s
shows one pool more (5) and if delete this pool again the counter in
ceph -s goes back to 4 again.

How can I fix the system so that ceph -s also understands that are
actually no pools? There must be some inconsistency. Any ideas?

Thanks
Rainer
--
Rainer Krienke, Uni Koblenz, Rechenzentrum, A22, Universitaetsstrasse  1
56070 Koblenz, Web: http://www.uni-koblenz.de/~krienke, Tel: +49261287 1312
PGP: http://www.uni-koblenz.de/~krienke/mypgp.html, Fax: +49261287
1001312
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Need some advice about Pools and Erasure Coding

2019-04-29 Thread Rainer Krienke
I am planning to set up a ceph cluster and already implemented a test
cluster where we are going to use RBD images for data storage (9 hosts,
each host has 16 OSDs, each OSD 4TB).
We would like to use erasure coded (EC)  pools here, and so all OSD are
bluestore. Since several projects are going to store data on this ceph
cluster I think it would make sense to use several EC coded pools for
separation of the projects and access control.

Now I have some questions I hope someone can help me with:

- Do I still (nautilus) need two pools for EC based RBD images, one EC
data pool and a second replicated pool for metadatata?

- If I do need two pools for RBD images and I want to separate the data of
different projects by using different pools with EC coding then how
should I handle the metadata pool which contains probably only a small
amount of data compared to the data pool?  Does it make sense to have
*one* replicated metadata pool (eg the default rbd pool) for all
projects and one EC pool for each project, or would it be better to
create one replicated and one EC pool for each project?

- I also thought about the different k+m settings for a EC pool, for
example k=4, m=2 compared to k=8 and m=2. Both settings allow for two
OSDs to fail without any data loss, but I asked myself which of the two
settings would be more performant? On one hand distributing data to more
OSDs allows a higher parallel access to the data, that should result in
a faster access. On the other hand each OSD has a latency until
it can deliver its data shard. So is there a recommandation which of my
two k+m examples should be preferred?

Thanks in advance for your help
Rainer
-- 
Rainer Krienke, Uni Koblenz, Rechenzentrum, A22, Universitaetsstrasse 1
56070 Koblenz, Tel: +49261287 1312 Fax +49261287 100 1312
Web: http://userpages.uni-koblenz.de/~krienke
PGP: http://userpages.uni-koblenz.de/~krienke/mypgp.html
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Problems with osd creation in Ubuntu 18.04, ceph 13.2.4-1bionic

2019-02-17 Thread Rainer Krienke
Hello,

thanks for your answer, but zapping the disk did not make any
difference. I still get the same error.  Looking at the debug output I
found this error message that is probably the root of all trouble:

# ceph-volume lvm prepare --bluestore --data /dev/sdg

stderr: 2019-02-18 08:29:25.544 7fdaa50ed240 -1
bluestore(/var/lib/ceph/osd/ceph-0/) _read_fsid unparsable uuid

I found the bugreport below that seems to be exactly that problem I have:
http://tracker.ceph.com/issues/15386

However there seems to be no solution  up to now.

Does anyone have more information how to get around this problem?

Thanks
Rainer

Am 15.02.19 um 18:12 schrieb David Turner:
> I have found that running a zap before all prepare/create commands with
> ceph-volume helps things run smoother.  Zap is specifically there to
> clear everything on a disk away to make the disk ready to be used as an
> OSD.  Your wipefs command is still fine, but then I would lvm zap the
> disk before continuing.  I would run the commands like [1] this.  I also
> prefer the single command lvm create as opposed to lvm prepare and lvm
> activate.  Try that out and see if you still run into the problems
> creating the BlueStore filesystem.
> 
> [1] ceph-volume lvm zap /dev/sdg
> ceph-volume lvm prepare --bluestore --data /dev/sdg
> 
> On Thu, Feb 14, 2019 at 10:25 AM Rainer Krienke  <mailto:krie...@uni-koblenz.de>> wrote:
> 
> Hi,
> 
> I am quite new to ceph and just try to set up a ceph cluster. Initially
> I used ceph-deploy for this but when I tried to create a BlueStore osd
> ceph-deploy fails. Next I tried the direct way on one of the OSD-nodes
> using ceph-volume to create the osd, but this also fails. Below you can
> see what  ceph-volume says.
> 
> I ensured that there was no left over lvm VG and LV on the disk sdg
> before I started the osd creation for this disk. The very same error
> happens also on other disks not just for /dev/sdg. All the disk have 4TB
> in size and the linux system is Ubuntu 18.04 and finally ceph is
> installed in version 13.2.4-1bionic from this repo:
> https://download.ceph.com/debian-mimic.
> 
> There is a VG and two LV's  on the system for the ubuntu system itself
> that is installed on two separate disks configured as software raid1 and
> lvm on top of the raid. But I cannot imagine that this might do any harm
> to cephs osd creation.
> 
> Does anyone have an idea what might be wrong?
> 
> Thanks for hints
> Rainer
> 
> root@ceph1:~# wipefs -fa /dev/sdg
> root@ceph1:~# ceph-volume lvm prepare --bluestore --data /dev/sdg
> Running command: /usr/bin/ceph-authtool --gen-print-key
> Running command: /usr/bin/ceph --cluster ceph --name
> client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring
> -i - osd new 14d041d6-0beb-4056-8df2-3920e2febce0
> Running command: /sbin/vgcreate --force --yes
> ceph-1433ffd0-0a80-481a-91f5-d7a47b78e17b /dev/sdg
>  stdout: Physical volume "/dev/sdg" successfully created.
>  stdout: Volume group "ceph-1433ffd0-0a80-481a-91f5-d7a47b78e17b"
> successfully created
> Running command: /sbin/lvcreate --yes -l 100%FREE -n
> osd-block-14d041d6-0beb-4056-8df2-3920e2febce0
> ceph-1433ffd0-0a80-481a-91f5-d7a47b78e17b
>  stdout: Logical volume "osd-block-14d041d6-0beb-4056-8df2-3920e2febce0"
> created.
> Running command: /usr/bin/ceph-authtool --gen-print-key
> Running command: /bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-0
> --> Absolute path not found for executable: restorecon
> --> Ensure $PATH environment variable contains common executable
> locations
> Running command: /bin/chown -h ceph:ceph
> 
> /dev/ceph-1433ffd0-0a80-481a-91f5-d7a47b78e17b/osd-block-14d041d6-0beb-4056-8df2-3920e2febce0
> Running command: /bin/chown -R ceph:ceph /dev/dm-8
> Running command: /bin/ln -s
> 
> /dev/ceph-1433ffd0-0a80-481a-91f5-d7a47b78e17b/osd-block-14d041d6-0beb-4056-8df2-3920e2febce0
> /var/lib/ceph/osd/ceph-0/block
> Running command: /usr/bin/ceph --cluster ceph --name
> client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring
> mon getmap -o /var/lib/ceph/osd/ceph-0/activate.monmap
>  stderr: got monmap epoch 1
> Running command: /usr/bin/ceph-authtool /var/lib/ceph/osd/ceph-0/keyring
> --create-keyring --name osd.0 --add-key
> AQAAY2VcU968HxAAvYWMaJZmriUc4H9bCCp8XQ==
>  stdout: creating /var/lib/ceph/osd/ceph-0/keyring
> added entity osd.0 auth auth(auid = 18446744073709551615
> key=AQAAY2VcU968HxAAvYWMaJZmriUc4H9bCCp8XQ== with 0 caps)
> Running command: /bin/

[ceph-users] Problems with osd creation in Ubuntu 18.04, ceph 13.2.4-1bionic

2019-02-14 Thread Rainer Krienke
gned long, unsigned long,
ceph::buffer::list*, IOContext*, bool)+0x4a7) [0x561371346817]
 stderr: 8: (BlueFS::_read(BlueFS::FileReader*,
BlueFS::FileReaderBuffer*, unsigned long, unsigned long,
ceph::buffer::list*, char*)+0x435) [0x5613713065c5]
 stderr: 9: (BlueFS::_replay(bool, bool)+0x214) [0x56137130c434]
 stderr: 10: (BlueFS::mount()+0x1f1) [0x561371310c81]
 stderr: 11: (BlueStore::_open_db(bool, bool)+0x17cd) [0x56137123704d]
 stderr: 12: (BlueStore::mkfs()+0x805) [0x561371267fe5]
 stderr: 13: (OSD::mkfs(CephContext*, ObjectStore*,
std::__cxx11::basic_string,
std::allocator > const&, uuid_d, int)+0x1b0) [0x561370e10480]
 stderr: 14: (main()+0x4222) [0x561370cf7462]
 stderr: 15: (__libc_start_main()+0xe7) [0x7f3fc3695b97]
 stderr: 16: (_start()+0x2a) [0x561370dc095a]
 stderr: NOTE: a copy of the executable, or `objdump -rdS `
is needed to interpret this.
 stderr: 0> 2019-02-14 13:45:54.840 7f3fcecb3240 -1 *** Caught signal
(Aborted) **
 stderr: in thread 7f3fcecb3240 thread_name:ceph-osd
 stderr: ceph version 13.2.4 (b10be4d44915a4d78a8e06aa31919e74927b142e)
mimic (stable)
 stderr: 1: (()+0x92aa40) [0x561371357a40]
 stderr: 2: (()+0x12890) [0x7f3fc47d7890]
 stderr: 3: (gsignal()+0xc7) [0x7f3fc36b2e97]
 stderr: 4: (abort()+0x141) [0x7f3fc36b4801]
 stderr: 5: (ceph::__ceph_assert_fail(char const*, char const*, int,
char const*)+0x250) [0x7f3fc60d3530]
 stderr: 6: (()+0x26d5a7) [0x7f3fc60d35a7]
 stderr: 7: (KernelDevice::read(unsigned long, unsigned long,
ceph::buffer::list*, IOContext*, bool)+0x4a7) [0x561371346817]
 stderr: 8: (BlueFS::_read(BlueFS::FileReader*,
BlueFS::FileReaderBuffer*, unsigned long, unsigned long,
ceph::buffer::list*, char*)+0x435) [0x5613713065c5]
 stderr: 9: (BlueFS::_replay(bool, bool)+0x214) [0x56137130c434]
 stderr: 10: (BlueFS::mount()+0x1f1) [0x561371310c81]
 stderr: 11: (BlueStore::_open_db(bool, bool)+0x17cd) [0x56137123704d]
 stderr: 12: (BlueStore::mkfs()+0x805) [0x561371267fe5]
 stderr: 13: (OSD::mkfs(CephContext*, ObjectStore*,
std::__cxx11::basic_string,
std::allocator > const&, uuid_d, int)+0x1b0) [0x561370e10480]
 stderr: 14: (main()+0x4222) [0x561370cf7462]
 stderr: 15: (__libc_start_main()+0xe7) [0x7f3fc3695b97]
 stderr: 16: (_start()+0x2a) [0x561370dc095a]
 stderr: NOTE: a copy of the executable, or `objdump -rdS `
is needed to interpret this.
--> Was unable to complete a new OSD, will rollback changes
Running command: /usr/bin/ceph --cluster ceph --name
client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring
osd purge-new osd.0 --yes-i-really-mean-it
 stderr: purged osd.0
-->  RuntimeError: Command failed with exit code 250: /usr/bin/ceph-osd
--cluster ceph --osd-objectstore bluestore --mkfs -i 0 --monmap
/var/lib/ceph/osd/ceph-0/activate.monmap --keyfile - --osd-data
/var/lib/ceph/osd/ceph-0/ --osd-uuid
14d041d6-0beb-4056-8df2-3920e2febce0 --setuser ceph --setgroup ceph
-- 
Rainer Krienke, Uni Koblenz, Rechenzentrum, A22, Universitaetsstrasse 1
56070 Koblenz, Tel: +49261287 1312 Fax +49261287 100 1312
Web: http://userpages.uni-koblenz.de/~krienke
PGP: http://userpages.uni-koblenz.de/~krienke/mypgp.html
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com