Re: [ceph-users] RGW pools don't show up in luminous

2018-08-24 Thread Robert Stanford
Casey - this was exactly it. My ceph-mgr had issues. I didn't know this was necessary for ceph df to work. Thank you R On Fri, Aug 24, 2018 at 8:56 AM Casey Bodley wrote: > > > On 08/23/2018 01:22 PM, Robert Stanford wrote: > > > > I installed a new Ceph cluster with Luminous, after a

Re: [ceph-users] cephfs kernel client hangs

2018-08-24 Thread Yan, Zheng
Are there hang request in /sys/kernel/debug/ceph//osdc On Fri, Aug 24, 2018 at 9:32 PM Zhenshi Zhou wrote: > > I'm afaid that the client hangs again...the log shows: > > 2018-08-24 21:27:54.714334 [WRN] slow request 62.607608 seconds old, > received at 2018-08-24 21:26:52.106633:

Re: [ceph-users] ceph-fuse slow cache?

2018-08-24 Thread Gregory Farnum
On Fri, Aug 24, 2018 at 1:20 AM Stefan Kooman wrote: > Hi Gregory, > > Quoting Gregory Farnum (gfar...@redhat.com): > > This is quite strange. Given that you have a log, I think what you want > to > > do is find one request in the log, trace it through its lifetime, and see > > where the time is

[ceph-users] mimic - troubleshooting prometheus

2018-08-24 Thread Steven Vacaroaia
Hi, Any idea/suggestions for troubleshooting prometheus ? what logs /commands are available to find out why OSD servers specific data ( IOPS, disk and network data) is not scrapped but cluster specific data ( pools, capacity ..etc) is ? Increasing log level for MGR showed only the following

Re: [ceph-users] Mimic prometheus plugin -no socket could be created

2018-08-24 Thread Steven Vacaroaia
To have prometheus plugin working you HAVE to tell it to listen to an IPV4 address ...like this ceph config set mgr mgr/prometheus/server_addr 0.0.0.0 On Fri, 24 Aug 2018 at 12:44, Jones de Andrade wrote: > Hi all. > > I'm new to ceph, and after having serious problems in ceph stages 0, 1 and

[ceph-users] Ceph-Deploy error on 15/71 stage

2018-08-24 Thread Jones de Andrade
(Please forgive my previous email: I was using another message and completely forget to update the subject) Hi all. I'm new to ceph, and after having serious problems in ceph stages 0, 1 and 2 that I could solve myself, now it seems that I have hit a wall harder than my head. :) When I run

Re: [ceph-users] Mimic prometheus plugin -no socket could be created

2018-08-24 Thread Jones de Andrade
Hi all. I'm new to ceph, and after having serious problems in ceph stages 0, 1 and 2 that I could solve myself, now it seems that I have hit a wall harder than my head. :) When I run salt-run state.orch ceph.stage.deploy, i monitor I see it going up to here: ### [14/71] ceph.sysctl on

Re: [ceph-users] ceph auto repair. What is wrong?

2018-08-24 Thread Fyodor Ustinov
Hi! Did not help. :( HEALTH_WARN 3 osds down; 1 host (3 osds) down; 1 rack (3 osds) down; Degraded data redundancy: 112 pgs undersized OSD_DOWN 3 osds down osd.24 (root=default,rack=R-26-7-1,host=S-26-7-1-1) is down osd.25 (root=default,rack=R-26-7-1,host=S-26-7-1-1) is down

Re: [ceph-users] Clients report OSDs down/up (dmesg) nothing in Ceph logs (flapping OSDs)

2018-08-24 Thread Eugen Block
Update: I changed the primary affinity of one OSD back to 1.0 to test if those metrics change, and indeed they do: OSD.24 immediately shows values greater than 0. I guess the metrics are completely unrelated to the flapping. So the search goes on... Zitat von Eugen Block : An hour ago

Re: [ceph-users] Reminder: bi-weekly dashboard sync call today (15:00 CET)

2018-08-24 Thread Ernesto Puerta
My mistake, Lenz. That recording is just the 7 minutes of fun before everyone joined. This is the proper one (~1 hour): https://bluejeans.com/s/HUofE Ernesto ERNESTO PUERTA SENIOR SOFTWARE ENGINEER, CEPH R Red Hat On Fri, Aug 24, 2018 at 4:38 PM Lenz Grimmer wrote: > > On 08/24/2018 02:00

[ceph-users] rbd + openshift cause cpu stuck now and then

2018-08-24 Thread Jeffrey Zhang
I am testing openshift with ceph rbd, it works as expected. except that sometimes the container which has a rbd volume start slowly. And the load on the node that containers running will pretty high, until following error raise in dmesg. After some google, i found one similar issue at[0]. seems

Re: [ceph-users] Reminder: bi-weekly dashboard sync call today (15:00 CET)

2018-08-24 Thread Lenz Grimmer
On 08/24/2018 02:00 PM, Lenz Grimmer wrote: > On 08/24/2018 10:59 AM, Lenz Grimmer wrote: > >> JFYI, the team working on the Ceph Manager Dashboard has a bi-weekly >> conference call that discusses the ongoing development and gives an >> update on recent improvements/features. >> >> Today, we

Re: [ceph-users] Migrating from pre-luminous multi-root crush hierachy

2018-08-24 Thread Paul Emmerich
The proper way would be to do this change atomically by adjusting the crush hierarchy and rules at the same time by editing and setting the crush map manually. Paul 2018-08-24 9:40 GMT+02:00 Konstantin Shalygin : > On 08/24/2018 01:57 PM, Buchberger, Carsten wrote: >> >> Hi Konstantin, >> >>

Re: [ceph-users] ceph auto repair. What is wrong?

2018-08-24 Thread Fyodor Ustinov
Hi! I.e. I have to do ceph config set mon mon_osd_down_out_subtree_limit row and restart every mon? On 08/24/18 12:44, Paul Emmerich wrote: Ceph doesn't mark out whole racks by default, set mon_osd_down_out_subtree_limit to something higher like row or pod. Paul 2018-08-24 10:50 GMT+02:00

Re: [ceph-users] radosgw: need couple of blind (indexless) buckets, how-to?

2018-08-24 Thread Casey Bodley
On 08/24/2018 06:44 AM, Konstantin Shalygin wrote: Answer to myself. radosgw-admin realm create --rgw-realm=default --default radosgw-admin zonegroup modify --rgw-zonegroup=default --rgw-realm=default radosgw-admin period update --commit radosgw-admin zonegroup placement add

Re: [ceph-users] RGW pools don't show up in luminous

2018-08-24 Thread Casey Bodley
On 08/23/2018 01:22 PM, Robert Stanford wrote:  I installed a new Ceph cluster with Luminous, after a long time working with Jewel.  I created my RGW pools the same as always (pool create default.rgw.buckets.data etc.), but they don't show up in ceph df with Luminous.  Has the command

Re: [ceph-users] cephfs kernel client hangs

2018-08-24 Thread Zhenshi Zhou
I'm afaid that the client hangs again...the log shows: 2018-08-24 21:27:54.714334 [WRN] slow request 62.607608 seconds old, received at 2018-08-24 21:26:52.106633: client_request(client.213528:241811 getattr pAsLsXsFs #0x12e7e5a 2018-08-24 21:26:52.106425 caller_uid=0, caller_gid=0{})

Re: [ceph-users] Reminder: bi-weekly dashboard sync call today (15:00 CET)

2018-08-24 Thread Lenz Grimmer
On 08/24/2018 10:59 AM, Lenz Grimmer wrote: > JFYI, the team working on the Ceph Manager Dashboard has a bi-weekly > conference call that discusses the ongoing development and gives an > update on recent improvements/features. > > Today, we plan to give a demo of the new dashboard landing page

Re: [ceph-users] radosgw: need couple of blind (indexless) buckets, how-to?

2018-08-24 Thread Konstantin Shalygin
Answer to myself. radosgw-admin realm create --rgw-realm=default --default radosgw-admin zonegroup modify --rgw-zonegroup=default --rgw-realm=default radosgw-admin period update --commit radosgw-admin zonegroup placement add --rgw-zonegroup="default" \   --placement-id="indexless-placement"

Re: [ceph-users] Stability Issue with 52 OSD hosts

2018-08-24 Thread Andras Pataki
We pin half the OSDs to each socket (and to the corresponding memory).  Since the disk controller and the network card is connected only to one socket, this still probably produces quite a bit of QPI traffic. It is also worth investigating how the network does under high load.  We did run into

Re: [ceph-users] ceph auto repair. What is wrong?

2018-08-24 Thread Paul Emmerich
Ceph doesn't mark out whole racks by default, set mon_osd_down_out_subtree_limit to something higher like row or pod. Paul 2018-08-24 10:50 GMT+02:00 Christian Balzer : > Hello, > > On Fri, 24 Aug 2018 11:30:34 +0300 (EEST) Fyodor Ustinov wrote: > >> Hi! >> >> I wait about hour. >> > Aside from

Re: [ceph-users] Stability Issue with 52 OSD hosts

2018-08-24 Thread Marc Roos
Can this be related to numa issues? I have also dual processor nodes, and was wondering if there is some guide on how to optimize for numa. -Original Message- From: Tyler Bishop [mailto:tyler.bis...@beyondhosting.net] Sent: vrijdag 24 augustus 2018 3:11 To: Andras Pataki Cc:

Re: [ceph-users] Ceph RGW Index Sharding In Jewel

2018-08-24 Thread Alexandru Cucu
You should probably have a look at ceph-ansible as it has a "take-over-existing-cluster" playbook. I think versions older than 2.0 support Ceph versions older than Jewel. --- Alex Cucu On Fri, Aug 24, 2018 at 4:31 AM Russell Holloway wrote: > > Thanks. Unfortunately even my version of hammer is

[ceph-users] Reminder: bi-weekly dashboard sync call today (15:00 CET)

2018-08-24 Thread Lenz Grimmer
Hi all, JFYI, the team working on the Ceph Manager Dashboard has a bi-weekly conference call that discusses the ongoing development and gives an update on recent improvements/features. Today, we plan to give a demo of the new dashboard landing page (See https://tracker.ceph.com/issues/24573 and

Re: [ceph-users] ceph auto repair. What is wrong?

2018-08-24 Thread Christian Balzer
Hello, On Fri, 24 Aug 2018 11:30:34 +0300 (EEST) Fyodor Ustinov wrote: > Hi! > > I wait about hour. > Aside from verifying those timeout values in your cluster, what's your mon_osd_down_out_subtree_limit set to? Christian > - Original Message - > From: "Wido den Hollander" > To:

Re: [ceph-users] ceph auto repair. What is wrong?

2018-08-24 Thread Fyodor Ustinov
Hi! I wait about hour. - Original Message - From: "Wido den Hollander" To: "Fyodor Ustinov" , ceph-users@lists.ceph.com Sent: Friday, 24 August, 2018 09:52:23 Subject: Re: [ceph-users] ceph auto repair. What is wrong? On 08/24/2018 06:11 AM, Fyodor Ustinov wrote: > Hi! > > I have

Re: [ceph-users] ceph-fuse slow cache?

2018-08-24 Thread Stefan Kooman
Hi Gregory, Quoting Gregory Farnum (gfar...@redhat.com): > This is quite strange. Given that you have a log, I think what you want to > do is find one request in the log, trace it through its lifetime, and see > where the time is elapsed. You may find a bifurcation, where some > categories of

Re: [ceph-users] Shared WAL/DB device partition for multiple OSDs?

2018-08-24 Thread Eugen Block
Hi, I don't know why but, I noticed in the ceph-volume-systemd.log (above in bold), that there are 2 different lines corresponding to the lvm-1 (normally associated to the osd.1) ? One seems to have the correct id, while the other has a bad one...and it's looks like he's trying to start

Re: [ceph-users] RGW pools don't show up in luminous

2018-08-24 Thread Konstantin Shalygin
I installed a new Ceph cluster with Luminous, after a long time working with Jewel. I created my RGW pools the same as always (pool create default.rgw.buckets.data etc.), but they don't show up in ceph df with Luminous. Has the command changed? Since Luminous you don't need to create pools.

Re: [ceph-users] Shared WAL/DB device partition for multiple OSDs?

2018-08-24 Thread Hervé Ballans
Le 23/08/2018 à 18:44, Alfredo Deza a écrit : ceph-volume-systemd.log (extract) [2018-08-20 11:26:26,386][systemd][INFO ] raw systemd input received: lvm-6-ba351d69-5c48-418e-a377-4034f503af93 [2018-08-20 11:26:26,386][systemd][INFO ] raw systemd input received:

Re: [ceph-users] Migrating from pre-luminous multi-root crush hierachy

2018-08-24 Thread Konstantin Shalygin
On 08/24/2018 01:57 PM, Buchberger, Carsten wrote: Hi Konstantin, sounds easy;-) If i apply the new rule to the existing pools there won't be any osds to satisfy the requirements of the rule - because the osds are not in the new root yet. Isn't that a problem ? Thank you Your IO will

[ceph-users] PG auto repair with BlueStore

2018-08-24 Thread Wido den Hollander
Hi, osd_scrub_auto_repair still defaults to false and I was wondering how we think about enabling this feature by default. Would we say it's safe to enable this with BlueStore? Wido ___ ceph-users mailing list ceph-users@lists.ceph.com

Re: [ceph-users] ceph auto repair. What is wrong?

2018-08-24 Thread Wido den Hollander
On 08/24/2018 06:11 AM, Fyodor Ustinov wrote: > Hi! > > I have fresh ceph cluster. 12 host and 3 osd on each host (one - hdd and two > - ssd). Each host located in own rack. > > I make such crush configuration on fresh ceph installation: > >sudo ceph osd crush add-bucket R-26-3-1 rack >

Re: [ceph-users] Migrating from pre-luminous multi-root crush hierachy

2018-08-24 Thread Konstantin Shalygin
We recently upgrade to luminous (you can see the device-classes in the output). So it should be possible to have one single root, no fake hosts and just use the device-class. We added some hosts/osds recently which back a new pools, so we also created a new hierarchy and crush rules for those.