Casey - this was exactly it. My ceph-mgr had issues. I didn't know this
was necessary for ceph df to work. Thank you
R
On Fri, Aug 24, 2018 at 8:56 AM Casey Bodley wrote:
>
>
> On 08/23/2018 01:22 PM, Robert Stanford wrote:
> >
> > I installed a new Ceph cluster with Luminous, after a
Are there hang request in /sys/kernel/debug/ceph//osdc
On Fri, Aug 24, 2018 at 9:32 PM Zhenshi Zhou wrote:
>
> I'm afaid that the client hangs again...the log shows:
>
> 2018-08-24 21:27:54.714334 [WRN] slow request 62.607608 seconds old,
> received at 2018-08-24 21:26:52.106633:
On Fri, Aug 24, 2018 at 1:20 AM Stefan Kooman wrote:
> Hi Gregory,
>
> Quoting Gregory Farnum (gfar...@redhat.com):
> > This is quite strange. Given that you have a log, I think what you want
> to
> > do is find one request in the log, trace it through its lifetime, and see
> > where the time is
Hi,
Any idea/suggestions for troubleshooting prometheus ?
what logs /commands are available to find out why OSD servers specific
data ( IOPS, disk and network data) is not scrapped but cluster specific
data ( pools, capacity ..etc) is ?
Increasing log level for MGR showed only the following
To have prometheus plugin working you HAVE to tell it to listen to an IPV4
address ...like this
ceph config set mgr mgr/prometheus/server_addr 0.0.0.0
On Fri, 24 Aug 2018 at 12:44, Jones de Andrade wrote:
> Hi all.
>
> I'm new to ceph, and after having serious problems in ceph stages 0, 1 and
(Please forgive my previous email: I was using another message and
completely forget to update the subject)
Hi all.
I'm new to ceph, and after having serious problems in ceph stages 0, 1 and
2 that I could solve myself, now it seems that I have hit a wall harder
than my head. :)
When I run
Hi all.
I'm new to ceph, and after having serious problems in ceph stages 0, 1 and
2 that I could solve myself, now it seems that I have hit a wall harder
than my head. :)
When I run salt-run state.orch ceph.stage.deploy, i monitor I see it going
up to here:
###
[14/71] ceph.sysctl on
Hi!
Did not help. :(
HEALTH_WARN 3 osds down; 1 host (3 osds) down; 1 rack (3 osds) down;
Degraded data redundancy: 112 pgs undersized
OSD_DOWN 3 osds down
osd.24 (root=default,rack=R-26-7-1,host=S-26-7-1-1) is down
osd.25 (root=default,rack=R-26-7-1,host=S-26-7-1-1) is down
Update:
I changed the primary affinity of one OSD back to 1.0 to test if those
metrics change, and indeed they do:
OSD.24 immediately shows values greater than 0.
I guess the metrics are completely unrelated to the flapping.
So the search goes on...
Zitat von Eugen Block :
An hour ago
My mistake, Lenz. That recording is just the 7 minutes of fun before
everyone joined.
This is the proper one (~1 hour): https://bluejeans.com/s/HUofE
Ernesto
ERNESTO PUERTA
SENIOR SOFTWARE ENGINEER, CEPH R
Red Hat
On Fri, Aug 24, 2018 at 4:38 PM Lenz Grimmer wrote:
>
> On 08/24/2018 02:00
I am testing openshift with ceph rbd, it works as expected. except that
sometimes the container which has a rbd volume start slowly. And the load
on the node that containers running will pretty high, until following error
raise in dmesg.
After some google, i found one similar issue at[0]. seems
On 08/24/2018 02:00 PM, Lenz Grimmer wrote:
> On 08/24/2018 10:59 AM, Lenz Grimmer wrote:
>
>> JFYI, the team working on the Ceph Manager Dashboard has a bi-weekly
>> conference call that discusses the ongoing development and gives an
>> update on recent improvements/features.
>>
>> Today, we
The proper way would be to do this change atomically by adjusting the
crush hierarchy and rules at the same time by editing and setting the
crush map manually.
Paul
2018-08-24 9:40 GMT+02:00 Konstantin Shalygin :
> On 08/24/2018 01:57 PM, Buchberger, Carsten wrote:
>>
>> Hi Konstantin,
>>
>>
Hi!
I.e. I have to do
ceph config set mon mon_osd_down_out_subtree_limit row
and restart every mon?
On 08/24/18 12:44, Paul Emmerich wrote:
Ceph doesn't mark out whole racks by default, set
mon_osd_down_out_subtree_limit to something higher like row or pod.
Paul
2018-08-24 10:50 GMT+02:00
On 08/24/2018 06:44 AM, Konstantin Shalygin wrote:
Answer to myself.
radosgw-admin realm create --rgw-realm=default --default
radosgw-admin zonegroup modify --rgw-zonegroup=default --rgw-realm=default
radosgw-admin period update --commit
radosgw-admin zonegroup placement add
On 08/23/2018 01:22 PM, Robert Stanford wrote:
I installed a new Ceph cluster with Luminous, after a long time
working with Jewel. I created my RGW pools the same as always (pool
create default.rgw.buckets.data etc.), but they don't show up in ceph
df with Luminous. Has the command
I'm afaid that the client hangs again...the log shows:
2018-08-24 21:27:54.714334 [WRN] slow request 62.607608 seconds old,
received at 2018-08-24 21:26:52.106633: client_request(client.213528:241811
getattr pAsLsXsFs #0x12e7e5a 2018-08-24 21:26:52.106425 caller_uid=0,
caller_gid=0{})
On 08/24/2018 10:59 AM, Lenz Grimmer wrote:
> JFYI, the team working on the Ceph Manager Dashboard has a bi-weekly
> conference call that discusses the ongoing development and gives an
> update on recent improvements/features.
>
> Today, we plan to give a demo of the new dashboard landing page
Answer to myself.
radosgw-admin realm create --rgw-realm=default --default
radosgw-admin zonegroup modify --rgw-zonegroup=default --rgw-realm=default
radosgw-admin period update --commit
radosgw-admin zonegroup placement add --rgw-zonegroup="default" \
--placement-id="indexless-placement"
We pin half the OSDs to each socket (and to the corresponding memory).
Since the disk controller and the network card is connected only to one
socket, this still probably produces quite a bit of QPI traffic.
It is also worth investigating how the network does under high load. We
did run into
Ceph doesn't mark out whole racks by default, set
mon_osd_down_out_subtree_limit to something higher like row or pod.
Paul
2018-08-24 10:50 GMT+02:00 Christian Balzer :
> Hello,
>
> On Fri, 24 Aug 2018 11:30:34 +0300 (EEST) Fyodor Ustinov wrote:
>
>> Hi!
>>
>> I wait about hour.
>>
> Aside from
Can this be related to numa issues? I have also dual processor nodes,
and was wondering if there is some guide on how to optimize for numa.
-Original Message-
From: Tyler Bishop [mailto:tyler.bis...@beyondhosting.net]
Sent: vrijdag 24 augustus 2018 3:11
To: Andras Pataki
Cc:
You should probably have a look at ceph-ansible as it has a
"take-over-existing-cluster" playbook. I think versions older than 2.0
support Ceph versions older than Jewel.
---
Alex Cucu
On Fri, Aug 24, 2018 at 4:31 AM Russell Holloway
wrote:
>
> Thanks. Unfortunately even my version of hammer is
Hi all,
JFYI, the team working on the Ceph Manager Dashboard has a bi-weekly
conference call that discusses the ongoing development and gives an
update on recent improvements/features.
Today, we plan to give a demo of the new dashboard landing page (See
https://tracker.ceph.com/issues/24573 and
Hello,
On Fri, 24 Aug 2018 11:30:34 +0300 (EEST) Fyodor Ustinov wrote:
> Hi!
>
> I wait about hour.
>
Aside from verifying those timeout values in your cluster, what's your
mon_osd_down_out_subtree_limit set to?
Christian
> - Original Message -
> From: "Wido den Hollander"
> To:
Hi!
I wait about hour.
- Original Message -
From: "Wido den Hollander"
To: "Fyodor Ustinov" , ceph-users@lists.ceph.com
Sent: Friday, 24 August, 2018 09:52:23
Subject: Re: [ceph-users] ceph auto repair. What is wrong?
On 08/24/2018 06:11 AM, Fyodor Ustinov wrote:
> Hi!
>
> I have
Hi Gregory,
Quoting Gregory Farnum (gfar...@redhat.com):
> This is quite strange. Given that you have a log, I think what you want to
> do is find one request in the log, trace it through its lifetime, and see
> where the time is elapsed. You may find a bifurcation, where some
> categories of
Hi,
I don't know why but, I noticed in the ceph-volume-systemd.log
(above in bold), that there are 2 different lines corresponding to
the lvm-1 (normally associated to the osd.1) ?
One seems to have the correct id, while the other has a bad
one...and it's looks like he's trying to start
I installed a new Ceph cluster with Luminous, after a long time working
with Jewel. I created my RGW pools the same as always (pool create
default.rgw.buckets.data etc.), but they don't show up in ceph df with
Luminous. Has the command changed?
Since Luminous you don't need to create pools.
Le 23/08/2018 à 18:44, Alfredo Deza a écrit :
ceph-volume-systemd.log (extract)
[2018-08-20 11:26:26,386][systemd][INFO ] raw systemd input received:
lvm-6-ba351d69-5c48-418e-a377-4034f503af93
[2018-08-20 11:26:26,386][systemd][INFO ] raw systemd input received:
On 08/24/2018 01:57 PM, Buchberger, Carsten wrote:
Hi Konstantin,
sounds easy;-) If i apply the new rule to the existing pools there won't be
any osds to satisfy the requirements of the rule - because the osds are not in
the new root yet.
Isn't that a problem ?
Thank you
Your IO will
Hi,
osd_scrub_auto_repair still defaults to false and I was wondering how we
think about enabling this feature by default.
Would we say it's safe to enable this with BlueStore?
Wido
___
ceph-users mailing list
ceph-users@lists.ceph.com
On 08/24/2018 06:11 AM, Fyodor Ustinov wrote:
> Hi!
>
> I have fresh ceph cluster. 12 host and 3 osd on each host (one - hdd and two
> - ssd). Each host located in own rack.
>
> I make such crush configuration on fresh ceph installation:
>
>sudo ceph osd crush add-bucket R-26-3-1 rack
>
We recently upgrade to luminous (you can see the device-classes in the output).
So it should be possible to have one single root, no fake hosts and just use
the device-class.
We added some hosts/osds recently which back a new pools, so we also created a
new hierarchy and crush rules for those.
34 matches
Mail list logo