Re: [ceph-users] cephfs degraded on ceph luminous 12.2.2

2018-01-11 Thread Alessandro De Salvo
Hi,
took quite some time to recover the pgs, and indeed the problem with the
mds instances was due to the activating pgs. Once they were cleared the
fs went back to the original state.
I had to restart a few times some OSds though, in order to get all the
pgs activated, and I didn't hit the limits on the max pgs, but I'm close
to, so I have set them to 300 just to be safe (AFAIK it was the limit
set to prior releases of ceph, not sure why it was lowered to 200 now).
Thanks,

Alessandro

On Tue, 2018-01-09 at 09:01 +0100, Burkhard Linke wrote:
> Hi,
> 
> 
> On 01/08/2018 05:40 PM, Alessandro De Salvo wrote:
> > Thanks Lincoln,
> >
> > indeed, as I said the cluster is recovering, so there are pending ops:
> >
> >
> > pgs: 21.034% pgs not active
> >  1692310/24980804 objects degraded (6.774%)
> >  5612149/24980804 objects misplaced (22.466%)
> >  458 active+clean
> >  329 active+remapped+backfill_wait
> >  159 activating+remapped
> >  100 active+undersized+degraded+remapped+backfill_wait
> >  58  activating+undersized+degraded+remapped
> >  27  activating
> >  22  active+undersized+degraded+remapped+backfilling
> >  6   active+remapped+backfilling
> >  1   active+recovery_wait+degraded
> >
> >
> > If it's just a matter to wait for the system to complete the recovery 
> > it's fine, I'll deal with that, but I was wondendering if there is a 
> > more suble problem here.
> >
> > OK, I'll wait for the recovery to complete and see what happens, thanks.
> 
> The blocked MDS might be caused by the 'activating' PGs. Do you have a 
> warning about too much PGs per OSD? If that is the case, 
> activating/creating/peering/whatever on the affected OSDs is blocked, 
> which leads to blocked requests etc.
> 
> You can resolve this be increasing the number of allowed PGs per OSD 
> ('mon_max_pg_per_osd'). AFAIK it needs to be set for mon, mgr and osd 
> instances. There was also been some discussion about this setting on the 
> mailing list in the last weeks.
> 
> Regards,
> Burkhard
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] cephfs degraded on ceph luminous 12.2.2

2018-01-09 Thread Burkhard Linke

Hi,


On 01/08/2018 05:40 PM, Alessandro De Salvo wrote:

Thanks Lincoln,

indeed, as I said the cluster is recovering, so there are pending ops:


    pgs: 21.034% pgs not active
 1692310/24980804 objects degraded (6.774%)
 5612149/24980804 objects misplaced (22.466%)
 458 active+clean
 329 active+remapped+backfill_wait
 159 activating+remapped
 100 active+undersized+degraded+remapped+backfill_wait
 58  activating+undersized+degraded+remapped
 27  activating
 22  active+undersized+degraded+remapped+backfilling
 6   active+remapped+backfilling
 1   active+recovery_wait+degraded


If it's just a matter to wait for the system to complete the recovery 
it's fine, I'll deal with that, but I was wondendering if there is a 
more suble problem here.


OK, I'll wait for the recovery to complete and see what happens, thanks.


The blocked MDS might be caused by the 'activating' PGs. Do you have a 
warning about too much PGs per OSD? If that is the case, 
activating/creating/peering/whatever on the affected OSDs is blocked, 
which leads to blocked requests etc.


You can resolve this be increasing the number of allowed PGs per OSD 
('mon_max_pg_per_osd'). AFAIK it needs to be set for mon, mgr and osd 
instances. There was also been some discussion about this setting on the 
mailing list in the last weeks.


Regards,
Burkhard
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] cephfs degraded on ceph luminous 12.2.2

2018-01-08 Thread Sergey Malinin
You cannot force mds quit "replay" state for obvious reason of keeping data 
consistent. You might raise mds_beacon_grace to a somewhat reasonable value 
that would allow MDS to replay the journal without being marked laggy and 
eventually blacklisted.


From: ceph-users <ceph-users-boun...@lists.ceph.com> on behalf of Alessandro De 
Salvo <alessandro.desa...@roma1.infn.it>
Sent: Monday, January 8, 2018 7:40:59 PM
To: Lincoln Bryant; ceph-users@lists.ceph.com
Subject: Re: [ceph-users] cephfs degraded on ceph luminous 12.2.2

Thanks Lincoln,

indeed, as I said the cluster is recovering, so there are pending ops:


 pgs: 21.034% pgs not active
  1692310/24980804 objects degraded (6.774%)
  5612149/24980804 objects misplaced (22.466%)
  458 active+clean
  329 active+remapped+backfill_wait
  159 activating+remapped
  100 active+undersized+degraded+remapped+backfill_wait
  58  activating+undersized+degraded+remapped
  27  activating
  22  active+undersized+degraded+remapped+backfilling
  6   active+remapped+backfilling
  1   active+recovery_wait+degraded


If it's just a matter to wait for the system to complete the recovery
it's fine, I'll deal with that, but I was wondendering if there is a
more suble problem here.

OK, I'll wait for the recovery to complete and see what happens, thanks.

Cheers,


 Alessandro


Il 08/01/18 17:36, Lincoln Bryant ha scritto:
> Hi Alessandro,
>
> What is the state of your PGs? Inactive PGs have blocked CephFS
> recovery on our cluster before. I'd try to clear any blocked ops and
> see if the MDSes recover.
>
> --Lincoln
>
> On Mon, 2018-01-08 at 17:21 +0100, Alessandro De Salvo wrote:
>> Hi,
>>
>> I'm running on ceph luminous 12.2.2 and my cephfs suddenly degraded.
>>
>> I have 2 active mds instances and 1 standby. All the active
>> instances
>> are now in replay state and show the same error in the logs:
>>
>>
>>  mds1 
>>
>> 2018-01-08 16:04:15.765637 7fc2e92451c0  0 ceph version 12.2.2
>> (cf0baba3b47f9427c6c97e2144b094b7e5ba) luminous (stable),
>> process
>> (unknown), pid 164
>> starting mds.mds1 at -
>> 2018-01-08 16:04:15.785849 7fc2e92451c0  0 pidfile_write: ignore
>> empty
>> --pid-file
>> 2018-01-08 16:04:20.168178 7fc2e1ee1700  1 mds.mds1 handle_mds_map
>> standby
>> 2018-01-08 16:04:20.278424 7fc2e1ee1700  1 mds.1.20635 handle_mds_map
>> i
>> am now mds.1.20635
>> 2018-01-08 16:04:20.278432 7fc2e1ee1700  1 mds.1.20635
>> handle_mds_map
>> state change up:boot --> up:replay
>> 2018-01-08 16:04:20.278443 7fc2e1ee1700  1 mds.1.20635 replay_start
>> 2018-01-08 16:04:20.278449 7fc2e1ee1700  1 mds.1.20635  recovery set
>> is 0
>> 2018-01-08 16:04:20.278458 7fc2e1ee1700  1 mds.1.20635  waiting for
>> osdmap 21467 (which blacklists prior instance)
>>
>>
>>  mds2 
>>
>> 2018-01-08 16:04:16.870459 7fd8456201c0  0 ceph version 12.2.2
>> (cf0baba3b47f9427c6c97e2144b094b7e5ba) luminous (stable),
>> process
>> (unknown), pid 295
>> starting mds.mds2 at -
>> 2018-01-08 16:04:16.881616 7fd8456201c0  0 pidfile_write: ignore
>> empty
>> --pid-file
>> 2018-01-08 16:04:21.274543 7fd83e2bc700  1 mds.mds2 handle_mds_map
>> standby
>> 2018-01-08 16:04:21.314438 7fd83e2bc700  1 mds.0.20637 handle_mds_map
>> i
>> am now mds.0.20637
>> 2018-01-08 16:04:21.314459 7fd83e2bc700  1 mds.0.20637
>> handle_mds_map
>> state change up:boot --> up:replay
>> 2018-01-08 16:04:21.314479 7fd83e2bc700  1 mds.0.20637 replay_start
>> 2018-01-08 16:04:21.314492 7fd83e2bc700  1 mds.0.20637  recovery set
>> is 1
>> 2018-01-08 16:04:21.314517 7fd83e2bc700  1 mds.0.20637  waiting for
>> osdmap 21467 (which blacklists prior instance)
>> 2018-01-08 16:04:21.393307 7fd837aaf700  0 mds.0.cache creating
>> system
>> inode with ino:0x100
>> 2018-01-08 16:04:21.397246 7fd837aaf700  0 mds.0.cache creating
>> system
>> inode with ino:0x1
>>
>> The cluster is recovering as we are changing some of the osds, and
>> there
>> are a few slow/stuck requests, but I'm not sure if this is the cause,
>> as
>> there is apparently no data loss (until now).
>>
>> How can I force the MDSes to quit the replay state?
>>
>> Thanks for any help,
>>
>>
>>   Alessandro
>>
>>
>> ___
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] cephfs degraded on ceph luminous 12.2.2

2018-01-08 Thread Alessandro De Salvo

Thanks Lincoln,

indeed, as I said the cluster is recovering, so there are pending ops:


    pgs: 21.034% pgs not active
 1692310/24980804 objects degraded (6.774%)
 5612149/24980804 objects misplaced (22.466%)
 458 active+clean
 329 active+remapped+backfill_wait
 159 activating+remapped
 100 active+undersized+degraded+remapped+backfill_wait
 58  activating+undersized+degraded+remapped
 27  activating
 22  active+undersized+degraded+remapped+backfilling
 6   active+remapped+backfilling
 1   active+recovery_wait+degraded


If it's just a matter to wait for the system to complete the recovery 
it's fine, I'll deal with that, but I was wondendering if there is a 
more suble problem here.


OK, I'll wait for the recovery to complete and see what happens, thanks.

Cheers,


    Alessandro


Il 08/01/18 17:36, Lincoln Bryant ha scritto:

Hi Alessandro,

What is the state of your PGs? Inactive PGs have blocked CephFS
recovery on our cluster before. I'd try to clear any blocked ops and
see if the MDSes recover.

--Lincoln

On Mon, 2018-01-08 at 17:21 +0100, Alessandro De Salvo wrote:

Hi,

I'm running on ceph luminous 12.2.2 and my cephfs suddenly degraded.

I have 2 active mds instances and 1 standby. All the active
instances
are now in replay state and show the same error in the logs:


 mds1 

2018-01-08 16:04:15.765637 7fc2e92451c0  0 ceph version 12.2.2
(cf0baba3b47f9427c6c97e2144b094b7e5ba) luminous (stable),
process
(unknown), pid 164
starting mds.mds1 at -
2018-01-08 16:04:15.785849 7fc2e92451c0  0 pidfile_write: ignore
empty
--pid-file
2018-01-08 16:04:20.168178 7fc2e1ee1700  1 mds.mds1 handle_mds_map
standby
2018-01-08 16:04:20.278424 7fc2e1ee1700  1 mds.1.20635 handle_mds_map
i
am now mds.1.20635
2018-01-08 16:04:20.278432 7fc2e1ee1700  1 mds.1.20635
handle_mds_map
state change up:boot --> up:replay
2018-01-08 16:04:20.278443 7fc2e1ee1700  1 mds.1.20635 replay_start
2018-01-08 16:04:20.278449 7fc2e1ee1700  1 mds.1.20635  recovery set
is 0
2018-01-08 16:04:20.278458 7fc2e1ee1700  1 mds.1.20635  waiting for
osdmap 21467 (which blacklists prior instance)


 mds2 

2018-01-08 16:04:16.870459 7fd8456201c0  0 ceph version 12.2.2
(cf0baba3b47f9427c6c97e2144b094b7e5ba) luminous (stable),
process
(unknown), pid 295
starting mds.mds2 at -
2018-01-08 16:04:16.881616 7fd8456201c0  0 pidfile_write: ignore
empty
--pid-file
2018-01-08 16:04:21.274543 7fd83e2bc700  1 mds.mds2 handle_mds_map
standby
2018-01-08 16:04:21.314438 7fd83e2bc700  1 mds.0.20637 handle_mds_map
i
am now mds.0.20637
2018-01-08 16:04:21.314459 7fd83e2bc700  1 mds.0.20637
handle_mds_map
state change up:boot --> up:replay
2018-01-08 16:04:21.314479 7fd83e2bc700  1 mds.0.20637 replay_start
2018-01-08 16:04:21.314492 7fd83e2bc700  1 mds.0.20637  recovery set
is 1
2018-01-08 16:04:21.314517 7fd83e2bc700  1 mds.0.20637  waiting for
osdmap 21467 (which blacklists prior instance)
2018-01-08 16:04:21.393307 7fd837aaf700  0 mds.0.cache creating
system
inode with ino:0x100
2018-01-08 16:04:21.397246 7fd837aaf700  0 mds.0.cache creating
system
inode with ino:0x1

The cluster is recovering as we are changing some of the osds, and
there
are a few slow/stuck requests, but I'm not sure if this is the cause,
as
there is apparently no data loss (until now).

How can I force the MDSes to quit the replay state?

Thanks for any help,


  Alessandro


___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] cephfs degraded on ceph luminous 12.2.2

2018-01-08 Thread Lincoln Bryant
Hi Alessandro,

What is the state of your PGs? Inactive PGs have blocked CephFS
recovery on our cluster before. I'd try to clear any blocked ops and
see if the MDSes recover.

--Lincoln

On Mon, 2018-01-08 at 17:21 +0100, Alessandro De Salvo wrote:
> Hi,
> 
> I'm running on ceph luminous 12.2.2 and my cephfs suddenly degraded.
> 
> I have 2 active mds instances and 1 standby. All the active
> instances 
> are now in replay state and show the same error in the logs:
> 
> 
>  mds1 
> 
> 2018-01-08 16:04:15.765637 7fc2e92451c0  0 ceph version 12.2.2 
> (cf0baba3b47f9427c6c97e2144b094b7e5ba) luminous (stable),
> process 
> (unknown), pid 164
> starting mds.mds1 at -
> 2018-01-08 16:04:15.785849 7fc2e92451c0  0 pidfile_write: ignore
> empty 
> --pid-file
> 2018-01-08 16:04:20.168178 7fc2e1ee1700  1 mds.mds1 handle_mds_map
> standby
> 2018-01-08 16:04:20.278424 7fc2e1ee1700  1 mds.1.20635 handle_mds_map
> i 
> am now mds.1.20635
> 2018-01-08 16:04:20.278432 7fc2e1ee1700  1 mds.1.20635
> handle_mds_map 
> state change up:boot --> up:replay
> 2018-01-08 16:04:20.278443 7fc2e1ee1700  1 mds.1.20635 replay_start
> 2018-01-08 16:04:20.278449 7fc2e1ee1700  1 mds.1.20635  recovery set
> is 0
> 2018-01-08 16:04:20.278458 7fc2e1ee1700  1 mds.1.20635  waiting for 
> osdmap 21467 (which blacklists prior instance)
> 
> 
>  mds2 
> 
> 2018-01-08 16:04:16.870459 7fd8456201c0  0 ceph version 12.2.2 
> (cf0baba3b47f9427c6c97e2144b094b7e5ba) luminous (stable),
> process 
> (unknown), pid 295
> starting mds.mds2 at -
> 2018-01-08 16:04:16.881616 7fd8456201c0  0 pidfile_write: ignore
> empty 
> --pid-file
> 2018-01-08 16:04:21.274543 7fd83e2bc700  1 mds.mds2 handle_mds_map
> standby
> 2018-01-08 16:04:21.314438 7fd83e2bc700  1 mds.0.20637 handle_mds_map
> i 
> am now mds.0.20637
> 2018-01-08 16:04:21.314459 7fd83e2bc700  1 mds.0.20637
> handle_mds_map 
> state change up:boot --> up:replay
> 2018-01-08 16:04:21.314479 7fd83e2bc700  1 mds.0.20637 replay_start
> 2018-01-08 16:04:21.314492 7fd83e2bc700  1 mds.0.20637  recovery set
> is 1
> 2018-01-08 16:04:21.314517 7fd83e2bc700  1 mds.0.20637  waiting for 
> osdmap 21467 (which blacklists prior instance)
> 2018-01-08 16:04:21.393307 7fd837aaf700  0 mds.0.cache creating
> system 
> inode with ino:0x100
> 2018-01-08 16:04:21.397246 7fd837aaf700  0 mds.0.cache creating
> system 
> inode with ino:0x1
> 
> The cluster is recovering as we are changing some of the osds, and
> there 
> are a few slow/stuck requests, but I'm not sure if this is the cause,
> as 
> there is apparently no data loss (until now).
> 
> How can I force the MDSes to quit the replay state?
> 
> Thanks for any help,
> 
> 
>  Alessandro
> 
> 
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com