I got swamped today.  I should be able to look tomorrow.  Sorry!
-Sam

On Mon, Aug 12, 2013 at 9:39 PM, Stefan Priebe - Profihost AG
<s.pri...@profihost.ag> wrote:
> Did you take a look?
>
> Stefan
>
> Am 11.08.2013 um 05:50 schrieb Samuel Just <sam.j...@inktank.com>:
>
>> Great!  I'll take a look on Monday.
>> -Sam
>>
>> On Sat, Aug 10, 2013 at 12:08 PM, Stefan Priebe <s.pri...@profihost.ag> 
>> wrote:
>>> Hi Samual,
>>>
>>> Am 09.08.2013 23:44, schrieb Samuel Just:
>>>
>>>> I think Stefan's problem is probably distinct from Mike's.
>>>>
>>>> Stefan: Can you reproduce the problem with
>>>>
>>>> debug osd = 20
>>>> debug filestore = 20
>>>> debug ms = 1
>>>> debug optracker = 20
>>>>
>>>> on a few osds (including the restarted osd), and upload those osd logs
>>>> along with the ceph.log from before killing the osd until after the
>>>> cluster becomes clean again?
>>>
>>>
>>> done - you'll find the logs at cephdrop folder:
>>> slow_requests_recovering_cuttlefish
>>>
>>> osd.52 was the one recovering
>>>
>>> Thanks!
>>>
>>> Greets,
>>> Stefan
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majord...@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to