Re: [ceph-users] osd_snap_trim_sleep keeps locks PG during sleep?

2017-02-21 Thread Samuel Just
@lists.ceph.com] *On Behalf > Of *Samuel Just > *Sent:* 09 February 2017 19:22 > *To:* Nick Fisk > *Cc:* ceph-users@lists.ceph.com > > *Subject:* Re: [ceph-users] osd_snap_trim_sleep keeps locks PG during > sleep? > > > > Ok, https://github.com/athanatos/c

Re: [ceph-users] osd_snap_trim_sleep keeps locks PG during sleep?

2017-02-09 Thread Samuel Just
Ok, https://github.com/athanatos/ceph/tree/wip-snap-trim-sleep (based on master) passed a rados suite. It adds a configurable limit to the number of pgs which can be trimming on any OSD (default: 2). PGs trimming will be in snaptrim state, PGs waiting to trim will be in snaptrim_wait state. I su

Re: [ceph-users] osd_snap_trim_sleep keeps locks PG during sleep?

2017-02-03 Thread Samuel Just
intended recipient of this message or received it > erroneously, please notify the sender and delete it, together with any > attachments, and be advised that any dissemination or copying of this > message is prohibited. > > -- > > ---

Re: [ceph-users] osd_snap_trim_sleep keeps locks PG during sleep?

2017-02-03 Thread Samuel Just
ete it, together with any > attachments, and be advised that any dissemination or copying of this > message is prohibited. > > -- > > -- > *From:* Samuel Just [sj...@redhat.com] > *Sent:* Thursday, January 26, 2017 1:14 P

Re: [ceph-users] osd_snap_trim_sleep keeps locks PG during sleep?

2017-01-26 Thread Samuel Just
am On Fri, Jan 20, 2017 at 2:05 PM, Nick Fisk wrote: > Hi Sam, > > > > I have a test cluster, albeit small. I’m happy to run tests + graph > results with a wip branch and work out reasonable settings…etc > > > > *From:* Samuel Just [mailto:sj...@redhat.com] > *S

Re: [ceph-users] osd_snap_trim_sleep keeps locks PG during sleep?

2017-01-19 Thread Samuel Just
> -- > > -- > *From:* ceph-users [ceph-users-boun...@lists.ceph.com] on behalf of David > Turner [david.tur...@storagecraft.com] > *Sent:* Thursday, January 19, 2017 3:25 PM > *To:* Samuel Just; Nick Fisk > > *Cc:* cep

Re: [ceph-users] osd_snap_trim_sleep keeps locks PG during sleep?

2017-01-19 Thread Samuel Just
application can priority tag individual > IO's for CFQ to handle, instead of the current limitation of priority per > thread, I realise this is probably very very hard or impossible. But it would > allow Ceph to control IO queue's right down to the disk. > >> -Original M

Re: [ceph-users] osd_snap_trim_sleep keeps locks PG during sleep?

2017-01-19 Thread Samuel Just
fixing a bug with that. > > Nick > >> -Original Message- >> From: Samuel Just [mailto:sj...@redhat.com] >> Sent: 19 January 2017 15:47 >> To: Dan van der Ster >> Cc: Nick Fisk ; ceph-users >> Subject: Re: [ceph-users] osd_snap_trim_sleep k

Re: [ceph-users] osd_snap_trim_sleep keeps locks PG during sleep?

2017-01-19 Thread Samuel Just
Snaptrimming is now in the main op threadpool along with scrub, recovery, and client IO. I don't think it's a good idea to use any of the _sleep configs anymore -- the intention is that by setting the priority low, they won't actually be scheduled much. -Sam On Thu, Jan 19, 2017 at 5:40 AM, Dan v

Re: [ceph-users] Why would "osd marked itself down" will not recognised?

2017-01-12 Thread Samuel Just
That would work. -Sam On Thu, Jan 12, 2017 at 1:40 PM, Gregory Farnum wrote: > On Thu, Jan 12, 2017 at 1:37 PM, Samuel Just wrote: >> Oh, this is basically working as intended. What happened is that the >> mon died before the pending map was actually committed. The OSD has a

Re: [ceph-users] Why would "osd marked itself down" will not recognised?

2017-01-12 Thread Samuel Just
r. I will look tomorrow > at the output again and create an ticket. > > > Thanks > > > Udo > > > On 12.01.2017 20:02, Samuel Just wrote: >> How long did it take for the cluster to recover? >> -Sam >> >> On Thu, Jan 12, 2017 at 10:54 AM, Gregor

Re: [ceph-users] Why would "osd marked itself down" will not recognised?

2017-01-12 Thread Samuel Just
How long did it take for the cluster to recover? -Sam On Thu, Jan 12, 2017 at 10:54 AM, Gregory Farnum wrote: > On Thu, Jan 12, 2017 at 2:03 AM, wrote: >> Hi all, >> I had just reboot all 3 nodes (one after one) of an small Proxmox-VE >> ceph-cluster. All nodes are mons and have two OSDs. >> Du

Re: [ceph-users] Any librados C API users out there?

2017-01-11 Thread Samuel Just
Jason: librbd itself uses the librados C++ api though, right? -Sam On Wed, Jan 11, 2017 at 9:37 AM, Jason Dillaman wrote: > +1 > > I'd be happy to tweak the internals of librbd to support pass-through > of C buffers all the way to librados. librbd clients like QEMU use the > C API and this curren

Re: [ceph-users] High CPU usage by ceph-mgr on idle Ceph cluster

2017-01-10 Thread Samuel Just
c/ceph/ > > > Bryan > > On 1/10/17, 10:51 AM, "Samuel Just" wrote: > >>Can you push that branch somewhere? I don't have a v11.1.1 or that sha1. >>-Sam >> >>On Tue, Jan 10, 2017 at 9:41 AM, Stillwell, Bryan J >> wrote: >>> This is f

Re: [ceph-users] High CPU usage by ceph-mgr on idle Ceph cluster

2017-01-10 Thread Samuel Just
Can you push that branch somewhere? I don't have a v11.1.1 or that sha1. -Sam On Tue, Jan 10, 2017 at 9:41 AM, Stillwell, Bryan J wrote: > This is from: > > ceph version 11.1.1 (87597971b371d7f497d7eabad3545d72d18dd755) > > On 1/10/17, 10:23 AM, "Samuel Just" wro

Re: [ceph-users] High CPU usage by ceph-mgr on idle Ceph cluster

2017-01-10 Thread Samuel Just
What ceph sha1 is that? Does it include 6c3d015c6854a12cda40673848813d968ff6afae which fixed the messenger spin? -Sam On Tue, Jan 10, 2017 at 9:00 AM, Stillwell, Bryan J wrote: > On 1/10/17, 5:35 AM, "John Spray" wrote: > >>On Mon, Jan 9, 2017 at 11:46 PM, Stillwell, Bryan J >> wrote: >>> Last

Re: [ceph-users] PGs stuck active+remapped and osds lose data?!

2017-01-10 Thread Samuel Just
Shinobu isn't correct, you have 9/9 osds up and running. up does not equal acting because crush is having trouble fulfilling the weights in your crushmap and the acting set is being padded out with an extra osd which happens to have the data to keep you up to the right number of replicas. Please

Re: [ceph-users] pg stuck in peering while power failure

2017-01-10 Thread Samuel Just
{ "name": "Started\/Primary\/Peering", "enter_time": "2017-01-10 13:43:34.933074", "past_intervals": [ { "first": 75858, "last": 75860, "maybe_went_rw": 1, "up

Re: [ceph-users] Kraken 11.x feedback

2016-12-09 Thread Samuel Just
Is there a particular reason you are sticking to the versions with shorter support periods? -Sam On Fri, Dec 9, 2016 at 11:38 AM, Ben Hines wrote: > Anyone have any good / bad experiences with Kraken? I haven't seen much > discussion of it. Particularly from the RGW front. > > I'm still on Infern

Re: [ceph-users] 10.2.4 Jewel released

2016-12-07 Thread Samuel Just
Actually, Greg and Sage are working up other branches, nvm. -Sam On Wed, Dec 7, 2016 at 2:52 PM, Samuel Just wrote: > I just pushed a branch wip-14120-10.2.4 with a possible fix. > > https://github.com/ceph/ceph/pull/12349/ is a fix for a known bug > which didn't quite make it

Re: [ceph-users] 10.2.4 Jewel released

2016-12-07 Thread Samuel Just
I just pushed a branch wip-14120-10.2.4 with a possible fix. https://github.com/ceph/ceph/pull/12349/ is a fix for a known bug which didn't quite make it into 10.2.4, it's possible that 165e5abdbf6311974d4001e43982b83d06f9e0cc which did made the bug much more likely to happen. wip-14120-10.2.4 ha

Re: [ceph-users] How are replicas spread in default crush configuration?

2016-11-23 Thread Samuel Just
On Wed, Nov 23, 2016 at 4:11 PM, Chris Taylor wrote: > Kevin, > > After changing the pool size to 3, make sure the min_size is set to 1 to > allow 2 of the 3 hosts to be offline. If you do this, the flip side is that while in that configuration losing that single host will render your data unreco

Re: [ceph-users] how possible is that ceph cluster crash

2016-11-23 Thread Samuel Just
I don't mind trying to get this done if its felt to be worthwhile. > > Nick > >> -Original Message- >> From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of >> Samuel Just >> Sent: 19 November 2016 00:31 >> To: Nick Fisk >

Re: [ceph-users] how possible is that ceph cluster crash

2016-11-18 Thread Samuel Just
t; >> This is like your mother telling not to cross the road when you were 4 >> years of age but not telling you it was because you could be flattened >> by a car :) >> >> Can you expand on your answer? If you are in a DC with AB power, >> redundant UPS, dual feed

Re: [ceph-users] how possible is that ceph cluster crash

2016-11-18 Thread Samuel Just
Never *ever* use nobarrier with ceph under *any* circumstances. I cannot stress this enough. -Sam On Fri, Nov 18, 2016 at 10:39 AM, Craig Chi wrote: > Hi Nick and other Cephers, > > Thanks for your reply. > >> 2) Config Errors >> This can be an easy one to say you are safe from. But I would say

Re: [ceph-users] After OSD Flap - FAILED assert(oi.version == i->first)

2016-11-17 Thread Samuel Just
Puzzling, added a question to the ticket. -Sam On Thu, Nov 17, 2016 at 4:32 AM, Nick Fisk wrote: > Hi Sam, > > I've updated the ticket with logs from the wip run. > > Nick > >> -Original Message- >> From: Samuel Just [mailto:sj...@redhat.com] >>

Re: [ceph-users] After OSD Flap - FAILED assert(oi.version == i->first)

2016-11-15 Thread Samuel Just
http://tracker.ceph.com/issues/17916 I just pushed a branch wip-17916-jewel based on v10.2.3 with some additional debugging. Once it builds, would you be able to start the afflicted osds with that version of ceph-osd and debug osd = 20 debug ms = 1 debug filestore = 20 and get me the log? -Sam

Re: [ceph-users] cache tier on rgw index pool

2016-09-21 Thread Samuel Just
I seriously doubt that it's ever going to be a winning strategy to let rgw index objects go to a cold tier. Some practical problems: 1) We don't track omap size (the leveldb entries for an object) because it would turn writes into rmw's -- so they always show up as 0 size. Thus, the target_max_by

Re: [ceph-users] Same pg scrubbed over and over (Jewel)

2016-09-21 Thread Samuel Just
Ah, same question then. If we can get logging on the primary for one of those pgs, it should be fairly obvious. -Sam On Wed, Sep 21, 2016 at 4:08 AM, Pavan Rallabhandi wrote: > We find this as well in our fresh built Jewel clusters, and seems to happen > only with a handful of PGs from couple o

Re: [ceph-users] crash of osd using cephfs jewel 10.2.2, and corruption

2016-09-21 Thread Samuel Just
Looks like the OSD didn't like an error return it got from the underlying fs. Can you reproduce with debug filestore = 20 debug osd = 20 debug ms = 1 on the osd and post the whole log? -Sam On Wed, Sep 21, 2016 at 12:10 AM, Peter Maloney wrote: > Hi, > > I created a one disk osd with data and

Re: [ceph-users] Same pg scrubbed over and over (Jewel)

2016-09-21 Thread Samuel Just
Can you reproduce with logging on the primary for that pg? debug osd = 20 debug filestore = 20 debug ms = 1 Since restarting the osd may be a workaround, can you inject the debug values without restarting the daemon? -Sam On Wed, Sep 21, 2016 at 2:44 AM, Tobias Böhm wrote: > Hi, > > there is an

Re: [ceph-users] OSD daemon randomly stops

2016-09-02 Thread Samuel Just
Probably an EIO. You can reproduce with debug filestore = 20 to confirm. -Sam On Fri, Sep 2, 2016 at 10:18 AM, Reed Dier wrote: > OSD has randomly stopped for some reason. Lots of recovery processes > currently running on the ceph cluster. OSD log with assert below: > > -14> 2016-09-02 11:32:38.

Re: [ceph-users] Strange copy errors in osd log

2016-09-01 Thread Samuel Just
If it's bluestore, this is pretty likely to be a bluestore bug. If you are interested in experimenting with bluestore, you probably want to watch developements on the master branch, it's undergoing a bunch of changes right now. -Sam On Thu, Sep 1, 2016 at 1:54 PM, Виталий Филиппов wrote: > Hi! I

Re: [ceph-users] Backfilling pgs not making progress

2016-08-11 Thread Samuel Just
use peering and >> backfill to restart) >> 5) link all logs to the bug >> >> Thanks! >> -Sam >> >> On Tue, Jul 26, 2016 at 9:11 AM, Samuel Just wrote: >> > Hmm, nvm, it's not an lfn object anyway. >> > -Sam >> > >> &g

Re: [ceph-users] Backfilling pgs not making progress

2016-08-10 Thread Samuel Just
, 2016 at 9:11 AM, Samuel Just wrote: > Hmm, nvm, it's not an lfn object anyway. > -Sam > > On Tue, Jul 26, 2016 at 7:07 AM, Brian Felton wrote: >> If I search on osd.580, I find >> default.421929.15\uTEPP\s84316222-6ddd-4ac9-8283-6fa1cdcf9b88\sbackups\s2016063009

Re: [ceph-users] [Ceph-community] Noobie question about OSD fail

2016-07-27 Thread Samuel Just
osd min down reports = 2 Set that to 1? -Sam On Wed, Jul 27, 2016 at 10:24 AM, Patrick McGarry wrote: > Moving this to ceph-user. > > > On Wed, Jul 27, 2016 at 8:36 AM, Kostya Velychkovsky > wrote: >> Hello. I have test CEPH cluster with 5 nodes: 3 MON and 2 OSD >> >> This is my ceph.conf >> >

Re: [ceph-users] Listing objects in a specified placement group / OSD

2016-07-27 Thread Samuel Just
Well, it's kind of deliberately obfuscated because PGs aren't a librados-level abstraction. Why do you want to list the objects in a PG? -Sam On Wed, Jul 27, 2016 at 8:10 AM, David Blundell wrote: > Hi, > > > > I wasn’t sure if this is a ceph-users or ceph-devel question as it’s about > the API

Re: [ceph-users] How to get Active set of OSD Map in serial order of osd index

2016-07-27 Thread Samuel Just
e plugin for RDP (or RAID-DP) kind of > code. > > Thanks, > Syed > > On Wed, Jul 27, 2016 at 4:12 AM, Samuel Just wrote: >> >> Why do you want them in serial increasing order? >> -Sam >> >> On Tue, Jul 26, 2016 at 2:43 PM, Samuel Just wrote:

Re: [ceph-users] How to get Active set of OSD Map in serial order of osd index

2016-07-26 Thread Samuel Just
Why do you want them in serial increasing order? -Sam On Tue, Jul 26, 2016 at 2:43 PM, Samuel Just wrote: > How would such a code work if there were more than 24 osds? > -Sam > > On Tue, Jul 26, 2016 at 2:37 PM, Syed Hussain wrote: > >> Hi, >> >> I'm wor

Re: [ceph-users] How to get Active set of OSD Map in serial order of osd index

2016-07-26 Thread Samuel Just
How would such a code work if there were more than 24 osds? -Sam On Tue, Jul 26, 2016 at 2:37 PM, Syed Hussain wrote: > Hi, > > I'm working to develop an Erasure Code plugin (variation of ISA) that have > typical requirement that the active set of the Erasure Coded pool in serial > order. > For

Re: [ceph-users] Backfilling pgs not making progress

2016-07-26 Thread Samuel Just
about a dangling link, please > point me in the right direction. > > Brian > > On Tue, Jul 26, 2016 at 8:59 AM, Samuel Just wrote: >> >> Did you also confirm that the backfill target does not have any of >> those dangling links? I'd be looking for

Re: [ceph-users] Backfilling pgs not making progress

2016-07-26 Thread Samuel Just
e reviewing them with my team tomorrow morning to see if we can find > anything. Thanks for your assistance. > > Brian > > On Mon, Jul 25, 2016 at 3:33 PM, Samuel Just wrote: >> >> The next thing I'd want is for you to reproduce with >> >> debug osd = 20

Re: [ceph-users] Backfilling pgs not making progress

2016-07-25 Thread Samuel Just
the special ceph long filename handling. > > Also, we are not using XFS on our OSDs; we are using ZFS instead. > > If I'm misunderstanding the issue linked above and the corresponding thread, > please let me know. > > Brian > > > On Mon, Jul 25, 2016 at 1:32 PM, Samu

Re: [ceph-users] Backfilling pgs not making progress

2016-07-25 Thread Samuel Just
The next thing I'd want is for you to reproduce with debug osd = 20 debug filestore = 20 debug ms = 1 and post the file somewhere. -Sam On Mon, Jul 25, 2016 at 1:33 PM, Samuel Just wrote: > If you don't have the orphaned file link, it's not the same bug. > -Sam > >

Re: [ceph-users] Backfilling pgs not making progress

2016-07-25 Thread Samuel Just
You may have hit http://tracker.ceph.com/issues/14766. There was a thread on the list a while back about diagnosing and fixing it. -Sam On Mon, Jul 25, 2016 at 10:45 AM, Brian Felton wrote: > Greetings, > > Problem: After removing (out + crush remove + auth del + osd rm) three osds > on a single

Re: [ceph-users] ERROR: unable to open OSD superblock on /var/lib/ceph/osd/ceph-1: (13) Permission denied

2016-07-06 Thread Samuel Just
Try strace. -Sam On Wed, Jul 6, 2016 at 7:53 AM, RJ Nowling wrote: > Hi all, > > I'm trying to use the ceph-ansible playbooks to deploy onto baremetal. I'm > currently testing with the approach that uses an existing filesystem rather > than giving raw access to the disks. > > The OSDs are failin

Re: [ceph-users] slow request, waiting for rw locks / subops from osd doing deep scrub of pg in rgw.buckets.index

2016-06-21 Thread Samuel Just
.rgw.bucket.index.pool is the pool with rgw's index objects, right? The actual on-disk directory for one of those pgs would contain only empty files -- the actual index data is stored in the osd's leveldb instance. I suspect your index objects are very large (because the buckets contain many objec

Re: [ceph-users] Fio randwrite does not work on Centos 7.2 VM

2016-06-15 Thread Samuel Just
I think you hit the os process fd limit. You need to adjust it. -Sam On Wed, Jun 15, 2016 at 2:07 PM, Mansour Shafaei Moghaddam wrote: > It fails at "FileStore.cc: 2761". Here is a more complete log: > > -9> 2016-06-15 10:55:13.205014 7fa2dcd85700 -1 dump_open_fds unable to > open /proc/self

Re: [ceph-users] strange cache tier behaviour with cephfs

2016-06-13 Thread Samuel Just
I'd have to look more closely, but these days promotion is probabilistic and throttled. During each read of those objects, it will tend to promote a few more of them depending on how many promotions are in progress and how hot it thinks a particular object is. The lack of a speed up is a bummer,

Re: [ceph-users] Upgrade Errors.

2016-06-06 Thread Samuel Just
r > root as they were created while the older ceph processes were still running. > > Changing that after the fact was still a necessity and I will make sure that > those are also properly changed. > > On Mon, Jun 6, 2016 at 12:12 PM Samuel Just wrote: >> >> Oh, what wa

Re: [ceph-users] Upgrade Errors.

2016-06-06 Thread Samuel Just
Oh, what was the problem (for posterity)? -Sam On Mon, Jun 6, 2016 at 12:11 PM, Tu Holmes wrote: > It totally did and I see what the problem is. > > Thanks for your input. I truly appreciate it. > > > On Mon, Jun 6, 2016 at 12:01 PM Samuel Just wrote: >> >> If you

Re: [ceph-users] Upgrade Errors.

2016-06-06 Thread Samuel Just
If you reproduce with debug osd = 20 debug filestore = 20 debug ms = 1 that might make it clearer what is going on. -Sam On Mon, Jun 6, 2016 at 11:53 AM, Tu Holmes wrote: > Hey cephers. I have been following the upgrade documents and I have done > everything regarding upgrading the client to th

Re: [ceph-users] jewel upgrade and sortbitwise

2016-06-03 Thread Samuel Just
Sorry, I should have been more clear. The bug actually is due to a difference in an on disk encoding from hammer. An infernalis cluster would never had had such encodings and is fine. -Sam On Jun 3, 2016 6:53 AM, "Francois Lafont" wrote: > Hi, > > On 03/06/2016 05:39

[ceph-users] jewel upgrade and sortbitwise

2016-06-02 Thread Samuel Just
Due to http://tracker.ceph.com/issues/16113, it would be best to avoid setting the sortbitwise flag on jewel clusters upgraded from previous versions until we get a point release out with a fix. The symptom is that setting the sortbitwise flag on a jewel cluster upgraded from a previous version ca

Re: [ceph-users] Crashing OSDs (suicide timeout, following a single pool)

2016-06-02 Thread Samuel Just
I suspect the problem is that ReplicatedBackend::build_push_op assumes that osd_recovery_max_chunk (defaults to 8MB) of omap entries is about the same amount of work to get as 8MB of normal object data. The fix would be to add another config osd_recovery_max_omap_entries_per_chunk with a sane defa

Re: [ceph-users] OSD Restart results in "unfound objects"

2016-06-01 Thread Samuel Just
as you noted in the issue. > > The cluster runs jewel 10.2.1 and was created a long time ago, I think it was > giant. > > Thanks again! > > Uwe > >> Am 02.06.2016 um 00:19 schrieb Samuel Just : >> >> http://tracker.ceph.com/issues/16113 >> >> I t

Re: [ceph-users] OSD Restart results in "unfound objects"

2016-06-01 Thread Samuel Just
the upgrade (despite the well know systemd), > and after that the cluster didn't show any suspicious behavior. > > > --- > Diego Castro / The CloudFather > GetupCloud.com - Eliminamos a Gravidade > > 2016-06-01 18:57 GMT-03:00 Samuel Just : >> >> Was this cluster upg

Re: [ceph-users] OSD Restart results in "unfound objects"

2016-06-01 Thread Samuel Just
t unset sortbitwise flag. > > > --- > Diego Castro / The CloudFather > GetupCloud.com - Eliminamos a Gravidade > > 2016-06-01 13:39 GMT-03:00 Samuel Just : >> >> Can either of you reproduce with logs? That would make it a lot >> easier to track down if it's a bug. I

Re: [ceph-users] OSD Restart results in "unfound objects"

2016-06-01 Thread Samuel Just
Can either of you reproduce with logs? That would make it a lot easier to track down if it's a bug. I'd want debug osd = 20 debug ms = 1 debug filestore = 20 On all of the osds for a particular pg from when it is clean until it develops an unfound object. -Sam On Wed, Jun 1, 2016 at 5:36 AM, D

Re: [ceph-users] unfound objects - why and how to recover ? (bonus : jewel logs)

2016-05-27 Thread Samuel Just
Well, it's not supposed to do that if the backing storage is working properly. If the filesystem/disk controller/disk combination is not respecting barriers (or otherwise can lose committed data in a power failure) in your configuration, a power failure could cause a node to go backwards in time -

Re: [ceph-users] help removing an rbd image?

2016-05-26 Thread Samuel Just
. Suggestions? > > > Kevan > > > > On 5/24/16, 4:11 PM, "Kevan Rehm" wrote: > >>Okay, will do. If the problem goes away with filestore, I'll switch back >>to bluestore again and re-duplicate the problem. In that case, are there >>p

Re: [ceph-users] help removing an rbd image?

2016-05-24 Thread Samuel Just
My money is on bluestore. If you can try to reproduce on filestore, that would rapidly narrow it down. -Sam On Tue, May 24, 2016 at 1:53 PM, Kevan Rehm wrote: > Nope, not using tiering. > > Also, this is my second attempt, this is repeatable for me, I'm trying to > duplicate a previous occurrenc

Re: [ceph-users] ceph hang on pg list_unfound

2016-05-19 Thread Samuel Just
Restart osd.1 with debugging enabled debug osd = 20 debug filestore = 20 debug ms = 1 Then, run list_unfound once the pg is back in active+recovering. If it still hangs, post osd.1's log to the list along with the output of ceph osd dump and ceph pg dump. -Sam On Wed, May 18, 2016 at 6:20 PM, D

Re: [ceph-users] PG stuck incomplete after power failure.

2016-05-17 Thread Samuel Just
Try restarting the primary osd for that pg with osd_find_best_info_ignore_history_les set to true (don't leave it set long term). -Sam On Tue, May 17, 2016 at 7:50 AM, Hein-Pieter van Braam wrote: > Hello, > > Today we had a power failure in a rack housing our OSD servers. We had > 7 of our 30 to

Re: [ceph-users] Erasure Coded Pools Cause Irretrievable Objects and Possible Corruption

2016-05-05 Thread Samuel Just
Can you reproduce with debug ms = 1 debug objecter = 20 on the radosgw side? -Sam On Thu, May 5, 2016 at 8:28 AM, Brian Felton wrote: > Greetings, > > We are running a number of Ceph clusters in production to provide object > storage services. We have stumbled upon an issue where objects of ce

Re: [ceph-users] OSD Crashes

2016-04-29 Thread Samuel Just
egards > Somnath > > -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Garg, Pankaj > Sent: Friday, April 29, 2016 8:59 AM > To: Samuel Just > Cc: ceph-users@lists.ceph.com > Subject: Re: [ceph-users] OSD Crashes >

Re: [ceph-users] OSD Crashes

2016-04-29 Thread Samuel Just
Your fs is throwing an EIO on open. -Sam On Fri, Apr 29, 2016 at 8:54 AM, Garg, Pankaj wrote: > Hi, > > I had a fully functional Ceph cluster with 3 x86 Nodes and 3 ARM64 nodes, > each with 12 HDD Drives and 2SSD Drives. All these were initially running > Hammer, and then were successfully update

Re: [ceph-users] help troubleshooting some osd communication problems

2016-04-28 Thread Samuel Just
I'd guess that to make any progress we'll need debug ms = 20 on both sides of the connection when a message is lost. -Sam On Thu, Apr 28, 2016 at 2:38 PM, Mike Lovell wrote: > there was a problem on one of the clusters i manage a couple weeks ago where > pairs of OSDs would wait indefinitely on s

Re: [ceph-users] CEPH All OSD got segmentation fault after CRUSH edit

2016-04-26 Thread Samuel Just
I think? Probably worth reproducing on a vstart cluster to validate the fix. Didn't we introduce something in the mon to validate new crushmaps? Hammer maybe? -Sam On Tue, Apr 26, 2016 at 8:09 AM, Wido den Hollander wrote: > >> Op 26 april 2016 om 16:58 schreef Samuel Just : &g

Re: [ceph-users] CEPH All OSD got segmentation fault after CRUSH edit

2016-04-26 Thread Samuel Just
Can you attach the OSDMap (ceph osd getmap -o )? -Sam On Tue, Apr 26, 2016 at 2:07 AM, Henrik Svensson wrote: > Hi! > > We got a three node CEPH cluster with 10 OSD each. > > We bought 3 new machines with additional 30 disks that should reside in > another location. > Before adding these machine

Re: [ceph-users] Deprecating ext4 support

2016-04-14 Thread Samuel Just
It doesn't seem like it would be wise to run such systems on top of rbd. -Sam On Thu, Apr 14, 2016 at 11:05 AM, Jianjian Huo wrote: > On Wed, Apr 13, 2016 at 6:06 AM, Sage Weil wrote: >> On Tue, 12 Apr 2016, Jan Schermer wrote: >>> Who needs to have exactly the same data in two separate objects

Re: [ceph-users] ceph_assert_fail after upgrade from hammer to infernalis

2016-04-07 Thread Samuel Just
Well, that bug is a completely different backtrace, so it's probably safe to say that it's not related. It looks like you are missing maps, which suggests you have inconsistent osd filesystems. Did you powercycle this node with barriers disabled in xfs? Posting a log from startup to crash would

Re: [ceph-users] Scrubbing a lot

2016-03-29 Thread Samuel Just
f304ebf8bf1166aba6) > > If I run the command without root or sudo the command failed with a > Permission denied > > German > > 2016-03-29 17:24 GMT-03:00 Samuel Just : >> >> Or you needed to run it as root? >> -Sam >> >> On Tue, Mar 29, 2016 at 1:24

Re: [ceph-users] Scrubbing a lot

2016-03-29 Thread Samuel Just
Or you needed to run it as root? -Sam On Tue, Mar 29, 2016 at 1:24 PM, Samuel Just wrote: > Sounds like a version/compatibility thing. Are your rbd clients really old? > -Sam > > On Tue, Mar 29, 2016 at 1:19 PM, German Anders wrote: >> I've just upgrade to jewel, and

Re: [ceph-users] Scrubbing a lot

2016-03-29 Thread Samuel Just
"2", >> "filestore_queue_max_ops": "50", >> "filestore_split_multiple": "2", >> "fsid": "----0000", >> "internal_safe_to_sta

Re: [ceph-users] Scrubbing a lot

2016-03-29 Thread Samuel Just
That seems to be scrubbing pretty often. Can you attach a config diff from osd.4 (ceph daemon osd.4 config diff)? -Sam On Tue, Mar 29, 2016 at 9:30 AM, German Anders wrote: > Hi All, > > I've maybe a simple question, I've setup a new cluster with Infernalis > release, there's no IO going on at t

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-19 Thread Samuel Just
Ok, like I said, most files with _long at the end are *not orphaned*. The generation number also is *not* an indication of whether the file is orphaned -- some of the orphaned files will have as the generation number and others won't. For each long filename object in a pg you woul

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-19 Thread Samuel Just
seen and only about 230 have the user.cephos.lfn3 > file attribute. The files will have other attributes, just not > user.cephos.lfn3. > > Regards, > Jeff > > > On Wed, Mar 16, 2016 at 3:53 PM, Samuel Just wrote: >> >> Ok, like I said, most files with _long a

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-19 Thread Samuel Just
X\u7\uGAGTGG.tar.gz.2~\u1r\uFGidmpEP8GRsJkNLfAh9CokxkL_fa202ec9b4b3b217275a_0_long > > In the directory DIR_E here (from above), there is only one file without a > __head_ in the pathname -- the file aboveShould I be deleting both these > _long files without the __head_ in DIR_E

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-19 Thread Samuel Just
p. > Regards, > Jeff > > On Thu, Mar 17, 2016 at 11:58 AM, Samuel Just wrote: >> >> Oh, it's getting a stat mismatch. I think what happened is that on >> one of the earlier repairs it reset the stats to the wrong value (the >> orphan was causing the primary

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-19 Thread Samuel Just
ly have this one example so far.is the 79C of the hash > constant? Would the hash pick up another hex character if the pg splits > again? > > Thanks, > Jeff > > On Wed, Mar 16, 2016 at 10:24 AM, Samuel Just wrote: >> >> There is a directory structure hash, it&#x

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-15 Thread Samuel Just
time to put together a branch in the next week or two if you want to wait. It's still probably worth trying removing that object in 70.459. -Sam On Tue, Mar 15, 2016 at 6:03 PM, Samuel Just wrote: > The bug is entirely independent of hardware issues -- entirely a ceph > bug. xf

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-15 Thread Samuel Just
had hardware issues > during the remapping or would it have happened regardless of the hardware > issues? e.g. I'm not planning to add any additional hardware soon, but > would the bug pop again on an (unpatched) system not subject to any > remapping? > > thanks, > jeff &

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-15 Thread Samuel Just
.g. will the orphans always > be size 0? > > Thanks! > Jeff > > On Tue, Mar 15, 2016 at 5:35 PM, Samuel Just wrote: >> >> Ok, a branch merged to master which should fix this >> (https://github.com/ceph/ceph/pull/8136). It'll be backported in due >> cour

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-15 Thread Samuel Just
Ok, a branch merged to master which should fix this (https://github.com/ceph/ceph/pull/8136). It'll be backported in due course. The problem is that that patch won't clean orphaned files that already exist. Let me explain a bit about what the orphaned files look like. The problem is files with

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-08 Thread Samuel Just
s?The PGs there seem to now be remapped elsewhere. > > Regards, > Jeff > > > On Tue, Mar 8, 2016 at 2:09 PM, Samuel Just wrote: >> >> The pgs are not actually inconsistent (that is, I think that all of >> the real objects are present and healthy). I thin

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-08 Thread Samuel Just
Yep On Tue, Mar 8, 2016 at 12:08 PM, Jeffrey McDonald wrote: > Will do, I'm in the midst of the xfs filesystem check so it will be a bit > before I have the filesystem back. > jeff > > On Tue, Mar 8, 2016 at 2:05 PM, Samuel Just wrote: >> >> There are 3 other

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-08 Thread Samuel Just
ctive+clean > 43 active+clean+inconsistent >7 active+clean+scrubbing+deep > 7 active+clean+scrubbing > > Jeff > > On Tue, Mar 8, 2016 at 2:00 PM, Samuel Just wrote: >> >> Yeah, that procedure should have i

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-08 Thread Samuel Just
\u0386\uBC2E4WACXX\u2\uACTTGA.tar.gz.2~TWfBUlEkk4EPH4u\uNkMjmz65CRSkJA3._215ce1442b16dc173b77_0_long On Tue, Mar 8, 2016 at 12:00 PM, Samuel Just wrote: > Yeah, that procedure should have isolated any filesystem issues. Are > there still unfound objects? > -sam > > On Tue, Mar 8, 2

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-08 Thread Samuel Just
lowing fewer backfill > threads. > > Jeff > > On Tue, Mar 8, 2016 at 1:56 PM, Samuel Just wrote: >> >> By "migrated", you mean you marked them out one at a time and let ceph >> recover them over to new nodes? >> -Sam >> >> On Tue, M

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-08 Thread Samuel Just
hardware. > jeff > > On Tue, Mar 8, 2016 at 1:52 PM, Samuel Just wrote: >> >> ceph3 is not the same host as ceph03? >> -Sam >> >> On Tue, Mar 8, 2016 at 11:48 AM, Jeffrey McDonald >> wrote: >> > Hi Sam, >> > >> > 1) Are t

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-08 Thread Samuel Just
. > > > 4) Has anything unusual ever happened to the host which osd.307 is on? > Particularly a power loss? > > I don't recall anything. A couple of times the data center overheated (air > ) but these nodes are in a water-cooled enclosure and were OK. What I did > hav

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-08 Thread Samuel Just
e in the PGs and are labeled as unfound.Its important for us to >> maintain user confidence in the system so I need a fix as soon as >> possible.. >> >> The log files from the OSDs are here: >> >> https://drive.google.com/open?id=0Bzz8TrxFvfembkt2XzlCZFVJZFU &

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-07 Thread Samuel Just
Nevermind on http://tracker.ceph.com/issues/14766 , OSD::remove_dir uses the right collection_list_partial. -Sam On Mon, Mar 7, 2016 at 5:26 PM, Samuel Just wrote: > Yep, just as before. Actually, do it twice (wait for 'scrubbing' to > go away each time). > -Sam > > On

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-07 Thread Samuel Just
70.459 > 3- stop once the 70.459 pg goes inconsistent? > > Thanks, > Jeff > > > On Mon, Mar 7, 2016 at 6:52 PM, Samuel Just wrote: >> >> Hmm, I'll look into this a bit more tomorrow. Can you get the tree >> structure of the 70.459 pg directory on osd.30

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-07 Thread Samuel Just
On the plus side, I think I figured out http://tracker.ceph.com/issues/14766. -Sam On Mon, Mar 7, 2016 at 4:52 PM, Samuel Just wrote: > Hmm, I'll look into this a bit more tomorrow. Can you get the tree > structure of the 70.459 pg directory on osd.307 (find . will do fine). > -

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-07 Thread Samuel Just
Hmm, I'll look into this a bit more tomorrow. Can you get the tree structure of the 70.459 pg directory on osd.307 (find . will do fine). -Sam On Mon, Mar 7, 2016 at 4:50 PM, Jeffrey McDonald wrote: > 307 is on ceph03. > Jeff > > On Mon, Mar 7, 2016 at 6:48 PM, Samuel Just w

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-07 Thread Samuel Just
Which node is osd.307 on? -Sam On Mon, Mar 7, 2016 at 4:43 PM, Samuel Just wrote: > ' I didn't see the errors in the tracker on the new nodes, but they > were only receiving new data, not migrating it.' -- What do you mean > by that? > -Sam > > On Mon, Mar 7,

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-07 Thread Samuel Just
2 22:08:27 > UTC 2015 x86_64 x86_64 x86_64 GNU/Linux > ceph09: Linux ceph09 3.19.0-51-generic #58~14.04.1-Ubuntu SMP Fri Feb 26 > 22:02:58 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux > > > On Mon, Mar 7, 2016 at 6:39 PM, Samuel Just wrote: >> >> What filesystem and kernel

Re: [ceph-users] inconsistent PG -> unfound objects on an erasure coded system

2016-03-07 Thread Samuel Just
I'd rather you just scrubbed the same pg with the same osds and the same debugging. -Sam On Mon, Mar 7, 2016 at 4:40 PM, Samuel Just wrote: > Yes, the unfound objects are due to a bug in the repair command. I > suggest you don't repair anything, actually. I don't think

  1   2   3   4   >