[ceph-users] CephFS client df command showing raw space after adding second pool to mds

2019-01-03 Thread David C
, is this expected? Thanks, David ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Re: [ceph-users] Help with setting device-class rule on pool without causing data to move

2019-01-03 Thread David C
HDD and SDD hosts (I didn't have any mixed devices in hosts), set the Metadata pool to use the new rule, waited for recovery and then set the Metadata pool to use an SSD-only rule. Not sure if that intermediary stage was strictly necessary, I was concerned about inactive PGs. Thanks, David On Mon

Re: [ceph-users] list admin issues

2019-01-02 Thread David Galloway
eleted.' >> >> can anyone check MTA logs to see what the bounce is? > > Me too. Happens regularly and only on ceph-users, not on sepia or > ceph-maintainers, etc. David, Dan, could you or someone you know look > into this? > As far as I know, we don't have shell acces

[ceph-users] Help with setting device-class rule on pool without causing data to move

2018-12-30 Thread David C
Hi All I'm trying to set the existing pools in a Luminous cluster to use the hdd device-class but without moving data around. If I just create a new rule using the hdd class and set my pools to use that new rule it will cause a huge amount of data movement even though the pgs are all already on

Re: [ceph-users] Bluestore nvme DB/WAL size

2018-12-21 Thread David C
I'm in a similar situation, currently running filestore with spinners and journals on NVME partitions which are about 1% of the size of the OSD. If I migrate to bluestore, I'll still only have that 1% available. Per the docs, if my block.db device fills up, the metadata is going to spill back onto

Re: [ceph-users] active+recovering+degraded after cluster reboot

2018-12-15 Thread David C
Yep, that cleared it. Sorry for the noise! On Sun, Dec 16, 2018 at 12:16 AM David C wrote: > Hi Paul > > Thanks for the response. Not yet, just being a bit cautious ;) I'll go > ahead and do that. > > Thanks > David > > > On Sat, 15 Dec 2018, 23:39 Paul Emmeri

Re: [ceph-users] active+recovering+degraded after cluster reboot

2018-12-15 Thread David C
Hi Paul Thanks for the response. Not yet, just being a bit cautious ;) I'll go ahead and do that. Thanks David On Sat, 15 Dec 2018, 23:39 Paul Emmerich Did you unset norecover? > > > Paul > > -- > Paul Emmerich > > Looking for help with your Ceph cluster? Contac

[ceph-users] active+recovering+degraded after cluster reboot

2018-12-15 Thread David C
How would you recommend I proceed at this point? Thanks David ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[ceph-users] Why does "df" against a mounted cephfs report (vastly) different free space?

2018-12-12 Thread David Young
Hi all, I have a cluster used exclusively for cephfs (A EC "media" pool, and a standard metadata pool for the cephfs). "ceph -s" shows me: --- data: pools: 2 pools, 260 pgs objects: 37.18 M objects, 141 TiB usage: 177 TiB used, 114 TiB / 291 TiB avail pgs: 260

Re: [ceph-users] Luminous v12.2.10 released

2018-12-12 Thread David Galloway
Hey Dan, Thanks for bringing this to our attention. Looks like it did get left out. I just pushed the package and added a step to the release process to make sure packages don't get skipped again like that. - David On 12/12/2018 11:03 AM, Dan van der Ster wrote: > Hey Abhishek, > >

Re: [ceph-users] Deploying an Active/Active NFS Cluster over CephFS

2018-12-12 Thread David C
Hi Jeff Many thanks for this! Looking forward to testing it out. Could you elaborate a bit on why Nautilus is recommended for this set-up please. Would attempting this with a Luminous cluster be a non-starter? On Wed, 12 Dec 2018, 12:16 Jeff Layton (Sorry for the duplicate email to ganesha

Re: [ceph-users] Lost 1/40 OSDs at EC 4+1, now PGs are incomplete

2018-12-11 Thread David Young
cause some PG’s to not >> have at least K size available as you only have 1 extra M. >> >> As per the error you can get your pool back online by setting min_size to 4. >> >> However this would only be a temp fix while you get the OSD back online / >> rebuilt so you

[ceph-users] Lost 1/40 OSDs at EC 4+1, now PGs are incomplete

2018-12-11 Thread David Young
Hi all, I have a small 2-node cluster with 40 OSDs, using erasure coding 4+1 I lost osd38, and now I have 39 incomplete PGs. --- PG_AVAILABILITY Reduced data availability: 39 pgs inactive, 39 pgs incomplete pg 22.2 is incomplete, acting [19,33,10,8,29] (reducing pool media min_size from 5

Re: [ceph-users] CEPH DR RBD Mount

2018-11-30 Thread David C
Is that one big xfs filesystem? Are you able to mount with krbd? On Tue, 27 Nov 2018, 13:49 Vikas Rana Hi There, > > We are replicating a 100TB RBD image to DR site. Replication works fine. > > rbd --cluster cephdr mirror pool status nfs --verbose > > health: OK > > images: 1 total > > 1

Re: [ceph-users] Migrate OSD journal to SSD partition

2018-11-19 Thread David Turner
For this the procedure is generally to stop the osd, flush the journal, update the symlink on the osd to the new journal location, mkjournal, start osd. You shouldn't need to do anything in the ceph.conf file. On Thu, Nov 8, 2018 at 2:41 AM wrote: > Hi all, > > > > I have been trying to

Re: [ceph-users] Mimic - EC and crush rules - clarification

2018-11-16 Thread David Turner
only using as much space as 2x replication. On Thu, Nov 1, 2018 at 11:25 PM Wladimir Mutel wrote: > David Turner wrote: > > Yes, when creating an EC profile, it automatically creates a CRUSH rule > > specific for that EC profile. You are also correct that 2+1 doesn't >

Re: [ceph-users] ceph 12.2.9 release

2018-11-07 Thread David Turner
My big question is that we've had a few of these releases this year that are bugged and shouldn't be upgraded to... They don't have any release notes or announcement and the only time this comes out is when users finally ask about it weeks later. Why is this not proactively announced to avoid a

Re: [ceph-users] Any backfill in our cluster makes the cluster unusable and takes forever

2018-11-05 Thread David Turner
to update my omap backends. On Mon, Nov 5, 2018 at 4:26 PM Pavan Rallabhandi < prallabha...@walmartlabs.com> wrote: > Not sure I understand that, but starting Luminous, the filestore omap > backend is rocksdb by default. > > > > *From: *David Turner > *Date: *Monday,

Re: [ceph-users] Any backfill in our cluster makes the cluster unusable and takes forever

2018-11-05 Thread David Turner
s feature was supported in Jewel starting 10.2.11, ref > https://github.com/ceph/ceph/pull/18010 > > > > I thought you mentioned you were using Luminous 12.2.4. > > > > *From: *David Turner > *Date: *Friday, November 2, 2018 at 5:21 PM > > > *To: *Pavan Rallabhandi

Re: [ceph-users] Any backfill in our cluster makes the cluster unusable and takes forever

2018-11-02 Thread David Turner
as well that can be used to set up the omap backend db. On Fri, Nov 2, 2018, 4:26 PM Pavan Rallabhandi wrote: > It was Redhat versioned Jewel. But may be more relevantly, we are on > Ubuntu unlike your case. > > > > *From: *David Turner > *Date: *Friday, Novembe

Re: [ceph-users] Any backfill in our cluster makes the cluster unusable and takes forever

2018-11-02 Thread David Turner
. > > Thanks, > -Pavan. > > From: David Turner > Date: Monday, October 1, 2018 at 1:37 PM > To: Pavan Rallabhandi > Cc: ceph-users > Subject: EXT: Re: [ceph-users] Any backfill in our cluster makes the > cluster unusable and takes forever > > I tried m

Re: [ceph-users] Mimic - EC and crush rules - clarification

2018-11-01 Thread David Turner
Yes, when creating an EC profile, it automatically creates a CRUSH rule specific for that EC profile. You are also correct that 2+1 doesn't really have any resiliency built in. 2+2 would allow 1 node to go down while still having your data accessible. It will use 2x data to raw as opposed to

Re: [ceph-users] Packages for debian in Ceph repo

2018-10-30 Thread David Turner
What version of qemu-img are you using? I found [1] this when poking around on my qemu server when checking for rbd support. This version (note it's proxmox) has rbd listed as a supported format. [1] # qemu-img -V; qemu-img --help|grep rbd qemu-img version 2.11.2pve-qemu-kvm_2.11.2-1 Copyright

Re: [ceph-users] Balancer module not balancing perfectly

2018-10-30 Thread David Turner
>From the balancer module's code for v 12.2.7 I noticed [1] these lines which reference [2] these 2 config options for upmap. You might try using more max iterations or a smaller max deviation to see if you can get a better balance in your cluster. I would try to start with [3] these

Re: [ceph-users] OSD node reinstallation

2018-10-30 Thread David Turner
ceph.keyring >> >> --- >> Alex >> >> On Tue, Oct 30, 2018 at 4:48 AM David Turner >> wrote: >> > >> > Set noout, reinstall the OS without going the OSDs (including any >> journal partitions and maintaining any dmcrypt keys if you have >

Re: [ceph-users] OSD node reinstallation

2018-10-29 Thread David Turner
Set noout, reinstall the OS without going the OSDs (including any journal partitions and maintaining any dmcrypt keys if you have encryption), install ceph, make sure the ceph.conf file is correct,zip start OSDs, unset noout once they're back up and in. All of the data the OSD needs to start is on

Re: [ceph-users] reducing min_size on erasure coded pool may allow recovery ?

2018-10-29 Thread David Turner
min_size should be at least k+1 for EC. There are times to use k for emergencies like you had. I would suggest seeing it back to 3 once your back to healthy. As far as why you needed to reduce min_size, my guess would be that recovery would have happened as long as k copies were up. Were the PG's

Re: [ceph-users] Need advise on proper cluster reweighing

2018-10-28 Thread David Turner
Which version of Ceph are you running? Do you have any kernel clients? If yes, can which version kernel? These questions are all leading to see if you can enable the Luminous/Mimic mgr module balancer with upmap. If you can, it is hands down the best way to balance your cluster. On Sat, Oct 27,

Re: [ceph-users] Verifying the location of the wal

2018-10-28 Thread David Turner
If your had a specific location for the wal it would show up there. If there is no entry for the wal, then it is using the same seeing as the db. On Sun, Oct 28, 2018, 9:26 PM Robert Stanford wrote: > > Mehmet: it doesn't look like wal is mentioned in the osd metadata. I see > bluefs slow,

Re: [ceph-users] Migrate/convert replicated pool to EC?

2018-10-26 Thread David Turner
It is indeed adding a placement target and not removing it replacing the pool. The get/put wouldn't be a rados or even ceph command, you would do it through an s3 client. On Fri, Oct 26, 2018, 9:38 AM Matthew Vernon wrote: > Hi, > > On 26/10/2018 12:38, Alexandru Cucu wrote: > > > Have a look

Re: [ceph-users] RGW: move bucket from one placement to another

2018-10-25 Thread David Turner
Resharding a bucket won't affect the data in the bucket. After you change the placement for a bucket, you could update where the data is by re-writing all of the data in the bucket. On Thu, Oct 25, 2018 at 8:48 AM Jacek Suchenia wrote: > Hi > > We have a bucket created with LocationConstraint

Re: [ceph-users] Migrate/convert replicated pool to EC?

2018-10-25 Thread David Turner
There are no tools to migrate in either direction between EC and Replica. You can't even migrate an EC pool to a new EC profile. With RGW you can create a new data pool and new objects will be written to the new pool. If your objects have a lifecycle, then eventually you'll be to the new pool

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread David Turner
WAL on the NVMe? Is this "data disk" > still an ssd? > > > > On Mon, Oct 22, 2018 at 3:34 PM David Turner > wrote: > >> And by the data disk I mean that I didn't specify a location for the DB >> partition. >> >> On Mon, Oct 22, 2018 at 4:06

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread David Turner
lization > on the cluster than what's in the pools, the excess equal to about wal size > * number of osds... > > On Mon, Oct 22, 2018 at 3:35 PM David Turner > wrote: > >> My DB doesn't have a specific partition anywhere, but there's still a >> symlink for it to t

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread David Turner
nfiguration/bluestore-config-ref/ > > On Mon, Oct 22, 2018 at 3:13 PM Robert Stanford > wrote: > >> >> We're out of sync, I think. You have your DB on your data disk so your >> block.db symlink points to that disk, right? There is however no wal >> symlink?

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread David Turner
Track down where it says they point to? Does it match what you expect? It does for me. I have my DB on my data disk and my WAL on a separate NVMe. On Mon, Oct 22, 2018 at 3:21 PM Robert Stanford wrote: > > David - is it ensured that wal and db both live where the symlink > block.

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread David Turner
And by the data disk I mean that I didn't specify a location for the DB partition. On Mon, Oct 22, 2018 at 4:06 PM David Turner wrote: > Track down where it says they point to? Does it match what you expect? > It does for me. I have my DB on my data disk and my WAL on a separat

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread David Turner
You can always just go to /var/lib/ceph/osd/ceph-{osd-num}/ and look at where the symlinks for block and block.wal point to. On Mon, Oct 22, 2018 at 12:29 PM Robert Stanford wrote: > > That's what they say, however I did exactly this and my cluster > utilization is higher than the total pool

Re: [ceph-users] ceph df space usage confusion - balancing needed?

2018-10-22 Thread David Turner
I haven't had crush-compat do anything helpful for balancing my clusters. upmap has been amazing and balanced my clusters far better than anything else I've ever seen. I would go so far as to say that upmap can achieve a perfect balance. It seems to evenly distribute the PGs for each pool onto

Re: [ceph-users] bluestore compression enabled but no data compressed

2018-10-19 Thread David Turner
wrote: > Hi David, > > sorry for the slow response, we had a hell of a week at work. > > OK, so I had compression mode set to aggressive on some pools, but the > global option was not changed, because I interpreted the documentation as > "pool settings take precedence

Re: [ceph-users] Troubleshooting hanging storage backend whenever there is any cluster change

2018-10-18 Thread David Turner
2TB SSD. > > > > Log of osd.19 is here: > > https://pastebin.com/raw/6DWwhS0A > > > > Am 13.10.2018 um 21:20 schrieb Stefan Priebe - Profihost AG: > >> Hi David, > >> > >> i think this should be the problem - form a new log from today: &

Re: [ceph-users] ceph pg/pgp number calculation

2018-10-18 Thread David Turner
Not all pools need the same amount of PGs. When you get to so many pools you want to start calculating how much data each pool will have. If 1 of your pools will have 80% of your data in it, it should have 80% of your PGs. The metadata pools for rgw likely won't need more than 8 or so PGs each. If

Re: [ceph-users] SSD for MON/MGR/MDS

2018-10-15 Thread David Turner
Mgr and MDS do not use physical space on a disk. Mons do use the disk and benefit from SSDs, but they write a lot of stuff all the time. Depending why the SSDs aren't suitable for OSDs, they might not be suitable for mons either. On Mon, Oct 15, 2018, 7:16 AM ST Wong (ITSC) wrote: > Hi all, > >

Re: [ceph-users] bluestore compression enabled but no data compressed

2018-10-12 Thread David Turner
settings which you can see there about minimum ratios, blob sizes, etc). To compress current data, you need to re-write it. On Fri, Oct 12, 2018 at 10:41 AM Frank Schilder wrote: > Hi David, > > thanks for your quick answer. When I look at both references, I see > exactly the same comman

Re: [ceph-users] Anyone tested Samsung 860 DCT SSDs?

2018-10-12 Thread David Turner
What do you want to use these for? "5 Year or 0.2 DWPD" is the durability of this drive which is absolutely awful for most every use in Ceph. Possibly if you're using these for data disks (not DB or WAL) and you plan to have a more durable media to host the DB+WAL on... this could work. Or if

Re: [ceph-users] Troubleshooting hanging storage backend whenever there is any cluster change

2018-10-12 Thread David Turner
The PGs per OSD does not change unless the OSDs are marked out. You have noout set, so that doesn't change at all during this test. All of your PGs peered quickly at the beginning and then were active+undersized the rest of the time, you never had any blocked requests, and you always had

Re: [ceph-users] bluestore compression enabled but no data compressed

2018-10-12 Thread David Turner
It's all of the settings that you found in your first email when you dumped the configurations and such. http://docs.ceph.com/docs/master/rados/configuration/bluestore-config-ref/#inline-compression On Fri, Oct 12, 2018 at 7:36 AM Frank Schilder wrote: > Hi David, > > thanks for your

Re: [ceph-users] Troubleshooting hanging storage backend whenever there is any cluster change

2018-10-11 Thread David Turner
83f2~1,d83f4~3,d83f8~3,d83fd~2,d8402~1,d8405~1,d8407~1,d840a~2,d840f~1,d8411~1,d8413~3,d8417~3,d841c~4,d8422~4,d8428~2,d842b~1,d842e~1,d8430~1,d8432~5,d843a~1,d843c~3,d8440~5,d8447~1,d844a~1,d844d~1,d844f~1,d8452~1,d8455~1,d8457~1,d8459~2,d845d~2,d8460~1,d8462~3,d8467~1,d8469~1,d846b~2,d846e~2,d8

Re: [ceph-users] Troubleshooting hanging storage backend whenever there is any cluster change

2018-10-11 Thread David Turner
My first guess is to ask what your crush rules are. `ceph osd crush rule dump` along with `ceph osd pool ls detail` would be helpful. Also if you have a `ceph status` output from a time where the VM RBDs aren't working might explain something. On Thu, Oct 11, 2018 at 1:12 PM Nils Fahldieck -

Re: [ceph-users] Inconsistent PG, repair doesn't work

2018-10-11 Thread David Turner
As a part of a repair is queuing a deep scrub. As soon as the repair part is over the deep scrub continues until it is done. On Thu, Oct 11, 2018, 12:26 PM Brett Chancellor wrote: > Does the "repair" function use the same rules as a deep scrub? I couldn't > get one to kick off, until I

Re: [ceph-users] https://ceph-storage.slack.com

2018-10-11 Thread David Turner
I have 4 other slack servers that I'm in for work and personal hobbies. It's just easier for me to maintain one more slack server than have a separate application for IRC. On Thu, Oct 11, 2018, 11:02 AM John Spray wrote: > On Thu, Oct 11, 2018 at 8:44 AM Marc Roos > wrote: > > > > > > Why

Re: [ceph-users] OSD log being spammed with BlueStore stupidallocator dump

2018-10-10 Thread David Turner
Not a resolution, but an idea that you've probably thought of. Disabling logging on any affected OSDs (possibly just all of them) seems like a needed step to be able to keep working with this cluster to finish the upgrade and get it healthier. On Wed, Oct 10, 2018 at 6:37 PM Wido den Hollander

Re: [ceph-users] https://ceph-storage.slack.com

2018-10-10 Thread David Turner
I would like an invite to. drakonst...@gmail.com On Wed, Sep 19, 2018 at 1:02 PM Gregory Farnum wrote: > Done. :) > > On Tue, Sep 18, 2018 at 12:15 PM Alfredo Daniel Rezinovsky < > alfredo.rezinov...@ingenieria.uncuyo.edu.ar> wrote: > >> Can anyone add me to this slack? >> >> with my email

Re: [ceph-users] HEALTH_WARN 2 osd(s) have {NOUP, NODOWN, NOIN, NOOUT} flags set

2018-10-10 Thread David Turner
There is a newer [1] feature to be able to set flags per OSD instead of cluster wide. This way you can prevent a problem host from marking its OSDs down while the rest ofthe cluster is capable of doing so. [2] These commands ought to clear up your status. [1]

Re: [ceph-users] Does anyone use interactive CLI mode?

2018-10-10 Thread David Turner
I know that it existed, but I've never bothered using it. In applications like Python where you can get a different reaction by interacting with it line by line and setting up an environment it is very helpful. Ceph, however, doesn't have any such environment variables that would make this more

Re: [ceph-users] Can't remove DeleteMarkers in rgw bucket

2018-10-09 Thread David Turner
I would suggest trying to delete the bucket using radosgw-admin. If you can't get that to work, then I would go towards deleting the actual RADOS objects. There are a few threads on the ML that talk about manually deleting a bucket. On Thu, Sep 20, 2018 at 2:04 PM Sean Purdy wrote: > Hi, > >

Re: [ceph-users] bluestore compression enabled but no data compressed

2018-10-09 Thread David Turner
When I've tested compression before there are 2 places you need to configure compression. On the OSDs in the configuration settings that you mentioned, but also on the [1] pools themselves. If you have the compression mode on the pools set to none, then it doesn't matter what the OSDs

Re: [ceph-users] Error-code 2002/API 405 S3 REST API. Creating a new bucket

2018-10-09 Thread David Turner
Can you outline the process you're using to access the REST API? It's hard to troubleshoot this without knowing how you were trying to do this. On Mon, Sep 17, 2018 at 7:09 PM Michael Schäfer wrote: > Hi, > > We have a problem with the radosgw using the S3 REST API. > Trying to create a new

Re: [ceph-users] radosgw bucket stats vs s3cmd du

2018-10-09 Thread David Turner
Have you looked at your Garbage Collection. I would guess that your GC is behind and that radosgw-admin is accounting for that space knowing that it hasn't been freed up yet, whiles 3cmd doesn't see it since it no longer shows in the listing. On Tue, Sep 18, 2018 at 4:45 AM Luis Periquito

Re: [ceph-users] list admin issues

2018-10-06 Thread David C
Same issue here, Gmail user, member of different lists but only get disabled on ceph-users. Happens about once a month but had three in Sept. On Sat, 6 Oct 2018, 18:28 Janne Johansson, wrote: > Den lör 6 okt. 2018 kl 15:06 skrev Elias Abacioglu > : > > > > Hi, > > > > I'm bumping this old

Re: [ceph-users] Any backfill in our cluster makes the cluster unusable and takes forever

2018-10-01 Thread David Turner
It really seems to be something with RocksDB on centOS. I still think > you can try removing “compression=kNoCompression” from the > filestore_rocksdb_options And/Or check if rocksdb is expecting snappy to be > enabled. > > Thanks, > -Pavan. > > From: David Turner

Re: [ceph-users] mount cephfs from a public network ip of mds

2018-09-30 Thread David Turner
r clients with only > 1Gb nic will go through 140.109.0.0 (1Gb LAN) to ask mon or to read/write > to osds. This is why my osds also have 1Gb and 10Gb nics with 140.109.0.0 > and 10.32.0.0 networking respectively. > > Cheers > Joshua > > On Sun, Sep 30, 2018 at 12:09 PM

Re: [ceph-users] mount cephfs from a public network ip of mds

2018-09-29 Thread David Turner
The cluster/private network is only used by the OSDs. Nothing else in ceph or its clients communicate using it. Everything other than osd to osd communication uses the public network. That includes the MONs, MDSs, clients, and anything other than an osd talking to an osd. Nothing else other than

Re: [ceph-users] Any backfill in our cluster makes the cluster unusable and takes forever

2018-09-27 Thread David Turner
ssion to be enabled, can you try > removing “compression=kNoCompression” from the filestore_rocksdb_options? > And/or you might want to check if rocksdb is expecting snappy to be enabled. > > From: David Turner > Date: Tuesday, September 18, 2018 at 6:01 PM > To: Pavan Rallabhand

Re: [ceph-users] Mimic upgrade failure

2018-09-20 Thread David Turner
hutting > down a mon to see if a single one was the problem, but they keep electing no > matter the mix of them that are up. Part of it feels like a networking > problem but I haven't been able to find a culprit yet as everything was > working normally before starting the upgrade. Other t

Re: [ceph-users] Ceph Mimic packages not available for Ubuntu Trusty

2018-09-19 Thread David Turner
No, Ceph Mimic will not be available for Ubuntu Trusty 14.04. That release is almost 4.5 years old now, you should start planning towards an OS upgrade. On Wed, Sep 19, 2018 at 8:54 AM Jakub Jaszewski wrote: > Hi Cephers, > > Any plans for Ceph Mimic packages for Ubuntu Trusty? I found only >

Re: [ceph-users] Any backfill in our cluster makes the cluster unusable and takes forever

2018-09-18 Thread David Turner
> > But the stack trace didn’t hint as if the superblock on the OSD is > still considering the omap backend to be leveldb and to do with the > compression. > > Thanks, > -Pavan. > > From: David Turner > Date: Tuesday, September 18, 2018 at 5:07 PM &g

Re: [ceph-users] Any backfill in our cluster makes the cluster unusable and takes forever

2018-09-18 Thread David Turner
, > -Pavan. > > From: ceph-users on behalf of David > Turner > Date: Tuesday, September 18, 2018 at 4:09 PM > To: ceph-users > Subject: EXT: [ceph-users] Any backfill in our cluster makes the cluster > unusable and takes forever > > I've finally learned enough ab

[ceph-users] Any backfill in our cluster makes the cluster unusable and takes forever

2018-09-18 Thread David Turner
I've finally learned enough about the OSD backend track down this issue to what I believe is the root cause. LevelDB compaction is the common thread every time we move data around our cluster. I've ruled out PG subfolder splitting, EC doesn't seem to be the root cause of this, and it is cluster

Re: [ceph-users] can we drop support of centos/rhel 7.4?

2018-09-14 Thread David Turner
It's odd to me because this feels like the opposite direction of the rest of Ceph. Making management and operating Ceph simpler and easier. Requiring fast OS upgrades on dot releases of Ceph versions is not that direction at all. On Fri, Sep 14, 2018, 9:25 AM David Turner wrote: > Release da

Re: [ceph-users] can we drop support of centos/rhel 7.4?

2018-09-14 Thread David Turner
Release dates RHEL 7.4 - July 2017 Luminous 12.2.0 - August 2017 CentOS 7.4 - September 2017 RHEL 7.5 - April 2018 CentOS 7.5 - May 2018 Mimic 13.2.0 - June 2018 In the world of sysadmins it takes time to let new releases/OS's simmer before beginning to test them let alone upgrading to them. It

[ceph-users] Updating CRUSH Tunables to Jewel from Hammer

2018-09-13 Thread David Turner
I have a stage cluster with 4 HDDs and an SSD in each host. I have an EC profile that specifically chooses HDDs for placement. Also several Replica pools that write to either HDD or SSD. This has all worked well for a while. When I updated the Tunables to Jewel on the cluster, all of a sudden

Re: [ceph-users] Performance predictions moving bluestore wall, db to ssd

2018-09-12 Thread David Turner
Sorry, I was wrong that it was you. I just double checked. But there is a new thread as of this morning about this topic where someone is running benchmark tests with numbers titled "Benchmark does not show gains with DB on SSD". On Wed, Sep 12, 2018 at 12:20 PM David Turner wro

Re: [ceph-users] Performance predictions moving bluestore wall, db to ssd

2018-09-12 Thread David Turner
You already have a thread talking about benchmarking the addition of WAL and DB partitions to an OSD. Why are you creating a new one about the exact same thing? As with everything, the performance increase isn't even solely answerable by which drives you have, there are a lot of factors that

Re: [ceph-users] Benchmark does not show gains with DB on SSD

2018-09-12 Thread David Turner
If you're writes are small enough (64k or smaller) they're being placed on the WAL device regardless of where your DB is. If you change your testing to use larger writes you should see a difference by adding the DB. Please note that the community has never recommended using less than 120GB DB

Re: [ceph-users] Ceph balancer "Error EAGAIN: compat weight-set not available"

2018-09-11 Thread David Turner
ceph balancer status ceph config-key dump | grep balancer ceph osd dump | grep min_compat_client ceph osd crush dump | grep straw ceph osd crush dump | grep profile ceph features You didn't mention it, but based on your error and my experiences over the last week getting the balancer working,

Re: [ceph-users] rbd-nbd on CentOS

2018-09-10 Thread David Turner
while allowing CentOS 7.5 clients map RBDs? On Mon, Sep 10, 2018 at 1:33 PM Ilya Dryomov wrote: > On Mon, Sep 10, 2018 at 7:19 PM David Turner > wrote: > > > > I haven't found any mention of this on the ML and Google's results are > all about compiling your own kernel

[ceph-users] rbd-nbd on CentOS

2018-09-10 Thread David Turner
I haven't found any mention of this on the ML and Google's results are all about compiling your own kernel to use NBD on CentOS. Is everyone that's using rbd-nbd on CentOS honestly compiling their own kernels for the clients? This feels like something that shouldn't be necessary anymore. I would

Re: [ceph-users] upgrade jewel to luminous with ec + cache pool

2018-09-10 Thread David Turner
Yes, migrating to 12.2.8 is fine. Migrating to not use the cache tier is as simple as changing the ec pool mode to allow EC over writes, changing the cache tier mode to forward, flushing the tier, and removing it. Basically once you have EC over writes just follow the steps in the docs for

Re: [ceph-users] Mixing EC and Replicated pools on HDDs in Ceph RGW Luminous

2018-09-09 Thread David Turner
You can indeed have multiple types of pools on the same disks. Go ahead and put the non-ec pool with a replicated ruleset on the HDDs with the EC data pool. I believe your correct that the non-ec pool gets cleared out when the upload is complete and the file is flushed to EC. On Sun, Sep 9, 2018,

Re: [ceph-users] Safe to use RBD mounts for Docker volumes on containerized Ceph nodes

2018-09-08 Thread David Turner
The problem is with the kernel pagecache. If that is still shared in a containerized environment with the OSDs in containers and RBDs which are married on The node outside of containers, then it is indeed still a problem. I would guess that's the case, but I do not know for certain. Using rbd-nbd

Re: [ceph-users] Mimic upgrade failure

2018-09-08 Thread David Turner
What osd/mon/etc config settings do you have that are not default? It might be worth utilizing nodown to stop osds from marking each other down and finish the upgrade to be able to set the minimum osd version to mimic. Stop the osds in a node, manually mark them down, start them back up in mimic.

Re: [ceph-users] advice with erasure coding

2018-09-08 Thread David Turner
I tested running VMs on EC back in Hammer. The performance was just bad. I didn't even need much io, but even performing standard maintenance was annoying enough that I abandoned the idea. I didn't really try to tweak settings to make it work and I only had a 3 node cluster running 2+1. I did use

Re: [ceph-users] CephFS tar archiving immediately after writing

2018-09-07 Thread David Turner
ther necessary functions) so you can't disable it on the > server. > -Greg > > On Fri, Sep 7, 2018 at 11:48 AM David Turner > wrote: > >> Is it be possible to disable this feature? Very few filesystems >> calculate the size of its folder's contents. I know I enjoy

Re: [ceph-users] CephFS tar archiving immediately after writing

2018-09-07 Thread David Turner
e > tree to enable good differential backups. Not sure what the status of that > is. > -Greg > > On Fri, Sep 7, 2018 at 11:06 AM David Turner > wrote: > >> We have an existing workflow that we've moved from one server sharing a >> local disk via NFS to secondary ser

[ceph-users] CephFS tar archiving immediately after writing

2018-09-07 Thread David Turner
We have an existing workflow that we've moved from one server sharing a local disk via NFS to secondary servers to all of them mounting CephFS. The primary server runs a script similar to [1] this, but since we've moved it into CephFS, we get [2] this error. We added the sync in there to try to

Re: [ceph-users] Ceph talks from Mounpoint.io

2018-09-06 Thread David Turner
They mentioned that they were going to send the slides to everyone that had their badges scanned at the conference. I haven't seen that email come out yet, though. On Thu, Sep 6, 2018 at 4:14 PM Gregory Farnum wrote: > Unfortunately I don't believe anybody collected the slide files, so they >

Re: [ceph-users] help needed

2018-09-06 Thread David Turner
The official ceph documentation recommendations for a db partition for a 4TB bluestore osd would be 160GB each. Samsung Evo Pro is not an Enterprise class SSD. A quick search of the ML will allow which SSDs people are using. As was already suggested, the better option is an HBA as opposed to a

Re: [ceph-users] [Ceph-community] How to setup Ceph OSD auto boot up on node reboot

2018-09-05 Thread David Turner
The magic sauce to get Filestore OSDs to start on a node reboot is to make sure that all of your udev magic is correct. In particular you need to have the correct UUID set for all partitions. I haven't dealt with it in a long time, but I've written up a few good ML responses about it. On Tue,

Re: [ceph-users] v12.2.8 Luminous released

2018-09-05 Thread David Turner
- > > * bluestore: set correctly shard for existed Collection (issue#24761, > pr#22860, Jianpeng Ma) > > * build/ops: Boost system library is no longer required to compile and > link example librados program (issue#25054, pr#23202, Nathan Cutler) > > * build/ops: Bring b

Re: [ceph-users] Luminous new OSD being over filled

2018-09-04 Thread David Turner
preventing your cluster from being poorly balanced. On Mon, Sep 3, 2018 at 12:08 PM David C wrote: > Hi Marc > > I like that approach although I think I'd go in smaller weight increments. > > Still a bit confused by the behaviour I'm seeing, it looks like I've got > things weighted c

Re: [ceph-users] Luminous RGW errors at start

2018-09-04 Thread David Turner
I was confused what could be causing this until Janne's email. I think they're correct that the cluster is preventing pool creation due to too many PGs per OSD. Double check how many PGs you have in each pool and what your defaults are for that. On Mon, Sep 3, 2018 at 7:19 AM Janne Johansson

Re: [ceph-users] Ceph Luminous - journal setting

2018-09-04 Thread David Turner
Are you planning on using bluestore or filestore? The settings for filestore haven't changed. If you're planning to use bluestore there is a lot of documentation in the ceph docs as well as a wide history of questions like this on the ML. On Mon, Sep 3, 2018 at 5:24 AM M Ranga Swami Reddy

Re: [ceph-users] Luminous new OSD being over filled

2018-09-03 Thread David C
ph osd crush reweight osd.28 2 > sudo -u ceph ceph osd crush reweight osd.29 2 > > Etc etc > > > -Original Message- > From: David C [mailto:dcsysengin...@gmail.com] > Sent: maandag 3 september 2018 14:34 > To: ceph-users > Subject: [ceph-users] Luminous

[ceph-users] Luminous new OSD being over filled

2018-09-03 Thread David C
this one with a lower crush weight and then continue adding the drives and then eventually even out the crush weights? Thanks David ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Re: [ceph-users] Luminous missing osd_backfill_full_ratio

2018-09-03 Thread David C
if you delay the "...require-osd-release luminous" for whatever reason as it appears to leave you with no backfillfull limit Still having a bit of an issue with new OSDs over filling but will start a new thread for that Cheers, On Thu, Aug 30, 2018 at 10:34 PM David Turner wrote: >

Re: [ceph-users] "no valid command found" when running "ceph-deploy osd create"

2018-09-02 Thread David Wahler
On Sun, Sep 2, 2018 at 1:31 PM Alfredo Deza wrote: > > On Sun, Sep 2, 2018 at 12:00 PM, David Wahler wrote: > > Ah, ceph-volume.log pointed out the actual problem: > > > > RuntimeError: Cannot use device (/dev/storage/bluestore). A vg/lv path > > or an existing de

Re: [ceph-users] Help Basically..

2018-09-02 Thread David Turner
Agreed on not going the disks until your cluster is healthy again. Making them out and seeing how healthy you can get in the meantime is a good idea. On Sun, Sep 2, 2018, 1:18 PM Ronny Aasen wrote: > On 02.09.2018 17:12, Lee wrote: > > Should I just out the OSD's first or completely zap them

Re: [ceph-users] "no valid command found" when running "ceph-deploy osd create"

2018-09-02 Thread David Wahler
the ceph-deploy logs are a bit confusing. I submitted a PR to add a brief note to the quick-start guide, in case anyone else makes the same mistake: https://github.com/ceph/ceph/pull/23879 Thanks for the assistance! -- David On Sun, Sep 2, 2018 at 7:44 AM Alfredo Deza wrote: > > There shoul

Re: [ceph-users] Help Basically..

2018-09-02 Thread David Turner
astien-han.fr/blog/2014/11/27/ceph-recover-osds-after-ssd-journal-failure/ > > Which they all started without a problem. > > > On Sun, 2 Sep 2018 at 15:43, David Turner wrote: > >> It looks like osds on the first failed node are having problems. What >> commands did

<    1   2   3   4   5   6   7   8   9   10   >