Re: [ceph-users] cephfs-data-scan tool

2018-09-27 Thread John Spray
On Thu, Sep 27, 2018 at 11:34 AM Sergey Malinin wrote: > > Can such behaviour be related to data pool cache tiering? Yes -- if there's a cache tier in use then deletions in the base pool can be delayed and then happen later when the cache entries get expired. You may find that for a full scan

Re: [ceph-users] cephfs-data-scan tool

2018-09-27 Thread Sergey Malinin
> > On 27.09.2018, at 15:04, John Spray wrote: > > On Thu, Sep 27, 2018 at 11:34 AM Sergey Malinin wrote: >> >> Can such behaviour be related to data pool cache tiering? > > Yes -- if there's a cache tier in use then deletions in the base pool > can be delayed and then happen later when the

[ceph-users] Mimic cluster is offline and not healing

2018-09-27 Thread by morphin
Hello, I am writing this e-mail about an incident that has started last weekend. There seems to something wrong with my e-mail. Some of my e-mails did not reach-out. So I decided to start an new thread here and start from begining. One can find the email related e-mail thread

Re: [ceph-users] Mimic cluster is offline and not healing

2018-09-27 Thread Stefan Kooman
Quoting by morphin (morphinwith...@gmail.com): > After 72 hours I believe we may hit a bug. Any help would be greatly > appreciated. Is it feasible for you to stop all client IO to the Ceph cluster? At least until it stabilizes again. "ceph osd pause" would do the trick (ceph osd unpause would

Re: [ceph-users] Mimic cluster is offline and not healing

2018-09-27 Thread by morphin
I should not have client I/O right now. All of my VMs are down right now. There is only a single pool. Here is my crush map: https://paste.ubuntu.com/p/Z9G5hSdqCR/ Cluster does not recover. After starting OSDs with the specified flags, OSD up count drops from 168 to 50 with in 24 hours. Stefan

[ceph-users] Cephfs new file in ganesha mount Input/output error

2018-09-27 Thread Marc Roos
If I add on one client a file to the cephfs, that is exported via ganesha and nfs mounted somewhere else. I can see it in the dir listing on the other nfs client. But trying to read it gives an Input/output error. Other files (older ones in the same dir I can read) Anyone had this also?

[ceph-users] [CEPH]-[RADOS] Deduplication feature status

2018-09-27 Thread Gaël THEROND
Hi folks! As I’ll soon start to work on a new really large an distributed CEPH project for cold data storage, I’m checking out a few features availability and status, with the need for deduplication among them. I found out an interesting video about that from Cephalocon APAC 2018 and a seven

[ceph-users] slow export of cephfs through samba

2018-09-27 Thread Chad W Seys
Hi all, I am exporting cephfs using samba. It is much slower over samba than direct. Anyone know how to speed it up? Benchmarked using bonnie++ 5 times either directly to cephfs mounted by kernel (v4.18.6) module: bonnie++ -> kcephfs or through a cifs kernel-module-mounted (protocol

Re: [ceph-users] [CEPH]-[RADOS] Deduplication feature status

2018-09-27 Thread ceph
As of today, there is no such feature in Ceph Best regards, On 09/27/2018 04:34 PM, Gaël THEROND wrote: > Hi folks! > > As I’ll soon start to work on a new really large an distributed CEPH > project for cold data storage, I’m checking out a few features availability > and status, with the need

[ceph-users] CRUSH puzzle: step weighted-take

2018-09-27 Thread Dan van der Ster
Dear Ceph friends, I have a CRUSH data migration puzzle and wondered if someone could think of a clever solution. Consider an osd tree like this: -2 4428.02979 room 0513-R-0050 -72911.81897 rack RA01 -4917.27899 rack RA05 -6917.25500

Re: [ceph-users] CRUSH puzzle: step weighted-take

2018-09-27 Thread Luis Periquito
I think your objective is to move the data without anyone else noticing. What I usually do is reduce the priority of the recovery process as much as possible. Do note this will make the recovery take a looong time, and will also make recovery from failures slow... ceph tell osd.* injectargs

Re: [ceph-users] Any backfill in our cluster makes the cluster unusable and takes forever

2018-09-27 Thread David Turner
I got pulled away from this for a while. The error in the log is "abort: Corruption: Snappy not supported or corrupted Snappy compressed block contents" and the OSD has 2 settings set to snappy by default, async_compressor_type and bluestore_compression_algorithm. Do either of these settings

Re: [ceph-users] Mimic cluster is offline and not healing

2018-09-27 Thread by morphin
I think I might find something. When I start an OSD its making High I/O around %95 and the other OSDs are also triggered and altogether they make same the I/O. This is true even if when I set noup flag. So all the OSDs are making high I/O when ever an OSD starts. I think this is too much. I have

Re: [ceph-users] Any backfill in our cluster makes the cluster unusable and takes forever

2018-09-27 Thread Pavan Rallabhandi
I see Filestore symbols on the stack, so the bluestore config doesn’t affect. And the top frame of the stack hints at a RocksDB issue, and there are a whole lot of these too: “2018-09-17 19:23:06.480258 7f1f3d2a7700 2 rocksdb:

[ceph-users] Is object name used by CRUSH algorithm?

2018-09-27 Thread Jin Mao
I am running luminous and the objects were copied from Isilon with a long and similar prefix in path like /dir1/dir2/dir3//mm/dd. The objects are copied to various buckets like bucket_MMDD/dir1/dir2/dir3//mm/dd. This setup minimize some internal code change when moving from NFS to

Re: [ceph-users] CRUSH puzzle: step weighted-take

2018-09-27 Thread Maged Mokhtar
On 27/09/18 17:18, Dan van der Ster wrote: Dear Ceph friends, I have a CRUSH data migration puzzle and wondered if someone could think of a clever solution. Consider an osd tree like this: -2 4428.02979 room 0513-R-0050 -72911.81897 rack RA01 -4

Re: [ceph-users] Mimic cluster is offline and not healing

2018-09-27 Thread by morphin
Good news... :) After I tried everything. I decide to re-create my MONs from OSD's and I used the script: https://paste.ubuntu.com/p/rNMPdMPhT5/ And it worked!!! I think when 2 server crashed and come back same time some how MON's confused and the maps just corrupted. After re-creation all the

Re: [ceph-users] ceph-ansible

2018-09-27 Thread solarflow99
Thanks guys, installing this package did the trick, it works now. On Mon, Sep 24, 2018 at 8:39 AM Ken Dreyer wrote: > Hi Alfredo, > > I've packaged the latest version in Fedora, but I didn't update EPEL. > I've submitted the update for EPEL now at >

Re: [ceph-users] CRUSH puzzle: step weighted-take

2018-09-27 Thread Goncalo Borges
Hi Dan Hope to find you ok. Here goes a suggestion from someone who has been sitting in the side line for the last 2 years but following stuff as much as possible Will weight set per pool help? This is only possible in luminous but according to the docs there is the possibility to adjust

Re: [ceph-users] Fwd: [Ceph-community] After Mimic upgrade OSD's stuck at booting.

2018-09-27 Thread Willem Jan Withagen
On 26/09/2018 12:41, Eugen Block wrote: Hi, I'm not sure how the recovery "still works" with the flag "norecover". Anyway, I think you should unset the flags norecover, nobackfill. Even if not all OSDs come back up you should allow the cluster to backfill PGs. Not sure, but unsetting

Re: [ceph-users] cephfs-data-scan tool

2018-09-27 Thread John Spray
On Thu, Sep 27, 2018 at 11:03 AM Sergey Malinin wrote: > > Hello, > Does anybody have experience with using cephfs-data-scan tool? > Questions I have are how long would it take to scan extents on filesystem > with 120M relatively small files? While running extents scan I noticed that > number

Re: [ceph-users] Cannot write to cephfs if some osd's are not available on the client network

2018-09-27 Thread John Spray
On Thu, Sep 27, 2018 at 10:16 AM Marc Roos wrote: > > > I have a test cluster and on a osd node I put a vm. The vm is using a > macvtap on the client network interface of the osd node. Making access > to local osd's impossible. > > the vm of course reports that it cannot access the local osd's.

Re: [ceph-users] cephfs-data-scan tool

2018-09-27 Thread Sergey Malinin
Can such behaviour be related to data pool cache tiering? > On 27.09.2018, at 13:14, Sergey Malinin wrote: > > I'm trying alternate metadata pool approach. I double checked that MDS > servers are down and both original and recovery fs are set not joinable. > > >> On 27.09.2018, at 13:10,

[ceph-users] qemu/rbd: threads vs native, performance tuning

2018-09-27 Thread Elias Abacioglu
Hi, I was reading this thread: http://lists.ceph.com/pipermail/ceph-users-ceph.com/2016-March/008486.html And I am trying to get better performance in my virtual machines. These are my RBD settings: "rbd_cache": "true", "rbd_cache_block_writes_upfront": "false",

[ceph-users] Cannot write to cephfs if some osd's are not available on the client network

2018-09-27 Thread Marc Roos
I have a test cluster and on a osd node I put a vm. The vm is using a macvtap on the client network interface of the osd node. Making access to local osd's impossible. the vm of course reports that it cannot access the local osd's. What I am getting is: - I cannot reboot this vm normally,

[ceph-users] cephfs-data-scan tool

2018-09-27 Thread Sergey Malinin
Hello, Does anybody have experience with using cephfs-data-scan tool? Questions I have are how long would it take to scan extents on filesystem with 120M relatively small files? While running extents scan I noticed that number of objects in data pool is decreasing over the time. Is that normal?

Re: [ceph-users] Cannot write to cephfs if some osd's are not available on the client network

2018-09-27 Thread Burkhard Linke
Hi, On 09/27/2018 11:15 AM, Marc Roos wrote: I have a test cluster and on a osd node I put a vm. The vm is using a macvtap on the client network interface of the osd node. Making access to local osd's impossible. the vm of course reports that it cannot access the local osd's. What I am

Re: [ceph-users] cephfs-data-scan tool

2018-09-27 Thread Sergey Malinin
I'm trying alternate metadata pool approach. I double checked that MDS servers are down and both original and recovery fs are set not joinable. > On 27.09.2018, at 13:10, John Spray wrote: > > On Thu, Sep 27, 2018 at 11:03 AM Sergey Malinin wrote: >> >> Hello, >> Does anybody have