Re: [ceph-users] Global, Synchronous Blocked Requests

2015-11-27 Thread Gregory Farnum
On Fri, Nov 27, 2015 at 10:24 PM, Brian Felton wrote: > Greetings Ceph Community, > > We are running a Hammer cluster (0.94.3-1) in production that recently > experienced asymptotic performance degradation. We've been migrating data > from an older non-Ceph cluster at a

Re: [ceph-users] MDS memory usage

2015-11-24 Thread Gregory Farnum
On Tue, Nov 24, 2015 at 10:26 PM, Mike Miller wrote: > Hi, > > in my cluster with 16 OSD daemons and more than 20 million files on cephfs, > the memory usage on MDS is around 16 GB. It seems that 'mds cache size' has > no real influence on the memory usage of the MDS. > >

Re: [ceph-users] MDS memory usage

2015-11-25 Thread Gregory Farnum
regards, > > Mike > > > On 11/25/15 12:51 PM, Gregory Farnum wrote: >> >> On Tue, Nov 24, 2015 at 10:26 PM, Mike Miller <millermike...@gmail.com> >> wrote: >>> >>> Hi, >>> >>> in my cluster with 16 OSD daemons and more

Re: [ceph-users] CRUSH Algorithm

2015-11-30 Thread Gregory Farnum
The code is in ceph/src/crush of the gut repo, but it's pretty opaque. If you go to the Ceph site and look through the pages there's one about "publications" (or maybe just documentation? I think publications) that hosts a paper on how CRUSH works. IIRC it's using the jenkins hash on the object

Re: [ceph-users] rbd_inst.create

2015-11-30 Thread Gregory Farnum
On Nov 27, 2015 3:34 AM, "NEVEU Stephane" wrote: > > Ok, I think I got it. It seems to come from here : > > tracker.ceph.com/issues/6047 > > > > I’m trying to snapshot an image while I previously made a snapshot of my pool… whereas it just works fine when using a

Re: [ceph-users] ceph infernalis pg creating forever

2015-11-20 Thread Gregory Farnum
This usually means your crush mapping for the pool in question is unsatisfiable. Check what the rule is doing. -Greg On Friday, November 20, 2015, German Anders wrote: > Hi all, I've finished the install of a new ceph cluster with infernalis > 9.2.0 release. But I'm

Re: [ceph-users] Objects per PG skew warning

2015-11-23 Thread Gregory Farnum
On Thu, Nov 19, 2015 at 8:56 PM, Richard Gray wrote: > Hi, > > Running 'health detail' on our Ceph cluster this morning, I notice a warning > about one of the pools having significantly more objects per placement group > than the cluster average. > > ceph> health detail

Re: [ceph-users] Ceph 0.94.5 with accelio

2015-11-23 Thread Gregory Farnum
On Mon, Nov 23, 2015 at 10:05 AM, German Anders wrote: > Hi all, > > I want to know if there's any improvement or update regarding ceph 0.94.5 > with accelio, I've an already configured cluster (with no data on it) and I > would like to know if there's a way to 'modify' the

Re: [ceph-users] op sequence

2015-11-23 Thread Gregory Farnum
On Mon, Nov 23, 2015 at 8:44 AM, louis wrote: > Hi, if I submit read or write io in a sequence from a ceph client, will > these sequence will be kept in osds side? Thanks Any writes from the same client, to the same object, will be ordered with respect to one another.

Re: [ceph-users] Fixing inconsistency

2015-11-23 Thread Gregory Farnum
On Wed, Nov 18, 2015 at 4:34 AM, Межов Игорь Александрович wrote: > Hi! > > As for my previous message, digging mailing list gave me only one method to > fix > inconsistency - truncate object files in a filesystem to a size, that they > have > in ceph metadata: > >

Re: [ceph-users] ceph-mon cpu 100%

2015-11-23 Thread Gregory Farnum
Yep. I think you can inject it into the running mons without restarting as well (injectargs). -Greg On Mon, Nov 23, 2015 at 11:46 AM, Yujian Peng wrote: > The mons in my production cluster have a very high cpu usage 100%. > I think it may be caused by the leveldb

Re: [ceph-users] Difference between step choose and step chooseleaf

2016-06-08 Thread Gregory Farnum
On Wed, Jun 8, 2016 at 8:22 AM, George Shuklin wrote: > Hello. > > Can someone help me to see difference between step choose and step > chooseleaf in CRUSH map? When you run "choose" on a CRUSH bucket type, it selects CRUSH bucket nodes of that type. If you run

Re: [ceph-users] ceph-fuse, fio largely better after migration Infernalis to Jewel, is my bench relevant?

2016-06-06 Thread Gregory Farnum
We had several metadata caching improvements in ceph-fuse recently which I think went in after Infernalis. That could explain it. -Greg On Monday, June 6, 2016, Francois Lafont wrote: > Hi, > > I have a little Ceph cluster in production with 5 cluster nodes and 2 > client

Re: [ceph-users] Blocked ops, OSD consuming memory, hammer

2016-05-24 Thread Gregory Farnum
On Tue, May 24, 2016 at 2:16 PM, Heath Albritton wrote: > Having some problems with my cluster. Wondering if I could get some > troubleshooting tips: > > Running hammer 0.94.5. Small cluster with cache tiering. 3 spinning > nodes and 3 SSD nodes. > > Lots of blocked ops.

Re: [ceph-users] "mount error 5 = Input/output error" with the CephFS file system from client node

2016-06-14 Thread Gregory Farnum
On Tue, Jun 14, 2016 at 4:29 AM, Rakesh Parkiti wrote: > Hello, > > Unable to mount the CephFS file system from client node with "mount error 5 > = Input/output error" > MDS was installed on a separate node. Ceph Cluster health is OK and mds > services are running.

Re: [ceph-users] CephFS in the wild

2016-06-05 Thread Gregory Farnum
On Wed, Jun 1, 2016 at 1:50 PM, Brady Deetz wrote: > Question: > I'm curious if there is anybody else out there running CephFS at the scale > I'm planning for. I'd like to know some of the issues you didn't expect that > I should be looking out for. I'd also like to simply see

Re: [ceph-users] Rebuilding/recreating CephFS journal?

2016-05-27 Thread Gregory Farnum
On Fri, May 27, 2016 at 9:44 AM, Stillwell, Bryan J wrote: > I have a Ceph cluster at home that I¹ve been running CephFS on for the > last few years. Recently my MDS server became damaged and while > attempting to fix it I believe I¹ve destroyed by CephFS journal

Re: [ceph-users] Rebuilding/recreating CephFS journal?

2016-05-27 Thread Gregory Farnum
On Fri, May 27, 2016 at 1:54 PM, Stillwell, Bryan J <bryan.stillw...@charter.com> wrote: > On 5/27/16, 11:27 AM, "Gregory Farnum" <gfar...@redhat.com> wrote: > >>On Fri, May 27, 2016 at 9:44 AM, Stillwell, Bryan J >><bryan.stillw...@charter.com> wrot

Re: [ceph-users] Rebuilding/recreating CephFS journal?

2016-05-27 Thread Gregory Farnum
What's the current full output of "ceph -s"? If you already had your MDS in damaged state, you might just need to mark it as repaired. That's a monitor command. On Fri, May 27, 2016 at 2:09 PM, Stillwell, Bryan J <bryan.stillw...@charter.com> wrote: > On 5/27/16, 3:01 PM, &qu

Re: [ceph-users] Rebuilding/recreating CephFS journal?

2016-05-27 Thread Gregory Farnum
epaired at this point. > > > > > > > On 5/27/16, 3:17 PM, "Gregory Farnum" <gfar...@redhat.com> wrote: > >>What's the current full output of "ceph -s"? >> >>If you already had your MDS in damaged state, you might just need

Re: [ceph-users] Retrieve mds sessions

2016-06-02 Thread Gregory Farnum
On Thu, Jun 2, 2016 at 6:27 AM, Antonios Matsoukas wrote: > I am trying to get a list of the sessions from the MDS but i receive the > following error: > > root@monitor1:/var/log/ceph# ceph daemon mds.storage1 session ls > admin_socket: exception getting command

Re: [ceph-users] Crashing OSDs (suicide timeout, following a single pool)

2016-06-02 Thread Gregory Farnum
very process described in this post from >> 04/09/2015: >> http://ceph-users.ceph.narkive.com/lwDkR2fZ/recovering-incomplete-pgs-with-ceph-objectstore-tool >> Hopefully it works in your case too and you can the cluster back to a state >> that you can make the CephFS directorie

Re: [ceph-users] Infernalis => Jewel: ceph-fuse regression concerning the automatic mount at boot?

2016-06-01 Thread Gregory Farnum
On Wed, Jun 1, 2016 at 10:23 AM, Francois Lafont wrote: > Hi, > > I have a Jewel Ceph cluster in OK state and I have a "ceph-fuse" Ubuntu > Trusty client with ceph Infernalis. The cephfs is mounted automatically > and perfectly during the boot via ceph-fuse and this line in

Re: [ceph-users] Crashing OSDs (suicide timeout, following a single pool)

2016-06-01 Thread Gregory Farnum
If that pool is your metadata pool, it looks at a quick glance like it's timing out somewhere while reading and building up the omap contents (ie, the contents of a directory). Which might make sense if, say, you have very fragmented leveldb stores combined with very large CephFS directories.

Re: [ceph-users] OSDs not coming up on one host

2016-06-16 Thread Gregory Farnum
re you need to provide monitor logs before anybody can help. If they aren't, you are skipping steps and need to include OSD logs and things. ;) -Greg > > Best regards, > Kostis > > On 17 April 2015 at 02:04, Gregory Farnum <g...@gregs42.com> wrote: >> The moni

Re: [ceph-users] Ceph file change monitor

2016-06-16 Thread Gregory Farnum
On Wed, Jun 15, 2016 at 5:19 AM, siva kumar <85s...@gmail.com> wrote: > Yes , We need to similar to inotify/fanotity . > > came through link > http://docs.ceph.com/docs/master/dev/osd_internals/watch_notify/?highlight=notify#watch-notify > > Just want to know if i can use this ? > > If yes means

Re: [ceph-users] Ceph Day Switzerland slides and video

2016-06-16 Thread Gregory Farnum
On Wed, Jun 15, 2016 at 11:30 AM, Dan van der Ster wrote: > Dear Ceph Community, > > Yesterday we had the pleasure of hosting Ceph Day Switzerland, and we > wanted to let you know that the slides and videos of most talks have > been posted online: > >

[ceph-users] cephfs snapshots

2016-06-22 Thread Gregory Farnum
> Snapshots would be excellent for a number of fairly obvious reasons. Are > any of the know issues with snapshots issues that result in the loss of > non-snapshot data or a cluster? > On Jun 22, 2016 2:16 PM, "Gregory Farnum" <gfar...@redhat.com> wrote: > >&

Re: [ceph-users] cephfs snapshots

2016-06-22 Thread Gregory Farnum
On Wednesday, June 22, 2016, Kenneth Waegeman wrote: > Hi all, > > In Jewel ceph fs snapshots are still experimental. Does someone has a clue > when this would become stable, or how experimental this is ? > We're not sure yet. Probably it will follow stable multi-MDS;

Re: [ceph-users] Issue while building Jewel on ARM

2016-06-20 Thread Gregory Farnum
On Mon, Jun 20, 2016 at 5:28 AM, Daleep Singh Bais wrote: > Dear All, > > I am getting below error message while trying to build Jewel on ARM. Any > help / suggestion will be appreciated. > > g++: error: unrecognized command line option '-momit-leaf-frame-pointer' > g++:

Re: [ceph-users] cluster down during backfilling, Jewel tunables and client IO optimisations

2016-06-20 Thread Gregory Farnum
On Mon, Jun 20, 2016 at 8:33 AM, Daniel Swarbrick wrote: > We have just updated our third cluster from Infernalis to Jewel, and are > experiencing similar issues. > > We run a number of KVM virtual machines (qemu 2.5) with RBD images, and > have seen a lot of

Re: [ceph-users] Mysterious cache-tier flushing behavior

2016-06-17 Thread Gregory Farnum
Sounds like you've got deleted objects in the cache tier getting flushed (i.e., deleted) in the base tier. -Greg On Thursday, June 16, 2016, Christian Balzer wrote: > > Hello devs and other sage(sic) people, > > Ceph 0.94.5, cache tier in writeback mode. > > As mentioned before,

Re: [ceph-users] Mysterious cache-tier flushing behavior

2016-06-17 Thread Gregory Farnum
dj.png > > I can assure you, in that cluster objects don't get deleted. > > Christian > > On Fri, 17 Jun 2016 08:57:31 -0700 Gregory Farnum wrote: > >> Sounds like you've got deleted objects in the cache tier getting flushed >> (i.e., deleted) in the base tier. >

Re: [ceph-users] where is the fsid field coming from in ceph -s ?

2016-01-11 Thread Gregory Farnum
On Sat, Jan 9, 2016 at 1:58 AM, Oliver Dzombic wrote: > Hi, > > fighting to add a new mon it somehow happend by mistake, that a new > cluster id got generated. > > So the output of "ceph -s" show a new cluster id. > > But the osd/mon are still running on the old cluster

Re: [ceph-users] cephfs - inconsistent nfs and samba directory listings

2016-01-13 Thread Gregory Farnum
On Wed, Jan 13, 2016 at 11:24 AM, Mike Carlson wrote: > Hello. > > Since we upgraded to Infernalis last, we have noticed a severe problem with > cephfs when we have it shared over Samba and NFS > > Directory listings are showing an inconsistent view of the files: > > > $ ls

Re: [ceph-users] ceph-fuse on Jessie not mounted at boot

2016-01-14 Thread Gregory Farnum
Try using "id=client.my_user". It's not taking daemonize arguments because auto-mount in fstab requires the use of CLI arguments (of which daemonize isn't a member), IIRC. -Greg On Wednesday, January 6, 2016, Florent B wrote: > Hi everyone, > > I have a problem with

Re: [ceph-users] where is the fsid field coming from in ceph -s ?

2016-01-14 Thread Gregory Farnum
t; > Anschrift: > > IP Interactive UG ( haftungsbeschraenkt ) > Zum Sonnenberg 1-3 > 63571 Gelnhausen > > HRB 93402 beim Amtsgericht Hanau > Geschäftsführung: Oliver Dzombic > > Steuer Nr.: 35 236 3622 1 > UST ID: DE274086107 > > > Am 11.01.20

Re: [ceph-users] where is the client

2016-01-14 Thread Gregory Farnum
There's not a great unified tracking solution, but newer MDS code has admin socket commands to dump client sessions. Look for those. This question is good for the user list, but if you can't send mail to dev lost you're probably using HTML email or something. vger.kernel.org has some pretty

Re: [ceph-users] CEPH Replication

2016-01-14 Thread Gregory Farnum
We went to 3 copies because 2 isn't safe enough for the default. With 3 copies and a properly configured system your data is approximately as safe as the data center it's in. With 2 copies the durability is a lot lower than that (two 9s versus four 9s or something). The actual safety numbers did

[ceph-users] CephFS

2016-01-18 Thread Gregory Farnum
On Sunday, January 17, 2016, James Gallagher > wrote: > Hi, > > I'm looking to implement the CephFS on my Firefly release (v0.80) with > an XFS native file system, but so far I'm having some difficulties.

Re: [ceph-users] Again - state of Ceph NVMe and SSDs

2016-01-18 Thread Gregory Farnum
On Sun, Jan 17, 2016 at 12:34 PM, Tyler Bishop wrote: > The changes you are looking for are coming from Sandisk in the ceph "Jewel" > release coming up. > > Based on benchmarks and testing, sandisk has really contributed heavily on > the tuning aspects and are

Re: [ceph-users] cephfs - inconsistent nfs and samba directory listings

2016-01-14 Thread Gregory Farnum
Thu, Jan 14, 2016 at 3:37 AM, Mike Carlson <m...@bayphoto.com> >> >> wrote: >> >> > Hey Greg, >> >> > >> >> > The inconsistent view is only over nfs/smb on top of our /ceph mount. >> >> > >> >> > When I look direct

Re: [ceph-users] Ceph and NFS

2016-01-18 Thread Gregory Farnum
On Mon, Jan 18, 2016 at 4:48 AM, Arthur Liu wrote: > > > On Mon, Jan 18, 2016 at 11:34 PM, Burkhard Linke > wrote: >> >> Hi, >> >> On 18.01.2016 10:36, david wrote: >>> >>> Hello All. >>> Does anyone provides Ceph

Re: [ceph-users] Observations after upgrading to latest Firefly (0.80.11)

2016-01-14 Thread Gregory Farnum
On Thu, Jan 14, 2016 at 12:50 AM, Kostis Fardelas wrote: > Hello cephers, > after being on 0.80.10 for a while, we upgraded to 0.80.11 and we > noticed the following things: > a. ~13% paxos refresh latency increase (from about 0.015 to 0.017 on average) > b. ~15% paxos commit

Re: [ceph-users] osd current.remove.me.somenumber

2016-06-27 Thread Gregory Farnum
On Sat, Jun 25, 2016 at 11:22 AM, Mike Miller wrote: > Hi, > > what is the meaning of the directory "current.remove.me.846930886" is > /var/lib/ceph/osd/ceph-14? If you're using btrfs, I believe that's a no-longer-required snapshot of the current state of the system. If

Re: [ceph-users] How to monitor health and connectivity of OSD

2016-02-08 Thread Gregory Farnum
On Mon, Feb 8, 2016 at 3:25 AM, Mariusz Gronczewski wrote: > Is there an equivalent of 'ceph health' but for OSD ? > > Like warning about slowness or troubles with communication between OSDs? > > I've spent good amount of time debugging what looked like stuck pgs

Re: [ceph-users] mds0: Client X failing to respond to capability release

2016-02-08 Thread Gregory Farnum
peedPartner GmbH >>>>>>>>> <m...@speedpartner.de> wrote: >>>>>>>>> Am 03.02.2016 um 12:11 schrieb Yan, Zheng: >>>>>>>>>>> On Feb 3, 2016, at 17:39, Michael Metz-Martini | SpeedPartner GmbH >>>>>>&g

Re: [ceph-users] Can't fix down+incomplete PG

2016-02-10 Thread Gregory Farnum
On Wed, Feb 10, 2016 at 5:52 AM, Scott Laird wrote: > Ah, I should have mentioned--size=3, min_size=1. > > I'm pretty sure that 'down_osds_we_would_probe' is the problem, but it's not > clear if there's a way to fix that. Marking OSDs lost is what's supposed to resolve that.

Re: [ceph-users] pg dump question

2016-02-04 Thread Gregory Farnum
On Thu, Feb 4, 2016 at 10:23 AM, WRIGHT, JON R (JON R) wrote: > New ceph user, so a basic question > > I have a newly setup Ceph cluster. Seems to be working ok. But . . . > > I'm looking at the output of ceph pg dump, and I see that in the osdstat > list at the bottom

Re: [ceph-users] Increasing time to save RGW objects

2016-02-08 Thread Gregory Farnum
On Mon, Feb 8, 2016 at 8:49 AM, Kris Jurka wrote: > > I've been testing the performance of ceph by storing objects through RGW. > This is on Debian with Hammer using 40 magnetic OSDs, 5 mons, and 4 RGW > instances. Initially the storage time was holding reasonably steady, but

Re: [ceph-users] plain upgrade hammer to infernalis?

2016-02-08 Thread Gregory Farnum
On Mon, Feb 8, 2016 at 10:00 AM, Dzianis Kahanovich wrote: > I want to know about plain (not systemd, no deployment tools, only own simple > "start-stop-daemon" scripts under Gentoo) upgrade hammer to infernalis and see > no recommendations. Can I simple node-by-node

Re: [ceph-users] why is there heavy read traffic during object delete?

2016-02-04 Thread Gregory Farnum
On Thu, Feb 4, 2016 at 4:37 PM, Stephen Lord wrote: > I setup a cephfs file system with a cache tier over an erasure coded tier as > an experiment: > > ceph osd erasure-code-profile set raid6 k=4 m=2 > ceph osd pool create cephfs-metadata 512 512 > ceph osd pool set

Re: [ceph-users] why is there heavy read traffic during object delete?

2016-02-04 Thread Gregory Farnum
On Thu, Feb 4, 2016 at 5:07 PM, Stephen Lord <steve.l...@quantum.com> wrote: > >> On Feb 4, 2016, at 6:51 PM, Gregory Farnum <gfar...@redhat.com> wrote: >> >> I presume we're doing reads in order to gather some object metadata >> from the cephfs-data pool

Re: [ceph-users] ceph 9.2.0 mds cluster went down and now constantly crashes with Floating point exception

2016-02-04 Thread Gregory Farnum
On Thu, Feb 4, 2016 at 1:42 AM, Kenneth Waegeman wrote: > Hi, > > Hi, we are running ceph 9.2.0. > Overnight, our ceph state went to 'mds mds03 is laggy' . When I checked the > logs, I saw this mds crashed with a stacktrace. I checked the other mdss, > and I saw the

Re: [ceph-users] lstat() hangs on single file

2016-02-11 Thread Gregory Farnum
On Thu, Feb 11, 2016 at 4:29 PM, Blade Doyle wrote: > > After several months of use without needing any administration at all, I > think I finally found something to debug. > > Attempting to "ls -l" within a directory on CephFS hangs - strace shows its > hanging on lstat():

Re: [ceph-users] Increasing time to save RGW objects

2016-02-09 Thread Gregory Farnum
On Tue, Feb 9, 2016 at 8:07 AM, Kris Jurka <ju...@ejurka.com> wrote: > > > On 2/8/2016 9:16 AM, Gregory Farnum wrote: >> >> On Mon, Feb 8, 2016 at 8:49 AM, Kris Jurka <ju...@ejurka.com> wrote: >>> >>> >>> I've been te

Re: [ceph-users] why is there heavy read traffic during object delete?

2016-02-11 Thread Gregory Farnum
bject promotion when deleting from cache pool > > Reviewed-by: Sage Weil <s...@redhat.com> > > > Is there any chance that I was basically seeing with the same thing from the > filesystem standpoint? > > Thanks > > Steve > >> On Feb 5, 2016, at 8:42 AM

Re: [ceph-users] why is there heavy read traffic during object delete?

2016-02-05 Thread Gregory Farnum
coded pool. > > Looks like I am a little too bleeding edge for this, or the contents of the > .ceph_ attribute are not an object_info_t ghobject_info_t You can get the EC stuff actually deleted by getting the cache pool to flush everything. That's discussed in the docs and in various mailing lis

Re: [ceph-users] Trying to understand the contents of .rgw.buckets.index

2016-01-29 Thread Gregory Farnum
On Fri, Jan 29, 2016 at 3:10 AM, Wido den Hollander wrote: > > > On 29-01-16 11:31, Micha Krause wrote: >> Hi, >> >> I'm having problems listing the contents of an s3 bucket with ~2M objects. >> >> I already found the new bucket index sharding feature, but I'm >> interested how

Re: [ceph-users] downloads.ceph.com no longer valid?

2016-01-27 Thread Gregory Farnum
Infrastructure guys say it's down and they are working on it. On Wed, Jan 27, 2016 at 11:01 AM, John Hogenmiller wrote: > I dug a bit more. > > download.ceph.com resolves (for me) to 173.236.253.173 and is not responding > to icmp, port 80, or 443 > >

Re: [ceph-users] CephFS - Trying to understand direct OSD connection to ceph-fuse cephfs clients

2016-02-01 Thread Gregory Farnum
On Sunday, January 31, 2016, Goncalo Borges wrote: > Dear CephFS experts... > > We are using Ceph and CephFS 9.2.0. CephFS clients are being mounted via > ceph-fuse. > > We recently noticed the firewall from certain CephFS clients dropping > connections with OSDs as

Re: [ceph-users] mds0: Client X failing to respond to capability release

2016-02-03 Thread Gregory Farnum
On Tue, Feb 2, 2016 at 10:09 PM, Michael Metz-Martini | SpeedPartner GmbH wrote: > Hi, > > we're experiencing some strange issues running ceph 0.87 in our, I > think, quite large cluster (taking number of objects as a measurement). > > mdsmap e721086: 1/1/1 up

Re: [ceph-users] hammer-0.94.5 + kernel-4.1.15 - cephfs stuck

2016-02-03 Thread Gregory Farnum
The quick and dirty cleanup is to restart the OSDs hosting those PGs. They might have gotten some stuck ops which didn't get woken up; a few bugs like that have gone by and are resolved in various stable branches (I'm not sure what release binaries they're in). On Wed, Feb 3, 2016 at 11:32 PM,

Re: [ceph-users] Need help to develop CEPH EC Plugin for array type of Erasure Code

2016-02-02 Thread Gregory Farnum
On Mon, Feb 1, 2016 at 7:33 AM, Syed Hussain wrote: > Hi, > > I've been working to develop a CEPH EC plgin for array type erasure code, > for example > RAID-DP (i.e RDP). Later I realized that I can't continue with (k+m) format > in CEPH as > like normal RS code if I want to

Re: [ceph-users] hammer-0.94.5 + kernel-4.1.15 - cephfs stuck

2016-02-03 Thread Gregory Farnum
On Wed, Feb 3, 2016 at 2:32 AM, Nikola Ciprich wrote: > Hello Gregory, > > in the meantime, I managed to break it further :( > > I tried getting rid of active+remapped pgs and got some undersized > instead.. nto sure whether this can be related.. > > anyways here's the

Re: [ceph-users] MDS: bad/negative dir size

2016-02-03 Thread Gregory Farnum
On Wed, Feb 3, 2016 at 3:16 AM, Yan, Zheng wrote: > >> On Feb 3, 2016, at 17:51, Markus Blank-Burian wrote: >> >> Hi, >> >> on ceph mds startup, I see the following two errors in the our logfiles >> (using ceph 9.2.0 and linux 4.4 cephfs kernel client): >>

Re: [ceph-users] Urgent help needed for ceph storage "mount error 5 = Input/output error"

2016-02-02 Thread Gregory Farnum
On Tue, Feb 2, 2016 at 1:10 PM, Zhao Xu wrote: > I am not lucky on the ceph-fuse > > [root@igc-head ~]# ceph-fuse -d -m igc-head,is1,i1,i2,i3:6789 /mnt/igcfs/ > 2016-02-03 04:55:08.756420 7fe3f7437780 0 ceph version 0.94.5 > (9764da52395923e0b32908d83a9f7304401fee43), process

Re: [ceph-users] Ceph Tech Talk - High-Performance Production Databases on Ceph

2016-01-29 Thread Gregory Farnum
This is super cool — thanks, Thorvald, for the realistic picture of how databases behave on rbd! On Thu, Jan 28, 2016 at 11:56 AM, Patrick McGarry wrote: > Hey cephers, > > Here are the links to both the video and the slides from the Ceph Tech > Talk today. Thanks again to

Re: [ceph-users] CephFS is not maintianing conistency

2016-02-01 Thread Gregory Farnum
On Monday, February 1, 2016, Mykola Dvornik wrote: > Hi guys, > > This is sort of rebuttal. > > I have a CephFS deployed and mounted on a couple of clients via ceph-fuse > (due to quota support and possibility to kill the ceph-fuse process to > avoid stale mounts). > >

Re: [ceph-users] downloads.ceph.com no longer valid?

2016-01-27 Thread Gregory Farnum
Nah, it's not hosted on Ceph. On Wed, Jan 27, 2016 at 1:39 PM, Tyler Bishop <tyler.bis...@beyondhosting.net> wrote: > tyte... ceph pool go rogue? > > - Original Message - > From: "Gregory Farnum" <gfar...@redhat.com> > To: "John Hogenmille

Re: [ceph-users] Rack weight imbalance

2016-02-22 Thread Gregory Farnum
On Mon, Feb 22, 2016 at 9:29 AM, George Mihaiescu wrote: > Hi, > > We have a fairly large Ceph cluster (3.2 PB) that we want to expand and we > would like to get your input on this. > > The current cluster has around 700 OSDs (4 TB and 6 TB) in three racks with > the largest

Re: [ceph-users] Incorrect output from ceph osd map command

2016-02-23 Thread Gregory Farnum
This is not a bug. The map command just says which PG/OSD an object maps to; it does not go out and query the osd to see if there actually is such an object. -Greg On Tuesday, February 23, 2016, Vickey Singh wrote: > Hello Guys > > I am getting wired output from osd

Re: [ceph-users] librados: how to get notified when a certain object is created

2016-02-23 Thread Gregory Farnum
On Saturday, February 20, 2016, Sorin Manolache wrote: > Hello, > > I can set a watch on an object in librados. Does this object have to exist > already at the moment I'm setting the watch on it? What happens if the > object does not exist? Is my watcher valid? Will I get

Re: [ceph-users] Incorrect output from ceph osd map command

2016-02-23 Thread Gregory Farnum
ct exists, you will need to use a RADOS client to fetch the object and see if it's there. "map" is a mapping command: given an object name, which PG/OSD does CRUSH map that name to? > > Need expert suggestion. > > > On Tue, Feb 23, 2016 at 7:20 PM, Gregory Farnum <gf

Re: [ceph-users] Infernalis, cephfs: difference between df and du

2016-01-20 Thread Gregory Farnum
On Wed, Jan 20, 2016 at 6:40 PM, Francois Lafont wrote: > Hi, > > On 19/01/2016 07:24, Adam Tygart wrote: >> It appears that with --apparent-size, du adds the "size" of the >> directories to the total as well. On most filesystems this is the >> block size, or the amount of

Re: [ceph-users] ceph fuse closing stale session while still operable

2016-01-21 Thread Gregory Farnum
On Thu, Jan 21, 2016 at 4:24 AM, Oliver Dzombic wrote: > Hi Greg, > > alright. > > After shutting down the whole cluster and start it with "none" as > authentication, i resettet the auth rights and restarted the whole > cluster again after setting back to cephx. > > Now it

Re: [ceph-users] cephfs triggers warnings "tar: file changed as we read it"

2016-01-21 Thread Gregory Farnum
On Fri, Jan 15, 2016 at 9:00 AM, HMLTH wrote: > Hello, > > I'm evaluating cephfs on a virtual machines cluster. I'm using Infernalis > (9.2.0) on debian Jessie as client and server. > > I'm trying to get some performance numbers on operations like tar/untar on > things like the

Re: [ceph-users] Infernalis, cephfs: difference between df and du

2016-01-21 Thread Gregory Farnum
On Thu, Jan 21, 2016 at 1:20 AM, HMLTH <hm...@t-hamel.fr> wrote: > > > > Gregory Farnum – Thu., 21. January 2016 4:02 >> >> On Wed, Jan 20, 2016 at 6:40 PM, Francois Lafont <flafdiv...@free.fr> >> wrote: >> > Hi, >> > &

Re: [ceph-users] cephfs triggers warnings "tar: file changed as we read it"

2016-01-22 Thread Gregory Farnum
On Thu, Jan 21, 2016 at 11:52 PM, Yan, Zheng <uker...@gmail.com> wrote: > On Fri, Jan 22, 2016 at 6:24 AM, Gregory Farnum <gfar...@redhat.com> wrote: >> On Fri, Jan 15, 2016 at 9:00 AM, HMLTH <hm...@t-hamel.fr> wrote: >>> Hello, >>> >>> I'

Re: [ceph-users] CephFS

2016-01-22 Thread Gregory Farnum
On Fri, Jan 22, 2016 at 9:19 AM, James Gallagher wrote: > Hi there, > > Got a quick question regarding the CephFileSystem. After implementing the > setup from the quick start guide and having a Admin-Node, Monitor/Metadata > Server, OSD0 and OSD1. If I were to

Re: [ceph-users] ceph fuse closing stale session while still operable

2016-01-22 Thread Gregory Farnum
What is the output of the objecter_requests command? It really looks to me like the writes aren't going out and you're backing up on memory, but I can't tell without that. Actually, please grab a dump of the perfcounters while you're at it, that will include info on dirty memory and bytes written

Re: [ceph-users] ceph fuse closing stale session while still operable

2016-01-22 Thread Gregory Farnum
On Fri, Jan 22, 2016 at 2:26 PM, Oliver Dzombic wrote: > Hi Greg, > > from the client the list is huge: > > Thats the situation while the dd's are stuck. > > [root@cn201 ~]# ceph daemon /var/run/ceph/ceph-client.admin.asok > objecter_requests > { > "ops": [ > {

Re: [ceph-users] ceph fuse closing stale session while still operable

2016-01-25 Thread Gregory Farnum
On Mon, Jan 25, 2016 at 3:58 PM, Oliver Dzombic wrote: > Hi, > > i switched now debugging to ms = 10 > > when starting the dd i can see in the logs of osd: > > 2016-01-26 00:47:16.530046 7f086f404700 1 -- 10.0.0.1:6806/49658 >> :/0 > pipe(0x1f83 sd=292 :6806 s=0 pgs=0

Re: [ceph-users] ceph fuse closing stale session while still operable

2016-01-20 Thread Gregory Farnum
On Wed, Jan 20, 2016 at 6:58 AM, Oliver Dzombic wrote: > Hi, > > i am testing on centos 6 x64 minimal install. > > i am mounting successfully: > > ceph-fuse -m 10.0.0.1:6789,10.0.0.2:6789,10.0.0.3:6789,10.0.0.4:6789 > /ceph-storage/ > > > [root@cn201 log]# df > Filesystem

Re: [ceph-users] ceph fuse closing stale session while still operable

2016-01-20 Thread Gregory Farnum
On Wed, Jan 20, 2016 at 4:03 PM, Oliver Dzombic wrote: > Hi Greg, > > thank you for your time! > > #ceph-s > >cluster > health HEALTH_WARN > 62 requests are blocked > 32 sec > noscrub,nodeep-scrub flag(s) set > monmap e9: 4 mons at >

Re: [ceph-users] CephFS: read-only access

2016-02-15 Thread Gregory Farnum
On February 15, 2016 at 1:35:45 AM, Burkhard Linke (burkhard.li...@computational.bio.uni-giessen.de) wrote: > Hi, > > I would like to provide access to a bunch of large files (bio sequence > databases) to our cloud users. Putting the files in a RBD instance > requires special care if several VMs

Re: [ceph-users] Recover unfound objects from crashed OSD's underlying filesystem

2016-02-17 Thread Gregory Farnum
On Wed, Feb 17, 2016 at 3:05 PM, Kostis Fardelas wrote: > Hello cephers, > due to an unfortunate sequence of events (disk crashes, network > problems), we are currently in a situation with one PG that reports > unfound objects. There is also an OSD which cannot start-up and >

Re: [ceph-users] Recover unfound objects from crashed OSD's underlying filesystem

2016-02-17 Thread Gregory Farnum
t it into a random OSD, and then let the cluster recover from that OSD. (Assuming all the data you need is there — just because you can see the files on disk doesn't mean all the separate metadata is available.) -Greg > > Regards, > Kostis > > > On 18 February 2016 at 01:22, Grego

Re: [ceph-users] Recover unfound objects from crashed OSD's underlying filesystem

2016-02-17 Thread Gregory Farnum
ile 3.5a9..export > d. start the osd > > Regards, > Kostis > > > On 18 February 2016 at 02:54, Gregory Farnum <gfar...@redhat.com > <javascript:;>> wrote: > > On Wed, Feb 17, 2016 at 4:44 PM, Kostis Fardelas <dante1...@gmail.com > <javascri

Re: [ceph-users] cephfs mmap performance?

2016-02-19 Thread Gregory Farnum
On Fri, Feb 19, 2016 at 6:23 AM, Jan Schermer wrote: > I don't think there's any point in MMAP-ing a virtual file. > And I'd be surprised if there weren't any bugs or performance issues... Yeah. The CephFS kclient supports mmap, but I'm not sure Apache will be happy with the

Re: [ceph-users] plain upgrade hammer to infernalis?

2016-02-19 Thread Gregory Farnum
On Fri, Feb 19, 2016 at 5:11 AM, Steffen Weißgerber <weissgerb...@ksnb.de> wrote: > > >>>> Gregory Farnum <gfar...@redhat.com> schrieb am Montag, 8. Februar 2016 um > 19:10: >> On Mon, Feb 8, 2016 at 10:00 AM, Dzianis Kahanovich >> <maha...@bspu.

Re: [ceph-users] lstat() hangs on single file

2016-02-13 Thread Gregory Farnum
On Sat, Feb 13, 2016 at 8:14 AM, Blade Doyle wrote: > Greg, Thats very useful info. I had not queried the admin sockets before > today, so I am learning new things! > > on the x86_64: mds, mon, and osd, and rbd + cephfs client > ceph version 0.94.5

Re: [ceph-users] cephx capabilities to forbid rbd creation

2016-03-15 Thread Gregory Farnum
On Tue, Mar 15, 2016 at 2:44 PM, Loris Cuoghi wrote: > So, one key per RBD. > Or, dynamically enable/disable access to each RBD in each hypervisor's key. > Uhm, something doesn't scale here. :P > (I wonder if there's any limit to a key's capabilities string...) > > But, as

Re: [ceph-users] [cephfs] About feature 'snapshot'

2016-03-18 Thread Gregory Farnum
e on? >> >> ceph mds set allow_new_snaps true --yes-i-really-mean-it >> >> John >> >> > Thank you >> > >> > 2016-03-17 21:41 GMT+08:00 Gregory Farnum <gfar...@redhat.com >> <javascript:_e(%7B%7D,'cvml','gfar...@redhat.com

Re: [ceph-users] DONTNEED fadvise flag

2016-03-18 Thread Gregory Farnum
On Wed, Mar 16, 2016 at 9:46 AM, Kenneth Waegeman wrote: > Hi all, > > Quick question: Does cephFS pass the fadvise DONTNEED flag and take it into > account? > I want to use the --drop-cache option of rsync 3.1.1 to not fill the cache > when rsyncing to cephFS It looks

Re: [ceph-users] data corruption with hammer

2016-03-18 Thread Gregory Farnum
This tracker ticket happened to go by my eyes today: http://tracker.ceph.com/issues/12814 . There isn't a lot of detail there but the headline matches. -Greg On Wed, Mar 16, 2016 at 2:02 AM, Nick Fisk wrote: > > >> -Original Message- >> From: ceph-users

Re: [ceph-users] Recovering a secondary replica from another secondary replica

2016-03-09 Thread Gregory Farnum
ng to allow recovery directly via replicas, we count on having PGs whose primary is different to distribute that work. :) -Greg > > 2016-03-10 0:42 GMT+03:00 Gregory Farnum <gfar...@redhat.com>: >> On Wed, Mar 9, 2016 at 2:21 AM, Александр Шишенко <gamepa...@gmail.com>

Re: [ceph-users] Recovering a secondary replica from another secondary replica

2016-03-09 Thread Gregory Farnum
On Wed, Mar 9, 2016 at 2:21 AM, Александр Шишенко wrote: > Hello, > > I have a development cluster of three OSD's. My aim is to make a > secondary replica recover from another secondary replica (not > primary). Is it possible to do so with minor changes to ceph-osd > source

Re: [ceph-users] CephFS question

2016-03-11 Thread Gregory Farnum
On Friday, March 11, 2016, Sándor Szombat wrote: > Hi guys! > > We use Ceph and we need a distributed storage cluster for our files. I > check CephFS but documentation says > we can only > use 1 MDS this time. >

Re: [ceph-users] ceph-disk from jewel has issues on redhat 7

2016-03-15 Thread Gregory Farnum
There's a ceph-disk suite from last August that Loïc set up, but based on the qa list it wasn't running for a while and isn't in great shape. :/ I know there are some CentOS7 boxes in the sepia lab but it might not be enough for a small and infrequently-run test to reliably get tested against

<    8   9   10   11   12   13   14   15   16   17   >