Re: [ceph-users] 13.2.4 odd memory leak?

2019-03-08 Thread Steffen Winther Sørensen
> On 8 Mar 2019, at 14.30, Mark Nelson wrote: > > > On 3/8/19 5:56 AM, Steffen Winther Sørensen wrote: >> >>> On 5 Mar 2019, at 10.02, Paul Emmerich >> <mailto:paul.emmer...@croit.io>> wrote: >>> >>> Yeah, there's a bug in 1

Re: [ceph-users] 13.2.4 odd memory leak?

2019-03-08 Thread Steffen Winther Sørensen
2019 at 9:00 AM Steffen Winther Sørensen > wrote: >> >> >> >> On 4 Mar 2019, at 16.09, Paul Emmerich wrote: >> >> Bloated to ~4 GB per OSD and you are on HDDs? >> >> Something like that yes. >> >> >> 13.2.3 backported the c

Re: [ceph-users] 13.2.4 odd memory leak?

2019-03-05 Thread Steffen Winther Sørensen
-- > Paul Emmerich > > Looking for help with your Ceph cluster? Contact us at https://croit.io > > croit GmbH > Freseniusstr. 31h > 81247 München > www.croit.io > Tel: +49 89 1896585 90 > > On Mon, Mar 4, 2019 at 3:55 PM Steffen Winther Sørensen > wrote: >&

[ceph-users] 13.2.4 odd memory leak?

2019-03-04 Thread Steffen Winther Sørensen
List Members, patched a centos 7 based cluster from 13.2.2 to 13.2.4 last monday, everything appeared working fine. Only this morning I found all OSDs in the cluster to be bloated in memory foot print, possible after weekend backup through MDS. Anyone else seeing possible memory leak in

Re: [ceph-users] Mimic 13.2.1 release date

2018-07-10 Thread Steffen Winther Sørensen
> On 9 Jul 2018, at 17.11, Wido den Hollander wrote: > > Hi, > > Is there a release date for Mimic 13.2.1 yet? > > There are a few issues which currently make deploying with Mimic 13.2.0 > a bit difficult, for example: > > - https://tracker.ceph.com/issues/24423 > -

Re: [ceph-users] FYI - Mimic segv in OSD

2018-07-09 Thread Steffen Winther Sørensen
> On 9 Jul 2018, at 15.49, John Spray wrote: > > On Mon, Jul 9, 2018 at 2:37 PM Steffen Winther Sørensen > wrote: >> >> Dunno if this has been seen before so just for info, 1 in 24 OSD just did >> this: >> >> Jul 9 15:13:35 n4 ceph-osd: *** Caugh

[ceph-users] FYI - Mimic segv in OSD

2018-07-09 Thread Steffen Winther Sørensen
Dunno if this has been seen before so just for info, 1 in 24 OSD just did this: Jul 9 15:13:35 n4 ceph-osd: *** Caught signal (Segmentation fault) ** Jul 9 15:13:35 n4 ceph-osd: in thread 7ff209282700 thread_name:msgr-worker-2 Jul 9 15:13:35 n4 kernel: msgr-worker-2[4697]: segfault at 0 ip

Re: [ceph-users] Performance tuning for SAN SSD config

2018-07-06 Thread Steffen Winther Sørensen
> On 6 Jul 2018, at 17.55, Matthew Stroud wrote: > > We have changed the IO scheduler to NOOP, which seems to yield the best > results. However, I haven’t look into messing around with tuned. Let me play > with that and see if I get different results. > > On 5 Jul 2018, at 16.51, Matthew

Re: [ceph-users] Spurious empty files in CephFS root pool when multiple pools associated

2018-07-06 Thread Steffen Winther Sørensen
> On 6 Jul 2018, at 14.45, Steffen Winther Sørensen wrote: > > > >> On 3 Jul 2018, at 15.36, Jesus Cea wrote: >> >> Signed PGP part >> On 03/07/18 15:09, Steffen Winther Sørensen wrote: >>> >>> >>>> On 3 Jul 2018,

Re: [ceph-users] Spurious empty files in CephFS root pool when multiple pools associated

2018-07-06 Thread Steffen Winther Sørensen
> On 3 Jul 2018, at 15.36, Jesus Cea wrote: > > Signed PGP part > On 03/07/18 15:09, Steffen Winther Sørensen wrote: >> >> >>> On 3 Jul 2018, at 12.53, Jesus Cea wrote: >>> >>> Hi there. >>> >>> I have an issue wit

Re: [ceph-users] Performance tuning for SAN SSD config

2018-07-06 Thread Steffen Winther Sørensen
> On 5 Jul 2018, at 16.51, Matthew Stroud wrote: > > Bump. I’m hoping I can get people more knowledgeable than me to take a look. > We back some of our ceph clusters with SAN SSD disk, particularly VSP G/F and > Purestorage. I’m curious what are some settings we should look into modifying >

Re: [ceph-users] Spurious empty files in CephFS root pool when multiple pools associated

2018-07-03 Thread Steffen Winther Sørensen
> On 3 Jul 2018, at 12.53, Jesus Cea wrote: > > Hi there. > > I have an issue with cephfs and multiple datapools inside. I have like > SIX datapools inside the cephfs, I control where files are stored using > xattrs in the directories. Couldn’t you just use 6xCephFS each w/metadata + data

[ceph-users] Recreating a purged OSD fails

2018-06-27 Thread Steffen Winther Sørensen
List, Had a failed disk behind an OSD in a Mimic Cluster 13.2.0, so I tried following the doc on removal of an OSD. I did: # ceph osd crush reweight osd.19 0 waited for rebalancing to finish and cont.: # ceph osd out 19 # systemctl stop ceph-osd@19 # ceph osd purge 19 --yes-i-really-mean-it

Re: [ceph-users] Increase queue_depth in KVM

2018-06-26 Thread Steffen Winther Sørensen
> On 26 Jun 2018, at 14.04, Damian Dabrowski wrote: > > Hi Stefan, thanks for reply. > > Unfortunately it didn't work. > > disk config: > >discard='unmap'/> > > > >name='volumes-nvme/volume-ce247187-a625-49f1-bacd-fc03df215395'> > >

Re: [ceph-users] Ceph Mimic on CentOS 7.5 dependency issue (liboath)

2018-06-24 Thread Steffen Winther Sørensen
On 24 Jun 2018, at 06.57, Brad Hubbard wrote: > > As Brian pointed out > > # yum -y install > https://dl.fedoraproject.org/pub/epel/epel-release-latest-7.noarch.rpm > Yeap has worked fine for me in CentOS 7.5 w/EPEL

[ceph-users] Howto add another client user id to a cluster

2018-06-22 Thread Steffen Winther Sørensen
Anyone, We’ve ceph clients that we want to let mount two cephfs from each their own ceph clusters. Both cluster are standard created w/ceph-deploy and possible only has knowledge of each their client.admin. How could we allow a new client id to access the 2. cluster eg. as admin2? On ceph

Re: [ceph-users] Minimal MDS for CephFS on OSD hosts

2018-06-19 Thread Steffen Winther Sørensen
> Den 19. jun. 2018 kl. 16.50 skrev Webert de Souza Lima > : > > Keep in mind that the mds server is cpu-bound, so during heavy workloads it > will eat up CPU usage, so the OSD daemons can affect or be affected by the > MDS daemon. > But it does work well. We've been running a few clusters

[ceph-users] Mimic 13.2 - Segv in ceph-osd

2018-06-18 Thread Steffen Winther Sørensen
List, Just a heads up, I found an osd that did a segv on a CentOS 7.5 node: # ceph --version ceph version 13.2.0 (79a10589f1f80dfe21e8f9794365ed98143071c4) mimic (stable) # cat /etc/centos-release CentOS Linux release 7.5.1804 (Core) Jun 17 07:01:18 n3 ceph-osd: *** Caught signal

Re: [ceph-users] Ceph bonding vs separate provate public network

2018-06-12 Thread Steffen Winther Sørensen
> On 12 Jun 2018, at 14.00, Steven Vacaroaia wrote: > > Hi, > > I am designing a new ceph cluster and was wondering whether I should bond the > 10 GB adapters or use one for public one for private > > The advantage of bonding is simplicity and, maybe, performance > The catch though is that

[ceph-users] Debian GPG key for Luminous

2018-06-07 Thread Steffen Winther Sørensen
Community, Where would I find the GPG release key for Debian Luminous? as I’m getting: W: GPG error: http://download.ceph.com/debian-luminous stretch InRelease: The following signatures couldn't be verified because the public key is not available: NO_PUBKEY E84AC2C0460F3994 when attempting

Re: [ceph-users] Move data from Hammer to Mimic

2018-05-30 Thread Steffen Winther Sørensen
> On 29 May 2018, at 23.29, Gregory Farnum wrote: > > On Tue, May 29, 2018 at 3:59 AM Steffen Winther Sørensen <mailto:ste...@gmail.com>> wrote: > (ie. would Jewel be able to connect to both clusters)? > > Yes; that should work without any issues Great, that

[ceph-users] Move data from Hammer to Mimic

2018-05-29 Thread Steffen Winther Sørensen
List, Got an old Hammer Cluster where I would like to migrate it’s data (rbd images) to a newly installed Mimic Cluster. Would this be possible if I could upgrade the clients from Hammer to Jewel (ie. would Jewel be able to connect to both clusters)? /Steffen

Re: [ceph-users] Mimic EPERM doing rm pool

2018-05-29 Thread Steffen Winther Sørensen
> On 29 May 2018, at 10.08, Eugen Block wrote: > > Hi, > >> [root@n1 ~]# ceph osd pool rm mytestpool mytestpool --yes-i-really-mean-it >> Error EPERM: WARNING: this will *PERMANENTLY DESTROY* all data stored > > if the command you posted is complete then you forgot one "really" in the >

[ceph-users] Mimic EPERM doing rm pool

2018-05-29 Thread Steffen Winther Sørensen
List, I’ve just installed a new mimic cluster and wonder why I can’t remove a initial test pool like this: [root@n1 ~]# ceph -s cluster: id: 2284bf30-a27e-4543-af8f-b2726207762a health: HEALTH_OK services: mon: 3 daemons, quorum n1,n2,n3 mgr: n1.ceph(active), standbys:

[ceph-users] Hammer patching on Wheezy?

2017-06-28 Thread Steffen Winther Sørensen
Ceph users, Got a Hammer cluster installed on old debian wheezy (7.11) boxes (I know :) root@node4:~# dpkg -l | grep -i ceph ii ceph 0.94.9-1~bpo70+1 amd64 distributed storage and file system ii ceph-common

Re: [ceph-users] Hammer patch on Wheezy + CephFS leaking space?

2017-06-27 Thread Steffen Winther Sørensen
> On 27 Jun 2017, at 11.45, John Spray <jsp...@redhat.com> wrote: > > On Tue, Jun 27, 2017 at 9:52 AM, Steffen Winther Sørensen > <ste...@gmail.com <mailto:ste...@gmail.com>> wrote: >> Ceph users, >> >> Got an old Hammer CephFS ins

[ceph-users] Hammer patch on Wheezy + CephFS leaking space?

2017-06-27 Thread Steffen Winther Sørensen
Ceph users, Got an old Hammer CephFS installed on old debian wheezy (7.11) boxes (I know :) root@node4:~# dpkg -l | grep -i ceph ii ceph 0.94.9-1~bpo70+1 amd64 distributed storage and file system ii ceph-common

Re: [ceph-users] Hammer OSD crash during deep scrub

2016-02-16 Thread Steffen Winther Soerensen
Steffen Winther Soerensen <ceph.user@...> writes: > Looks like an IO error during read maybe, > only nothing logged in syslog messages at the time. :) but it was logged in syslog at the time: Feb 15 01:28:14 node2 kernel: cciss :46:00.0: cmd 88003a900280 has CHECK CONDITI

[ceph-users] Hammer OSD crash during deep scrub

2016-02-16 Thread Steffen Winther Soerensen
I've had few OSD crash from time to time, latest like this: --- begin dump of recent events --- -12> 2016-02-15 01:28:15.386412 7f29c8828700 1 -- 10.0.3.2:6819/448052 <== osd.17 10.0.3.1:0/6746 181211 osd_ping(ping e12542 stamp 2016-02-15 01:28:15.385759) v2 47+0+0 (1302847072 0

Re: [ceph-users] Hammer on Debian Wheezy not pulling in update +0.94.5

2016-02-16 Thread Steffen Winther Soerensen
Christian Balzer <chibi@...> writes: > > > Hello, > > On Tue, 16 Feb 2016 08:49:00 + (UTC) Steffen Winther Soerensen wrote: > > > Got few OSDs crash from time to time in my Hammer 0.94.5 cluster and > > it seems Hammer is at 0.94.7 > > Wh

[ceph-users] Hammer on Debian Wheezy not pulling in update +0.94.5

2016-02-16 Thread Steffen Winther Soerensen
Got few OSDs crash from time to time in my Hammer 0.94.5 cluster and it seems Hammer is at 0.94.7 but why doesn't My Debian Wheezy nodes that pull patches above 0.94.5? root@node2:~# apt-get update ... Hit http://ceph.com wheezy Release Hit

Re: [ceph-users] S3 RadosGW - Create bucket OP

2015-03-10 Thread Steffen Winther
Yehuda Sadeh-Weinraub yehuda@... writes: According to the api specified here http://docs.aws.amazon.com/AmazonS3/latest/API/RESTBucketPUT.html, there's no response expected. I can only assume that the application tries to decode the xml if xml content type is returned. Also what I hinted App

Re: [ceph-users] EC Pool and Cache Tier Tuning

2015-03-09 Thread Steffen Winther
Nick Fisk nick@... writes: My Ceph cluster comprises of 4 Nodes each with the following:- 10x 3TB WD Red Pro disks - EC pool k=3 m=3 (7200rpm) 2x S3700 100GB SSD's (20k Write IOPs) for HDD Journals 1x S3700 400GB SSD (35k Write IOPs) for cache tier - 3x replica If I have following 4x node

Re: [ceph-users] S3 RadosGW - Create bucket OP

2015-03-09 Thread Steffen Winther
Yehuda Sadeh-Weinraub yehuda@... writes: If you're using apache, then it filters out zero Content-Length. Nothing much radosgw can do about it. You can try using the radosgw civetweb frontend, see if it changes anything. Thanks, only no difference... Req: PUT /mssCl/ HTTP/1.1 Host:

Re: [ceph-users] S3 RadosGW - Create bucket OP

2015-03-09 Thread Steffen Winther
Steffen W Sørensen stefws@... writes: Response: HTTP/1.1 200 OK Date: Fri, 06 Mar 2015 10:41:14 GMT Server: Apache/2.2.22 (Fedora) Connection: close Transfer-Encoding: chunked Content-Type: application/xml This response makes the App say: S3.createBucket, class S3, code

Re: [ceph-users] Calamari build in vagrants

2015-02-16 Thread Steffen Winther
Dan Mick dmick@... writes: 0cbcfbaa791baa3ee25c4f1a135f005c1d568512 on the 1.2.3 branch has the change to yo 1.1.0. I've just cherry-picked that to v1.3 and master. Do you mean that you merged 1.2.3 into master and branch 1.3? BTW I managed to clone and built branch 1.2.3 in my vagrant env.

Re: [ceph-users] initially conf calamari to know about my Ceph cluster(s)

2015-02-16 Thread Steffen Winther
Steffen Winther ceph.user@... writes: Trying to figure out how to initially configure calamari clients to know about my Ceph Cluster(s) when such aint install through ceph.deploy but through Proxmox pveceph. Assume I possible need to copy some client admin keys and configure my MON hosts

Re: [ceph-users] Calamari build in vagrants

2015-02-15 Thread Steffen Winther
Dan Mick dmick@... writes: 0cbcfbaa791baa3ee25c4f1a135f005c1d568512 on the 1.2.3 branch has the change to yo 1.1.0. I've just cherry-picked that to v1.3 and master. Thanks, but sorry for my newbie question :) How/where do I find this ID from the 1.2.3 branch? on git I guess... will look

[ceph-users] initially conf calamari to know about my Ceph cluster(s)

2015-02-15 Thread Steffen Winther
Trying to figure out how to initially configure calamari clients to know about my Ceph Cluster(s) when such aint install through ceph.deploy but through Proxmox pveceph. Assume I possible need to copy some client admin keys and configure my MON hosts somehow, any pointers to doc on this?

Re: [ceph-users] Calamari build in vagrants

2015-02-12 Thread Steffen Winther
Steffen Winther ceph.user@... writes: Trying to build calamari rpm+deb packages following this guide: http://karan-mj.blogspot.fi/2014/09/ceph-calamari-survival-guide.html Server packages works fine, but fails in clients for: dashboard manage admin login due to: yo 1.1.0 seems needed