Re: [ceph-users] Jewel + kernel 4.4 Massive performance regression (-50%)

2016-07-25 Thread Yoann Moulin
Hello Mark, > FWIW, on CentOS7 I actually saw a performance increase when upgrading from the > stock 3.10 kernel to 4.4.5 with Intel P3700 NVMe devices. I was encountering > some kind of strange concurrency/locking issues at the driver level that 4.4.5 > resolved. I think your best bet is to try

Re: [ceph-users] Jewel + kernel 4.4 Massive performance regression (-50%)

2016-07-25 Thread Yoann Moulin
Hello, > Am running ubuntu 16 with kernel 4.4-0.31-generic and my speed are similar. do you have journal on disk too ? > I did tests on ubuntu 14 and Ubuntu 16 and the speed is similar. I have around > 80-90MB/s of OSD speeds in both operating systems ok and could you do bench with kernel 4.2 ?

Re: [ceph-users] Unknown error (95->500) when creating buckets or putting files to RGW after upgrade from Infernalis to Jewel

2016-07-25 Thread nick
Hey Maciej, I compared the output of your commands with the output on our cluster and they are the same. So I do not see any problems on that site. After that I googled for the warning you get in the debug log: """ WARNING: set_req_state_err err_no=95 resorting to 500 """ I found some reports ab

Re: [ceph-users] How to remove OSD in JEWEL on Centos7

2016-07-25 Thread Stefan Lissmats
Hello! For me it works with something like systemctl stop ceph-osd@3 If you have autocompletion installed you should also be able to tab-complete after "systemctl stop" to list available services and for me it shows the mon and mds service too besides all other system services. Hope this helps.

Re: [ceph-users] 答复: 答复: 答复: how to transfer ceph cluster from the old network-and-hosts to a new one

2016-07-25 Thread Владимир Дробышевский
Hi! As I can see from here: osdmap e99: 9 osds: 9 up, 6 in; 64 remapped pgs that you have 3 OSDs down. Seems that you've tried to remove all of your ceph-node3 OSDs at once. This 53 active+remapped 48 active+undersized+degraded means that you have degraded objects: I would say that

[ceph-users] 答复: 答复: 答复: how to transfer ceph cluster from the old network-and-hosts to a new one

2016-07-25 Thread 朱 彤
Now the service could be found, thanks. However, both ceph osd tree and ceph status show osd.3 is still up, although ceph status also shows "degraded, stuck unclean..." I think is because of clock skew on a second MON. Besides, systemctl status ceph-osd@3 and systemctl status ceph-osd@4 give the

Re: [ceph-users] 答复: 答复: how to transfer ceph cluster from the old network-and-hosts to a new one

2016-07-25 Thread Владимир Дробышевский
Hi! You should use ceph-osd@ as a service name, not ceph, and systemctl as a service control utility. For example, 'systemctl stop ceph-osd@3' Best regards, Vladimir С уважением, Дробышевский Владимир Компания "АйТи Город" +7 343 192 ICQ# - 1921011 Аппаратное и программное обеспечение

Re: [ceph-users] Backfilling pgs not making progress

2016-07-25 Thread Brian Felton
Sam, I cranked up the logging on the backfill target (osd 580 on node 07) and the acting primary for the pg (453 on node 08, for what it's worth). The logs from the primary are very large, so pardon the tarballs. PG Primary Logs: https://www.dropbox.com/s/ipjobn2i5ban9km/backfill-primary-log.tgz

[ceph-users] How to remove OSD in JEWEL on Centos7

2016-07-25 Thread 朱 彤
Hi all, The first step is to mark an OSD down, in Hammer one can achieve this in this way: ceph osd out osd.3 /etc/init.d/ceph stop osd.3 However, in Jewel, package like ceph-10.2.2-0.el7.x86_64 no longer offers /etc/init.d/ceph, thus making it "unable to stop an OSD". I noticed the package of

Re: [ceph-users] cephfs failed to rdlock, waiting

2016-07-25 Thread Oliver Dzombic
Hi Greg, i switched the cache tier to forward, and began to evit everything. I restarted the mds, it was switching to another node. Still the same issue... So how can it be a pg full issue this way ? cluster a8171427-141c-4766-9e0f-533d86dd4ef8 health HEALTH_OK monmap e1: 3 mons

Re: [ceph-users] cephfs failed to rdlock, waiting

2016-07-25 Thread Oliver Dzombic
Hi Greg, well so what would be the next step to solve that ? For some reason its working like hell with the cache: cache io 500 MB/s evict, 627 op/s promote, 3 PG(s) evicting even there are no changes and no nothing. Where is this pressure coming from ? -- Mit freundlichen Gruessen / Best

Re: [ceph-users] cephfs failed to rdlock, waiting

2016-07-25 Thread Gregory Farnum
Yep, that seems more likely than anything else — there are no other running external ops to hold up a read lock, and if restarting the MDS isn't fixing it, then it's permanent state. So, RADOS. On Mon, Jul 25, 2016 at 7:53 PM, Oliver Dzombic wrote: > Hi Greg, > > > I can see that sometimes its sh

Re: [ceph-users] cephfs failed to rdlock, waiting

2016-07-25 Thread Oliver Dzombic
Hi Greg, I can see that sometimes its showing an evict (full) cluster a8171427-141c-4766-9e0f-533d86dd4ef8 health HEALTH_WARN noscrub,nodeep-scrub,sortbitwise flag(s) set monmap e1: 3 mons at {cephmon1=10.0.0.11:6789/0,cephmon2=10.0.0.12:6789/0,cephmon3=10.0.0.13:6789/

Re: [ceph-users] cephfs failed to rdlock, waiting

2016-07-25 Thread Gregory Farnum
On Mon, Jul 25, 2016 at 7:38 PM, Oliver Dzombic wrote: > Hi, > > currently some productive stuff is down, because it can not be accessed > through cephfs. > > Client server restart, did not help. > Cluster restart, did not help. > > Only ONE directory inside cephfs has this issue. > > All other di

Re: [ceph-users] mon_osd_nearfull_ratio (unchangeable) ?

2016-07-25 Thread Brad Hubbard
On Tue, Jul 26, 2016 at 12:16:35PM +1000, Goncalo Borges wrote: > Hi Brad > > Thanks for replying. > > Answers inline. > > > > > I am a bit confused about the 'unchachable' message we get in Jewel 10.2.2 > > > when I try to change some cluster configs. > > > > > > For example: > > > > > > 1./

[ceph-users] cephfs failed to rdlock, waiting

2016-07-25 Thread Oliver Dzombic
Hi, currently some productive stuff is down, because it can not be accessed through cephfs. Client server restart, did not help. Cluster restart, did not help. Only ONE directory inside cephfs has this issue. All other directories are working fine. MDS Server: Kernel 4.5.4 client server: Kern

Re: [ceph-users] mon_osd_nearfull_ratio (unchangeable) ?

2016-07-25 Thread Christian Balzer
Hello, On Tue, 26 Jul 2016 12:16:35 +1000 Goncalo Borges wrote: > Hi Brad > > Thanks for replying. > > Answers inline. > > > >> I am a bit confused about the 'unchachable' message we get in Jewel 10.2.2 > >> when I try to change some cluster configs. > >> > >> For example: > >> > >> 1./ if I

Re: [ceph-users] mon_osd_nearfull_ratio (unchangeable) ?

2016-07-25 Thread Goncalo Borges
Hi Brad Thanks for replying. Answers inline. I am a bit confused about the 'unchachable' message we get in Jewel 10.2.2 when I try to change some cluster configs. For example: 1./ if I try to change mon_osd_nearfull_ratio from 0.85 to 0.90, I get # ceph tell mon.* injectargs "--mon_osd

[ceph-users] 答复: 答复: how to transfer ceph cluster from the old network-and-hosts to a new one

2016-07-25 Thread 朱 彤
@Дробышевский thanks, I have tried, but # service ceph status Redirecting to /bin/systemctl status ceph.service ● ceph.service Loaded: not-found (Reason: No such file or directory) Active: inactive (dead) compared to another one having /etc/init.d/ceph and CentOS7 as well. $ service ceph s

Re: [ceph-users] mon_osd_nearfull_ratio (unchangeable) ?

2016-07-25 Thread Brad Hubbard
On Tue, Jul 26, 2016 at 11:01:49AM +1000, Goncalo Borges wrote: > Dear Cephers... Hi Goncalo, > > I am a bit confused about the 'unchachable' message we get in Jewel 10.2.2 > when I try to change some cluster configs. > > For example: > > 1./ if I try to change mon_osd_nearfull_ratio from 0.85

[ceph-users] mon_osd_nearfull_ratio (unchangeable) ?

2016-07-25 Thread Goncalo Borges
Dear Cephers... I am a bit confused about the 'unchachable' message we get in Jewel 10.2.2 when I try to change some cluster configs. For example: 1./ if I try to change mon_osd_nearfull_ratio from 0.85 to 0.90, I get # ceph tell mon.* injectargs "--mon_osd_nearfull_ratio 0.90" mon.rcc

Re: [ceph-users] Recovery stuck after adjusting to recent tunables

2016-07-25 Thread Brad Hubbard
On Tue, Jul 26, 2016 at 6:08 AM, Kostis Fardelas wrote: > Following up, I increased pg_num/pgp_num for my 3-replica pool to 128 These pg numbers seem low. Can you take a look at http://ceph.com/pgcalc/ and verify these values are appropriate for your environment and use case? I'd also take a go

[ceph-users] CephFS snapshot preferred behaviors

2016-07-25 Thread Gregory Farnum
All, I spent several days last week examining our current snapshot implementation and thinking about how it could be improved. As part of that ongoing effort, I'd love to know what user expectations are about behavior. (I'm going to open up a ceph-devel thread on the implementation details shortly,

Re: [ceph-users] Problem with RGW after update to Jewel

2016-07-25 Thread Frank Enderle
It most certainly looks very much like the same problem.. Is there a way to patch the configuration by hand to get the cluster back in a working state? -- From: Shilpa Manjarabad Jagannath Date: 25 July 2016 at 10:34:42 To: Frank Enderle

Re: [ceph-users] Problem with RGW after update to Jewel

2016-07-25 Thread Frank Enderle
Hi, here the outputs: radosgw-admin zone get --rgw-zone=default { "id": "default", "name": "default", "domain_root": ".rgw", "control_pool": ".rgw.control", "gc_pool": ".rgw.gc", "log_pool": ".log", "intent_log_pool": ".intent-log", "usage_log_pool": ".usage",

Re: [ceph-users] Monitors not reaching quorum

2016-07-25 Thread Sergio A. de Carvalho Jr.
Awesome, thanks so much, Joao. Here's the mon_status: https://gist.github.com/anonymous/2b80a9a75d134d9e539dfbc81615c055 I'm still trying to collect the logs, but while doing that I noticed that the log records are severely delayed compared to the system clock. For example, watching the logs with

Re: [ceph-users] Backfilling pgs not making progress

2016-07-25 Thread Samuel Just
If you don't have the orphaned file link, it's not the same bug. -Sam On Mon, Jul 25, 2016 at 12:55 PM, Brian Felton wrote: > Sam, > > I'm reviewing that thread now, but I'm not seeing a lot of overlap with my > cluster's situation. For one, I am unable to start either a repair or a > deep scrub

Re: [ceph-users] Backfilling pgs not making progress

2016-07-25 Thread Samuel Just
The next thing I'd want is for you to reproduce with debug osd = 20 debug filestore = 20 debug ms = 1 and post the file somewhere. -Sam On Mon, Jul 25, 2016 at 1:33 PM, Samuel Just wrote: > If you don't have the orphaned file link, it's not the same bug. > -Sam > > On Mon, Jul 25, 2016 at 12:55

Re: [ceph-users] Recovery stuck after adjusting to recent tunables

2016-07-25 Thread Kostis Fardelas
Following up, I increased pg_num/pgp_num for my 3-replica pool to 128 (being in argonaut tunables) and after a small recovery that followed, I switched to bobtail tunables. Remapping started and got stuck (!) again without any OSD down this time with 1 PG active+remapped. Tried restarting PG's OSDs

Re: [ceph-users] Backfilling pgs not making progress

2016-07-25 Thread Brian Felton
Sam, I'm reviewing that thread now, but I'm not seeing a lot of overlap with my cluster's situation. For one, I am unable to start either a repair or a deep scrub on any of the affected pgs. I've instructed all six of the pgs to scrub, deep-scrub, and repair, and the cluster has been gleefully i

Re: [ceph-users] Backfilling pgs not making progress

2016-07-25 Thread Samuel Just
You may have hit http://tracker.ceph.com/issues/14766. There was a thread on the list a while back about diagnosing and fixing it. -Sam On Mon, Jul 25, 2016 at 10:45 AM, Brian Felton wrote: > Greetings, > > Problem: After removing (out + crush remove + auth del + osd rm) three osds > on a single

Re: [ceph-users] Uncompactable Monitor Store at 69GB -- Re: Cluster in warn state, not sure what to do next.

2016-07-25 Thread Salwasser, Zac
Thank you for your help David. For reference, when I originally re-added the osds to the cluster I had set the weight to 0.0 and failed to set the host value. From: David Turner Date: Thursday, July 21, 2016 at 3:24 PM To: "Salwasser, Zac" , "ceph-users@lists.ceph.com" Cc: "Heller, Chris" S

[ceph-users] Backfilling pgs not making progress

2016-07-25 Thread Brian Felton
Greetings, Problem: After removing (out + crush remove + auth del + osd rm) three osds on a single host, I have six pgs that, after 10 days of recovery, are stuck in a state of active+undersized+degraded+remapped+backfilling. Cluster details: - 9 hosts (32 cores, 256 GB RAM, Ubuntu 14.04, 72 6TB

Re: [ceph-users] Monitors not reaching quorum

2016-07-25 Thread Joao Eduardo Luis
On 07/25/2016 05:55 PM, Sergio A. de Carvalho Jr. wrote: I just forced an NTP updated on all hosts to be sure it's down to clock skew. I also checked that hosts can reach all other hosts on port 6789. I then stopped monitor 0 (60z0m02) and started monitor 1 (60zxl02), but the 3 monitors left (1

Re: [ceph-users] Ceph performance calculator

2016-07-25 Thread Mark Nelson
Several years ago Mark Kampe proposed doing something like this. I was never totally convinced we could make something accurate enough quickly enough for it to be useful. If I were to attempt it, I would probably start out with a multiple regression approach based on seemingly important confi

Re: [ceph-users] Unknown error (95->500) when creating buckets or putting files to RGW after upgrade from Infernalis to Jewel

2016-07-25 Thread Naruszewicz, Maciej
Nick, Thanks a lot for you input so far. I re-ran the fix script from scratch and it turned out I made some mistakes in the process. I managed to run it correctly and now I am able to create buckets but I still can't upload anything. I looked for any issues in our configuration by searching at

Re: [ceph-users] Monitors not reaching quorum

2016-07-25 Thread Sergio A. de Carvalho Jr.
I just forced an NTP updated on all hosts to be sure it's down to clock skew. I also checked that hosts can reach all other hosts on port 6789. I then stopped monitor 0 (60z0m02) and started monitor 1 (60zxl02), but the 3 monitors left (1 - 60zxl02, 2 - 610wl02, 4 - 615yl02) were still having prob

Re: [ceph-users] CephFS Samba VFS RHEL packages

2016-07-25 Thread Bob R
Blair, Please do follow up with your findings. I've built samba v4.4.3 packages for centos 7, updated the kernel to v4.5.4 and tried a number of different configurations including kernel mounting at /cephfs and sharing /cephfs/dir without using ceph_vfs, and using ceph_vfs and targeting /dir in sm

Re: [ceph-users] 1 active+undersized+degraded+remapped+wait_backfill+backfill_toofull ???

2016-07-25 Thread M Ranga Swami Reddy
please share the "ceph osd tree" output. Is the osd.140 filled above 80%? if not, try to restart and see. Thanks Swami On Mon, Jul 25, 2016 at 9:45 PM, Vincent Godin wrote: > I restart osd.80 and till now : no bakfill_toofull anymore > > 2016-07-25 17:46 GMT+02:00 M Ranga Swami Reddy : >> >> can

Re: [ceph-users] Monitors not reaching quorum

2016-07-25 Thread Joao Eduardo Luis
On 07/25/2016 04:34 PM, Sergio A. de Carvalho Jr. wrote: Thanks, Joao. All monitors have the exact same mom map. I suspect you're right that there might be some communication problem though. I stopped monitor 1 (60zxl02), but the other 3 monitors still failed to reach a quorum. I could see moni

Re: [ceph-users] 1 active+undersized+degraded+remapped+wait_backfill+backfill_toofull ???

2016-07-25 Thread Vincent Godin
I restart osd.80 and till now : no bakfill_toofull anymore 2016-07-25 17:46 GMT+02:00 M Ranga Swami Reddy : > can you restart osd.80 and check see, if the recovery procced? > > Thanks > Swami > > On Mon, Jul 25, 2016 at 9:05 PM, Vincent Godin > wrote: > > Hi, > > > > I'm facing this problem. The

Re: [ceph-users] Jewel + kernel 4.4 Massive performance regression (-50%)

2016-07-25 Thread Lomayani S. Laizer
Hello, Am running ubuntu 16 with kernel 4.4-0.31-generic and my speed are similar. I did tests on ubuntu 14 and Ubuntu 16 and the speed is similar. I have around 80-90MB/s of OSD speeds in both operating systems Only issue am observing now with ubuntu 16 is sometime osd fails on rebooting until i

Re: [ceph-users] Jewel + kernel 4.4 Massive performance regression (-50%)

2016-07-25 Thread Mark Nelson
Hi Yoann, FWIW, on CentOS7 I actually saw a performance increase when upgrading from the stock 3.10 kernel to 4.4.5 with Intel P3700 NVMe devices. I was encountering some kind of strange concurrency/locking issues at the driver level that 4.4.5 resolved. I think your best bet is to try diff

Re: [ceph-users] 1 active+undersized+degraded+remapped+wait_backfill+backfill_toofull ???

2016-07-25 Thread M Ranga Swami Reddy
can you restart osd.80 and check see, if the recovery procced? Thanks Swami On Mon, Jul 25, 2016 at 9:05 PM, Vincent Godin wrote: > Hi, > > I'm facing this problem. The cluster is in Hammer 0.94.5 > > When i do a ceph health detail, i can see : > > pg 8.c1 is stuck unclean for 21691.555742, curr

[ceph-users] Jewel + kernel 4.4 Massive performance regression (-50%)

2016-07-25 Thread Yoann Moulin
Hello, (this is a repost, my previous message seems to be slipping under the radar) Does anyone get a similar behaviour to the one described below ? I found a big performance drop between kernel 3.13.0-88 (default kernel on Ubuntu Trusty 14.04) or kernel 4.2.0 and kernel 4.4.0.24.14 (default ker

[ceph-users] Fwd: 1 active+undersized+degraded+remapped+wait_backfill+backfill_toofull ???

2016-07-25 Thread Vincent Godin
The OSD 140 is 73.61% used and its backfill_full_ratio is 0.85 too -- Forwarded message -- From: Vincent Godin Date: 2016-07-25 17:35 GMT+02:00 Subject: 1 active+undersized+degraded+remapped+wait_backfill+backfill_toofull ??? To: ceph-users@lists.ceph.com Hi, I'm facing this p

[ceph-users] 1 active+undersized+degraded+remapped+wait_backfill+backfill_toofull ???

2016-07-25 Thread Vincent Godin
Hi, I'm facing this problem. The cluster is in Hammer 0.94.5 When i do a ceph health detail, i can see : pg 8.c1 is stuck unclean for 21691.555742, current state active+undersized+degraded+remapped+wait_backfill+backfill_toofull, last acting [140] pg 8.c1 is stuck undersized for 21327.027365, cu

Re: [ceph-users] Monitors not reaching quorum

2016-07-25 Thread Sergio A. de Carvalho Jr.
Thanks, Joao. All monitors have the exact same mom map. I suspect you're right that there might be some communication problem though. I stopped monitor 1 (60zxl02), but the other 3 monitors still failed to reach a quorum. I could see monitor 0 was still declaring victory but the others were alway

Re: [ceph-users] Monitors not reaching quorum

2016-07-25 Thread Sergio A. de Carvalho Jr.
We're having problems to start the 5th host (some BIOS problem, possibly), so I won't be able to recover its monitor any time soon. I knew having an even number of monitors wasn't ideal, and that's why I started 3 monitors first and waited until they reached quorum before starting the 4th monitor.

Re: [ceph-users] Monitors not reaching quorum

2016-07-25 Thread Joao Eduardo Luis
On 07/25/2016 03:41 PM, Sergio A. de Carvalho Jr. wrote: In the logs, there 2 monitors are constantly reporting that they won the leader election: 60z0m02 (monitor 0): 2016-07-25 14:31:11.644335 7f8760af7700 0 log_channel(cluster) log [INF] : mon.60z0m02@0 won leader election with quorum 0,2,4

Re: [ceph-users] Monitors not reaching quorum

2016-07-25 Thread Joao Eduardo Luis
On 07/25/2016 03:45 PM, Joshua M. Boniface wrote: My understanding is that you need an odd number of monitors to reach quorum. This seems to match what you're seeing: with 3, there is a definite leader, but with 4, there isn't. Have you tried starting both the 4th and 5th simultaneously and le

Re: [ceph-users] Monitors not reaching quorum

2016-07-25 Thread Joshua M. Boniface
My understanding is that you need an odd number of monitors to reach quorum. This seems to match what you're seeing: with 3, there is a definite leader, but with 4, there isn't. Have you tried starting both the 4th and 5th simultaneously and letting them both vote? -- Joshua M. Boniface Linux S

Re: [ceph-users] Monitors not reaching quorum

2016-07-25 Thread Sergio A. de Carvalho Jr.
In the logs, there 2 monitors are constantly reporting that they won the leader election: 60z0m02 (monitor 0): 2016-07-25 14:31:11.644335 7f8760af7700 0 log_channel(cluster) log [INF] : mon.60z0m02@0 won leader election with quorum 0,2,4 2016-07-25 14:31:44.521552 7f8760af7700 1 mon.60z0m02@0(le

[ceph-users] Monitors not reaching quorum

2016-07-25 Thread Sergio A. de Carvalho Jr.
Hi, I have a cluster of 5 hosts running Ceph 0.94.6 on CentOS 6.5. On each host, there is 1 monitor and 13 OSDs. We had an issue with the network and for some reason (which I still don't know why), the servers were restarted. One host is still down, but the monitors on the 4 remaining servers are

[ceph-users] RGW container deletion problem

2016-07-25 Thread Daniel Schneller
Hi! I created a bunch of test containers with some objects in them via RGW/Swift (Ubuntu, RGW via Apache, Ceph Hammer 0.94.1) Now I try to get rid of the test data. I manually staretd with one container: ~/rgwtest ➜ swift -v -V 1.0 -A http://localhost:8405/auth -U <...> -K <...> --insecure d

Re: [ceph-users] pgs stuck unclean after reweight

2016-07-25 Thread Christian Balzer
Hello, On Sun, 24 Jul 2016 00:54:09 + Goncalo Borges wrote: > Hi Christian > Thanks for the tips. > We do have monitoring in place but we are currently on a peak and the > occupancy increased tremendously in a couple of days time. > > I solved the problem of the stucked pgs by reweight (de

Re: [ceph-users] Infernalis -> Jewel, 10x+ RBD latency increase

2016-07-25 Thread Jason Dillaman
On Sun, Jul 24, 2016 at 5:47 AM, Martin Millnert wrote: > Are the replica writes of the primary OSD async/parallel? Yes, the primary OSD will asynchronously send the data to the replicas in parallel. -- Jason ___ ceph-users mailing list ceph-users@lis

Re: [ceph-users] 答复: how to transfer ceph cluster from the old network-and-hosts to a new one

2016-07-25 Thread Дробышевский , Владимир
Hi! CentOS 7 is using systemd, so you should stop osd with 'systemctl stop ceph-osd@' Best regards, Vladimir С уважением, Дробышевский Владимир Компания "АйТи Город" +7 343 192 Аппаратное и программное обеспечение IBM, Microsoft, Eset Поставка проектов "под ключ" Аутсорсинг ИТ-услуг 201

Re: [ceph-users] Pool full but empty fs AND Error EBUSY: pool 'pool_metadata_cephfs' is in use by CephFS

2016-07-25 Thread John Spray
On Mon, Jul 25, 2016 at 11:01 AM, kelvin woo wrote: > Hi All, > > # 1 # > I encountered 2 problems, I found that one of newly created ceph pool is > full that I do not know the reason. > > [root@ceph-adm ceph-cluster]# ceph -s > cluster 6dfd4779-3c75-49f4-bd47-6f4c31df0cb2 > health HEALTH

Re: [ceph-users] S3 API - Canonical user ID

2016-07-25 Thread Victor Efimov
Opened ticket in the bug tracker http://tracker.ceph.com/issues/16806 2016-07-19 1:50 GMT+03:00 Victor Efimov : > 2016-07-19 1:21 GMT+03:00 Robin H. Johnson : >> On Mon, Jul 18, 2016 at 10:48:16AM +0300, Victor Efimov wrote: >>> >> xmlns="http://s3.amazonaws.com/doc/2006-03-01/";>someownerSOMEOWNE

[ceph-users] Pool full but empty fs AND Error EBUSY: pool 'pool_metadata_cephfs' is in use by CephFS

2016-07-25 Thread kelvin woo
Hi All, # 1 # I encountered 2 problems, I found that one of newly created ceph pool is full that I do not know the reason. [root@ceph-adm ceph-cluster]# ceph -s cluster 6dfd4779-3c75-49f4-bd47-6f4c31df0cb2 health HEALTH_WARN pool 'pool_cephfs' is full monmap e1: 3 mons a

Re: [ceph-users] Try to install ceph hammer on CentOS7

2016-07-25 Thread Manuel Lausch
Hi, Thanks for your help. I found the failure. Via puppet I configured a versionlock. There I had a wrong version-epoch configured. Regards, Manuel Am 23.07.2016 um 05:11 schrieb Brad Hubbard: On Sat, Jul 23, 2016 at 1:41 AM, Ruben Kerkhof wrote: Please keep the mailing list on the CC. On

[ceph-users] 答复: how to transfer ceph cluster from the old network-and-hosts to a new one

2016-07-25 Thread 朱 彤
@Henrik Korkuc thanks for the tip, operating on it. In order to stop OSD, i need to run /etc/init.d/ceph stop osd.num, but just noticed /etc/init.d/ceph is missing. No such directory or file. I used ceph-deploy to install the cluster on Centos 7. Any idea? Thanks! _

Re: [ceph-users] Problem with RGW after update to Jewel

2016-07-25 Thread Shilpa Manjarabad Jagannath
- Original Message - > From: "Frank Enderle" > To: ceph-users@lists.ceph.com > Sent: Monday, July 25, 2016 1:28:10 AM > Subject: [ceph-users] Problem with RGW after update to Jewel > > Hi, > > a while ago I updated a cluster from Infernalis to Jewel. After the update > some problems occ

Re: [ceph-users] how to transfer ceph cluster from the old network-and-hosts to a new one

2016-07-25 Thread Henrik Korkuc
On 16-07-25 10:55, 朱 彤 wrote: Hi all, I m looking for a method to transfer ceph cluster. Now the cluster is located in network1 that has hosts A, B, C... And the target is to transfer it to network2 that has hosts a,b,c... What I can think of, is adding hosts a, b, c into the current clu

[ceph-users] how to transfer ceph cluster from the old network-and-hosts to a new one

2016-07-25 Thread 朱 彤
Hi all, I m looking for a method to transfer ceph cluster. Now the cluster is located in network1 that has hosts A, B, C... And the target is to transfer it to network2 that has hosts a,b,c... What I can think of, is adding hosts a, b, c into the current cluster like adding OSD and MON. Th

Re: [ceph-users] Problem with RGW after update to Jewel

2016-07-25 Thread Wido den Hollander
> Op 24 juli 2016 om 21:58 schreef Frank Enderle : > > > Hi, > > a while ago I updated a cluster from Infernalis to Jewel. After the update > some problems occured, which I fixed (I had to create some additional pool > which I was helped with in the IRC channel) - so the cluster now ran fine

Re: [ceph-users] change of dns names and IP addresses of cluster members

2016-07-25 Thread Wido den Hollander
> Op 22 juli 2016 om 15:26 schreef Andrei Mikhailovsky : > > > Hi Henrik, > > Many thanks for your answer. > > What settings in the ceph.conf are you referring to? These: > > mon_initial_members = > mon_host = > only mon_host is required. > I was under the impression that mon_initial_