Re: [ceph-users] Possible way to clean up leaked multipart objects?

2017-08-31 Thread William Schroeder
David, We would love some testing of the tool. Are you set up to compile and deploy Ceph changes? If your situation is not related to the leaked multipart objects due to retries, it will let you know nothing was fixed. That is still a useful test. Another variation of multipart leak comes

Re: [ceph-users] luminous ceph-osd crash

2017-08-31 Thread Marcin Dulak
Hi, /var/log/ceph/ceph-osd.0.log is attached. My sdb is 128MB and sdc (journal) is 16MB: [root@server0 ~]# ceph-disk list /dev/dm-0 other, xfs, mounted on / /dev/dm-1 swap, swap /dev/sda : /dev/sda1 other, 0x83 /dev/sda2 other, xfs, mounted on /boot /dev/sda3 other, LVM2_member /dev/sdb :

Re: [ceph-users] Possible way to clean up leaked multipart objects?

2017-08-31 Thread David Turner
Jewel 10.2.7. I found a discrepancy in object counts for a multisite configuration and it's looking like it might be orphaned multipart files causing it. It doesn't look like this PR has received much attention. Is there anything I can do to help you with testing/confirming a use case for this

Re: [ceph-users] [rgw][s3] Object not in objects list

2017-08-31 Thread Stanley Zhang
Your bucket index got corrupted. I believe there is no easy way to restore the index other than downloading existing objects and re-upload them, correct me if anybody else know a better way. You can check out all your objects in that bucket with: rados -p .rgw.buckets ls | grep

[ceph-users] Object gateway and LDAP Auth

2017-08-31 Thread Josh
Hello! I've setup LDAP authentication on an object gateway and am attempting to create a bucket via s3 using python's boto3. It works fine using the access and secret key for a radosgw user, but access is denied using a token generated via radosgw-token with the LDAP user's credentials. The user

[ceph-users] (no subject)

2017-08-31 Thread Marc Roos
Should these messages not be gone in 12.2.0? 2017-08-31 20:49:33.500773 7f5aa1756d40 -1 WARNING: the following dangerous and experimental features are enabled: bluestore 2017-08-31 20:49:33.501026 7f5aa1756d40 -1 WARNING: the following dangerous and experimental features are enabled: bluestore

Re: [ceph-users] RGW Multisite metadata sync init

2017-08-31 Thread David Turner
All of the messages from sync error list are listed below. The number on the left is how many times the error message is found. 1811 "message": "failed to sync bucket instance: (16) Device or resource busy" 7 "message": "failed to sync bucket

Re: [ceph-users] Very slow start of osds after reboot

2017-08-31 Thread Hervé Ballans
Hi Piotr, Just to verify one point, how are connected your disks (physically), in a NON-RAID or RAID0 mode ? rv Le 31/08/2017 à 16:24, Piotr Dzionek a écrit : For a last 3 weeks I have been running latest LTS Luminous Ceph release on CentOS7. It started with 4th RC and now I have Stable

Re: [ceph-users] a metadata lost problem when mds breaks down

2017-08-31 Thread Sage Weil
On Thu, 31 Aug 2017, Mark Meyers wrote: > Hi: > > I encountered a metadata lost problem of files when testing the > outcome of ceph mds's unpredictable breakdown. > The test shell script is like(not real code): > > while i < 10 > do > touch $i >

Re: [ceph-users] Very slow start of osds after reboot

2017-08-31 Thread Dan van der Ster
Random theory... I just noticed that the ceph-osd's are listed twice [1] in the output of systemctl list-dependencies. Is that correct?!!! -- dan [1] > systemctl list-dependencies ... ● ├─ceph-mds.target ● ├─ceph-mon.target ● ├─ceph-osd.target ● │ ├─ceph-osd@48.service ● │ ├─ceph-osd@49.service

[ceph-users] a metadata lost problem when mds breaks down

2017-08-31 Thread Mark Meyers
Hi: I encountered a metadata lost problem of files when testing the outcome of ceph mds's unpredictable breakdown. The test shell script is like(not real code): while i < 10 do touch $i i++ done echo c >

Re: [ceph-users] Very slow start of osds after reboot

2017-08-31 Thread Dan van der Ster
Hi, I see the same with jewel on el7 -- it started one of the recent point releases around ~10.2.5, IIRC. Problem seems to be the same -- daemon is started before the osd is mounted... then the service waits several seconds before trying again. Aug 31 15:41:47 ceph-osd: 2017-08-31

Re: [ceph-users] Changing the failure domain

2017-08-31 Thread David Turner
How long are you seeing these blocked requests for? Initially or perpetually? Changing the failure domain causes all PGs to peer at the same time. This would be the cause if it happens really quickly. There is no way to avoid all of them peering while making a change like this. After that, It

Re: [ceph-users] Very slow start of osds after reboot

2017-08-31 Thread Sean Purdy
Datapoint: I have the same issue on 12.1.1, three nodes, 6 disks per node. On Thu, 31 Aug 2017, Piotr Dzionek said: > For a last 3 weeks I have been running latest LTS Luminous Ceph release on > CentOS7. It started with 4th RC and now I have Stable Release. > Cluster runs fine, however I noticed

Re: [ceph-users] where is a RBD in use

2017-08-31 Thread Maxime Guyot
Hi Götz, Something like "rbd status image-spec" usually works for me. Man page says: "Show the status of the image, including which clients have it open." I'll tell you which IPs have it open which should help you to track it down. Cheers, Maxime On Thu, 31 Aug 2017 at 16:26 Götz Reinicke

[ceph-users] where is a RBD in use

2017-08-31 Thread Götz Reinicke
Hi, Is it possible to see which clients are using an RBD? … I found an RBD in one of my pools but cant remember if I ever use / mounted it to a client. Thx for feedback ! Regards . Götz ___ ceph-users mailing list ceph-users@lists.ceph.com

[ceph-users] Very slow start of osds after reboot

2017-08-31 Thread Piotr Dzionek
Hi, For a last 3 weeks I have been running latest LTS Luminous Ceph release on CentOS7. It started with 4th RC and now I have Stable Release. Cluster runs fine, however I noticed that if I do a reboot of one the nodes, it takes a really long time for cluster to be in ok status. Osds are

Re: [ceph-users] luminous ceph-osd crash

2017-08-31 Thread Sage Weil
Hi Marcin, Can you reproduce the crash with 'debug bluestore = 20' set, and then ceph-post-file /var/log/ceph/ceph-osd.0.log? My guess is that we're not handling a very small device properly? sage On Thu, 31 Aug 2017, Marcin Dulak wrote: > Hi, > > I have a virtual CentOS 7.3 test setup at:

[ceph-users] Changing the failure domain

2017-08-31 Thread Laszlo Budai
Dear all! In our Hammer cluster we are planning to switch our failure domain from host to chassis. We have performed some simulations, and regardless of the settings we have used some slow requests have appeared all the time. we had the the following settings: "osd_max_backfills": "1",

Re: [ceph-users] Ceph Day Netherlands: 20-09-2017

2017-08-31 Thread Etienne Menguy
Hi, Do you know if some of the talks will be in english? Étienne From: ceph-users on behalf of Wido den Hollander Sent: Thursday, August 24, 2017 17:09 To: ceph-us...@ceph.com Subject: [ceph-users] Ceph Day

[ceph-users] luminous ceph-osd crash

2017-08-31 Thread Marcin Dulak
Hi, I have a virtual CentOS 7.3 test setup at: https://github.com/marcindulak/github-test-local/blob/a339ff 7505267545f593fd949a6453a56cdfd7fe/vagrant-ceph-rbd-tutorial-centos7.sh It seems to crash reproducibly with luminous, and works with kraken. Is this a known issue?

Re: [ceph-users] [rgw][s3] Object not in objects list

2017-08-31 Thread Rudenko Aleksandr
Hi, Maybe someone have thoughts? --- Best regards, Alexander Rudenko On 30 Aug 2017, at 12:28, Rudenko Aleksandr > wrote: Hi, I use ceph 0.94.10(hammer) with radosgw as S3-compatible object store. I have few objects in some bucket with strange