Re: [ceph-users] Fwd: Ceph Storage Migration from SAN storage to Local Disks

2016-08-03 Thread Bharath Krishna
Hi Gaurav, There are several ways to do it depending on how you deployed your ceph cluster. Easiest way to do it is using ceph-ansible with purge-cluster yaml ready made to wipe off CEPH. https://github.com/ceph/ceph-ansible/blob/master/purge-cluster.yml You may need to configure ansible inven

Re: [ceph-users] Ceph-deploy on Jewel error

2016-08-03 Thread Chengwei Yang
On Thu, Aug 04, 2016 at 12:20:01AM +, EP Komarla wrote: > Hi All, > > > > I am trying to do a fresh install of Ceph Jewel on my cluster. I went through > all the steps in configuring the network, ssh, password, etc. Now I am at the > stage of running the ceph-deploy commands to install mo

Re: [ceph-users] Cephfs issue - able to mount with user key, not able to write

2016-08-03 Thread Goncalo Borges
Hi ... We also use a _/mount_user/_ key to mount cephfs with ceph-fuse. I remember that we had some troubles also. We use ceph-authtool to generate the key with the following syntax: ceph-authtool --create-keyring --gen-key -n --cap mds ''

Re: [ceph-users] Fwd: Ceph Storage Migration from SAN storage to Local Disks

2016-08-03 Thread Gaurav Goyal
Please suggest a procedure for this uninstallation process? Regards Gaurav Goyal On Wed, Aug 3, 2016 at 5:58 PM, Gaurav Goyal wrote: > Thanks for your prompt > response! > > Situation is bit different now. Customer want us to remove the ceph > storage configuration from scratch. Let is openst

[ceph-users] Ceph-deploy on Jewel error

2016-08-03 Thread EP Komarla
Hi All, I am trying to do a fresh install of Ceph Jewel on my cluster. I went through all the steps in configuring the network, ssh, password, etc. Now I am at the stage of running the ceph-deploy commands to install monitors and other nodes. I am getting the below error when I am deploying

Re: [ceph-users] Multi-device BlueStore OSDs multiple fsck failures

2016-08-03 Thread Stillwell, Bryan J
Thanks Somnath, I'll try moving my testing to master tomorrow to see if that improves the stability at all. Bryan On 8/3/16, 4:50 PM, "Somnath Roy" wrote: >Probably, it is better to move to latest master and reproduce this >defect. Lot of stuff has changed between this. >This is a good test ca

Re: [ceph-users] Multi-device BlueStore OSDs multiple fsck failures

2016-08-03 Thread Somnath Roy
Yes Greg, agreed, I found some corruption during BlueFS replay , could also be caught in detail if I run fsck() may be.. Will do it , but, in dev environment time consumed during fsck() could be a challenge (though I have no idea how long it will take per TB of data, never ran it) considering t

Re: [ceph-users] Multi-device BlueStore OSDs multiple fsck failures

2016-08-03 Thread Gregory Farnum
On Wed, Aug 3, 2016 at 3:50 PM, Somnath Roy wrote: > Probably, it is better to move to latest master and reproduce this defect. > Lot of stuff has changed between this. > This is a good test case and I doubt any of us testing by enabling fsck() on > mount/unmount. Given that the allocator keeps

Re: [ceph-users] Multi-device BlueStore OSDs multiple fsck failures

2016-08-03 Thread Somnath Roy
Probably, it is better to move to latest master and reproduce this defect. Lot of stuff has changed between this. This is a good test case and I doubt any of us testing by enabling fsck() on mount/unmount. Thanks & Regards Somnath -Original Message- From: ceph-users [mailto:ceph-users-b

[ceph-users] Multi-device BlueStore OSDs multiple fsck failures

2016-08-03 Thread Stillwell, Bryan J
I've been doing some benchmarking of BlueStore in 10.2.2 the last few days and have come across a failure that keeps happening after stressing the cluster fairly heavily. Some of the OSDs started failing and attempts to restart them fail to log anything in /var/log/ceph/, so I tried starting them

[ceph-users] [Troubleshooting] I have a watcher I can't get rid of...

2016-08-03 Thread K.C. Wong
I'm having a hard time removing an RBD that I no longer need. # rbd rm / 2016-08-03 15:00:01.085784 7ff9dfc997c0 -1 librbd: image has watchers - not removing Removing image: 0% complete...failed. rbd: error: image still has watchers This means the image is still open or the client using it crashe

Re: [ceph-users] Fwd: Ceph Storage Migration from SAN storage to Local Disks

2016-08-03 Thread Gaurav Goyal
Thanks for your prompt response! Situation is bit different now. Customer want us to remove the ceph storage configuration from scratch. Let is openstack system work without ceph. Later on install ceph with local disks. So I need to know a procedure to uninstall ceph and unconfigure it from open

Re: [ceph-users] I use fio with randwrite io to ceph image , it's run 2000 IOPS in the first time , and run 6000 IOPS in second time

2016-08-03 Thread Warren Wang - ISD
It¹s probably rbd cache taking effect. If you know all your clients are well behaved, you could set "rbd cache writethrough until flush" to false, instead of the default true, but understand the ramification. You could also just do it during benchmarking. Warren Wang From: ceph-users on behal

Re: [ceph-users] Fwd: Ceph Storage Migration from SAN storage to Local Disks

2016-08-03 Thread David Turner
If I'm understanding your question correctly that you're asking how to actually remove the SAN osds from ceph, then it doesn't matter what is using the storage (ie openstack, cephfs, krbd, etc) as the steps are the same. I'm going to assume that you've already added the new storage/osds to the

Re: [ceph-users] Fwd: Ceph Storage Migration from SAN storage to Local Disks

2016-08-03 Thread Gaurav Goyal
Hello David, Can you help me with steps/Procedure to uninstall Ceph storage from openstack environment? Regards Gaurav Goyal On Tue, Aug 2, 2016 at 11:57 AM, Gaurav Goyal wrote: > Hello David, > > Thanks a lot for detailed information! > > This is going to help me. > > > Regards > Gaurav Goya

Re: [ceph-users] Read Stalls with Multiple OSD Servers

2016-08-03 Thread Christoph Adomeit
Hi Tom, thank you very much for your hint regarding tcp_sack and sysctl network stack tuning. This pointed me in the right direction. We had seldom similar issues where under high network load reads stalled on osds. Enabling tcp_sack made the situation better for us and some more tuning compl

Re: [ceph-users] ceph-dbg package for Xenial (ubuntu-16.04.x) broken

2016-08-03 Thread J. Ryan Earl
Inspecting the ceph-dbg packages under http://download.ceph.com/debian-jewel/pool/main/c/ceph/ it looks like this is an ongoing issue and not specific to just 10.2.2. Specifically there are only 2 ceph-dbg package versions: ceph-dbg_10.0.2-1trusty_amd64.deb ceph-dbg_10.0.2-1~bpo80+1_amd64.deb Th

Re: [ceph-users] Cleaning Up Failed Multipart Uploads

2016-08-03 Thread Brian Felton
> > The problem is that operations can happen concurrently, so the decision > whether to remove or not to remove an entry is not very easy. We have seen > before that application initiated multiple uploads of the same part, but > the one that actually complete the last was not the last to upload (e

Re: [ceph-users] ceph-dbg package for Xenial (ubuntu-16.04.x) broken

2016-08-03 Thread Ken Dreyer
For some reason, during the v10.2.2 release, ceph-dbg_10.0.2-1xenial_amd64.deb did not get transferred to http://download.ceph.com/debian-jewel/pool/main/c/ceph/ - Ken On Wed, Aug 3, 2016 at 12:27 PM, J. Ryan Earl wrote: > Hello, > > New to the list. I'm working on performance tuning and testin

[ceph-users] ceph-dbg package for Xenial (ubuntu-16.04.x) broken

2016-08-03 Thread J. Ryan Earl
Hello, New to the list. I'm working on performance tuning and testing a new Ceph cluster built on Ubuntu 16.04 LTS and newest "Jewel" Ceph release. I'm in the process of collecting stack frames as part of a profiling inspection using FlameGraph (https://github.com/brendangregg/FlameGraph) to ins

Re: [ceph-users] How using block device after cluster ceph on?

2016-08-03 Thread Patrick McGarry
Moving this to ceph-user so the broader community can weigh in. However, I would recommend you please spell out your question in much more detail if possible. Using a fragment like this will most likely not get a response. Thanks. On Tue, Aug 2, 2016 at 7:55 PM, Leandro wrote: > Hi. > after the

Re: [ceph-users] Cleaning Up Failed Multipart Uploads

2016-08-03 Thread Yehuda Sadeh-Weinraub
On Wed, Aug 3, 2016 at 10:57 AM, Brian Felton wrote: > I should clarify: > > There doesn't seem to be a problem with list_multipart_parts -- upon > further review, it seems to be doing the right thing. What tipped me off > is that when one aborts a multipart upload where parts have been uploaded

Re: [ceph-users] Cleaning Up Failed Multipart Uploads

2016-08-03 Thread Brian Felton
I should clarify: There doesn't seem to be a problem with list_multipart_parts -- upon further review, it seems to be doing the right thing. What tipped me off is that when one aborts a multipart upload where parts have been uploaded more than once, the last copy of each part uploaded is successf

Re: [ceph-users] Cleaning Up Failed Multipart Uploads

2016-08-03 Thread Yehuda Sadeh-Weinraub
On Wed, Aug 3, 2016 at 10:10 AM, Brian Felton wrote: > This may just be me having a conversation with myself, but maybe this will > be helpful to someone else. > > Having dug and dug and dug through the code, I've come to the following > realizations: > >1. When a multipart upload is complete

Re: [ceph-users] Cleaning Up Failed Multipart Uploads

2016-08-03 Thread Brian Felton
This may just be me having a conversation with myself, but maybe this will be helpful to someone else. Having dug and dug and dug through the code, I've come to the following realizations: 1. When a multipart upload is completed, the function list_multipart_parts in rgw_op.cc is called. Th

Re: [ceph-users] Automount Failovered Multi MDS CephFS

2016-08-03 Thread John Spray
On Wed, Aug 3, 2016 at 5:24 PM, Lazuardi Nasution wrote: > Hi John, > > If I have multi MON, should I put all MON IPs on /etc/fstab? Yes, you put all your mons in your /etc/fstab entry. > Is there any > way to overcome MDS metadata bottleneck when only single MDS active? Nope, you're going to h

Re: [ceph-users] Automount Failovered Multi MDS CephFS

2016-08-03 Thread Lazuardi Nasution
Hi John, If I have multi MON, should I put all MON IPs on /etc/fstab? Is there any way to overcome MDS metadata bottleneck when only single MDS active? In case on loadbalanced file/web servers, which one is better, each server mount to other replicated/distributed FS (for example via GlusterFS) o

[ceph-users] Automount Failovered Multi MDS CephFS

2016-08-03 Thread Lazuardi Nasution
Hi, I'm looking for example about what to put on /etc/fstab if I want to auto mount CephFS on failovered multi MDS (only one MDS is active) especially with Jewel. My target is to build loadbalanced file/web servers with CephFS backend. Best regards, ___

[ceph-users] CDM Starting in 15m

2016-08-03 Thread Patrick McGarry
Just a reminder, the Ceph Developer Monthly planning meeting is starting online in approx 15m http://wiki.ceph.com/Planning -- Best Regards, Patrick McGarry Director Ceph Community || Red Hat http://ceph.com || http://community.redhat.com @scuttlemonkey || @ceph _

Re: [ceph-users] Automount Failovered Multi MDS CephFS

2016-08-03 Thread John Spray
On Wed, Aug 3, 2016 at 5:10 PM, Lazuardi Nasution wrote: > Hi, > > I'm looking for example about what to put on /etc/fstab if I want to auto > mount CephFS on failovered multi MDS (only one MDS is active) especially > with Jewel. My target is to build loadbalanced file/web servers with CephFS > ba

Re: [ceph-users] Automount Failovered Multi MDS CephFS

2016-08-03 Thread Daniel Schwager
Maybe something like this? 192.168.135.31:6789:/ /cephfs ceph name=cephfs,secretfile=/etc/ceph/client.cephfs,noatime 0 0 Best regards Daniel From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of Lazuardi Nasution Sent: Wednesday, August 03, 2016 6:10 PM

Re: [ceph-users] [Scst-devel] Thin Provisioning and Ceph RBD's

2016-08-03 Thread Alex Gorbachev
On Wed, Aug 3, 2016 at 9:59 AM, Alex Gorbachev wrote: > On Tue, Aug 2, 2016 at 10:49 PM, Vladislav Bolkhovitin wrote: >> Alex Gorbachev wrote on 08/02/2016 07:56 AM: >>> On Tue, Aug 2, 2016 at 9:56 AM, Ilya Dryomov wrote: On Tue, Aug 2, 2016 at 3:49 PM, Alex Gorbachev wrote: > On

Re: [ceph-users] [Scst-devel] Thin Provisioning and Ceph RBD's

2016-08-03 Thread Alex Gorbachev
On Tue, Aug 2, 2016 at 10:49 PM, Vladislav Bolkhovitin wrote: > Alex Gorbachev wrote on 08/02/2016 07:56 AM: >> On Tue, Aug 2, 2016 at 9:56 AM, Ilya Dryomov wrote: >>> On Tue, Aug 2, 2016 at 3:49 PM, Alex Gorbachev >>> wrote: On Mon, Aug 1, 2016 at 11:03 PM, Vladislav Bolkhovitin wr

Re: [ceph-users] Number of PGs: fix from start or change as we grow ?

2016-08-03 Thread Christian Balzer
Hello, On Wed, 3 Aug 2016 15:15:21 +0300 Maged Mokhtar wrote: > Hello, > > I would like to build a small cluster with 20 disks to start but in the > future would like to gradually increase it to maybe 200 disks. > Is it better to fix the number of PGs in the pool from the beginning or is it >

[ceph-users] Ubuntu 14.04 Striping / RBD / Single Thread Performance

2016-08-03 Thread w...@globe.de
Hi List, i am using Ceph Infernalis and Ubuntu 14.04 Kernel 3.13. 18 Data Server / 3 MON / 3 RBD Clients I want to use RBD on the Client with image format 2 and Striping. Is it supported? I want to create rbd with: rbd create testrbd -s 2T --image-format=2 --image-feature=striping --image-featu

Re: [ceph-users] Number of PGs: fix from start or change as we grow ?

2016-08-03 Thread Luis Periquito
Changing the number of PGs is one of the most expensive operations you can run, and should be avoided as much as possible. Having said that you should try to avoid having way too many PGs with very few OSDs, but it's certainly preferable to splitting PGs... On Wed, Aug 3, 2016 at 1:15 PM, Maged M

Re: [ceph-users] Intel SSD (DC S3700) Power_Loss_Cap_Test failure

2016-08-03 Thread Christian Balzer
Hello, On Wed, 3 Aug 2016 13:42:50 +0200 Jan Schermer wrote: > Christian, can you post your values for Power_Loss_Cap_Test on the drive > which is failing? > Sure: --- 175 Power_Loss_Cap_Test 0x0033 001 001 010Pre-fail Always FAILING_NOW 1 (47 942) --- Now according to the In

[ceph-users] Number of PGs: fix from start or change as we grow ?

2016-08-03 Thread Maged Mokhtar
Hello, I would like to build a small cluster with 20 disks to start but in the future would like to gradually increase it to maybe 200 disks. Is it better to fix the number of PGs in the pool from the beginning or is it better to start with a small number and then gradually change the number of

Re: [ceph-users] Intel SSD (DC S3700) Power_Loss_Cap_Test failure

2016-08-03 Thread Jan Schermer
Christian, can you post your values for Power_Loss_Cap_Test on the drive which is failing? Thanks Jan > On 03 Aug 2016, at 13:33, Christian Balzer wrote: > > > Hello, > > yeah, I was particular interested in the Power_Loss_Cap_Test bit, as it > seemed to be such an odd thing to fail (given t

[ceph-users] Cephfs quota implement

2016-08-03 Thread Daleep Singh Bais
Dear all, Further to my Cephfs testing, I am trying to put quota on the mount I have done on client end. I am getting error message when querying the same. ceph-fuse fuse.ceph-fuse 2.8T 5.5G 2.8T 1% /cephfs # setfattr -n ceph.quota.max_bytes -v 1 /cephfs/test1/ # getfa

Re: [ceph-users] Intel SSD (DC S3700) Power_Loss_Cap_Test failure

2016-08-03 Thread Christian Balzer
Hello, yeah, I was particular interested in the Power_Loss_Cap_Test bit, as it seemed to be such an odd thing to fail (given that's not single capacitor). As for your Reallocated_Sector_Ct, that's really odd and definitely a RMA worthy issue. For the record, Intel SSDs use (typically 24) secto

Re: [ceph-users] Intel SSD (DC S3700) Power_Loss_Cap_Test failure

2016-08-03 Thread Daniel Swarbrick
Right, I actually updated to smartmontools 6.5+svn4324, which now properly supports this drive model. Some of the smart attr names have changed, and make more sense now (and there are no more "Unknowns"): ID# ATTRIBUTE_NAME FLAGSVALUE WORST THRESH FAIL RAW_VALUE 5 Reallocated_Sector

Re: [ceph-users] Intel SSD (DC S3700) Power_Loss_Cap_Test failure

2016-08-03 Thread Jan Schermer
I'm a fool, I miscalculated the writes by a factor of 1000 of course :-) 600GB/month is not much for S36xx at all, must be some sort of defect then... Jan > On 03 Aug 2016, at 12:15, Jan Schermer wrote: > > Make sure you are reading the right attribute and interpreting it right. > update-smart

Re: [ceph-users] Intel SSD (DC S3700) Power_Loss_Cap_Test failure

2016-08-03 Thread Jan Schermer
Make sure you are reading the right attribute and interpreting it right. update-smart-drivedb sometimes makes wonders :) I wonder what isdct tool would say the drive's life expectancy is with this workload? Are you really writing ~600TB/month?? Jan > On 03 Aug 2016, at 12:06, Maxime Guyot wro

Re: [ceph-users] Intel SSD (DC S3700) Power_Loss_Cap_Test failure

2016-08-03 Thread Maxime Guyot
Hi, I haven’t had problems with Power_Loss_Cap_Test so far. Regarding Reallocated_Sector_Ct (SMART ID: 5/05h), you can check the “Available Reserved Space” (SMART ID: 232/E8h), the data sheet (http://www.intel.com/content/dam/www/public/us/en/documents/product-specifications/ssd-dc-s3610-spec.

Re: [ceph-users] Intel SSD (DC S3700) Power_Loss_Cap_Test failure

2016-08-03 Thread Daniel Swarbrick
Hi Christian, Intel drives are good, but apparently not infallible. I'm watching a DC S3610 480GB die from reallocated sectors. ID# ATTRIBUTE_NAME FLAGSVALUE WORST THRESH FAIL RAW_VALUE 5 Reallocated_Sector_Ct -O--CK 081 081 000-756 9 Power_On_Hours -O--C

Re: [ceph-users] CRUSH map utilization issue

2016-08-03 Thread Rob Reus
Hi, > I have never tried it, but gets back to my original question: Why the rack in > between and not add the hosts directly to the root? > > You should add the rack when you want to set the failure domain to racks and > thus replicate over multiple racks. > > In your case you want the failure

Re: [ceph-users] CRUSH map utilization issue

2016-08-03 Thread Christian Balzer
Hello, On Wed, 3 Aug 2016 08:35:49 + Rob Reus wrote: > Hi Wido, > > > This is indeed something I have tried, and confirmed to work, see the other > CRUSH map link I have provided in my original email. > > > However, I was wondering if achieving that same goal, but with only 1 root, > i

Re: [ceph-users] CRUSH map utilization issue

2016-08-03 Thread Wido den Hollander
> Op 3 augustus 2016 om 10:35 schreef Rob Reus : > > > Hi Wido, > > > This is indeed something I have tried, and confirmed to work, see the other > CRUSH map link I have provided in my original email. > Ah, double e-mails. > > However, I was wondering if achieving that same goal, but with

Re: [ceph-users] CRUSH map utilization issue

2016-08-03 Thread Rob Reus
Hi Wido, This is indeed something I have tried, and confirmed to work, see the other CRUSH map link I have provided in my original email. However, I was wondering if achieving that same goal, but with only 1 root, is possible/feasible. Thanks! Van: Wido de

Re: [ceph-users] CRUSH map utilization issue

2016-08-03 Thread Wido den Hollander
> Op 3 augustus 2016 om 10:08 schreef Rob Reus : > > > Hi all, > > > I built a CRUSH map, with the goal to distinguish between SSD and HDD storage > machines using only 1 root. The map can be found here: > http://pastebin.com/VQdB0CE9 > > > The issue I am having is this: > > > root@ceph2

[ceph-users] Cephfs issue - able to mount with user key, not able to write

2016-08-03 Thread Daleep Singh Bais
Dear All, I am trying to use CephFS in my setup. I have created a test setup with 01 MON and 03 OSD's. I have created a MDS server and able to mount it on client using FUSE. Using admin keyring, I am able to write to cephfs and sub-dir also. I am experiencing issue when I try to write to cephfs

[ceph-users] CRUSH map utilization issue

2016-08-03 Thread Rob Reus
Hi all, I built a CRUSH map, with the goal to distinguish between SSD and HDD storage machines using only 1 root. The map can be found here: http://pastebin.com/VQdB0CE9 The issue I am having is this: root@ceph2:~/crush_files# crushtool -i crushmap --test --show-utilization --rule 0 --num-

[ceph-users] CRUSH map utilization issue

2016-08-03 Thread Rob Reus
Hi all, I have built a CRUSH map myself, with the goal to distinguish between SSD storage machines, and HDD storage machines, using a custom type. This with in mind to have only 1 root (default). The map can be found here: http://pastebin.com/VQdB0CE9 Now the issue I am seeing is that when I