Hi Gaurav,
There are several ways to do it depending on how you deployed your ceph
cluster. Easiest way to do it is using ceph-ansible with purge-cluster yaml
ready made to wipe off CEPH.
https://github.com/ceph/ceph-ansible/blob/master/purge-cluster.yml
You may need to configure ansible inven
On Thu, Aug 04, 2016 at 12:20:01AM +, EP Komarla wrote:
> Hi All,
>
>
>
> I am trying to do a fresh install of Ceph Jewel on my cluster. I went through
> all the steps in configuring the network, ssh, password, etc. Now I am at the
> stage of running the ceph-deploy commands to install mo
Hi ...
We also use a _/mount_user/_ key to mount cephfs with ceph-fuse. I
remember that we had some troubles also. We use ceph-authtool to
generate the key with the following syntax:
ceph-authtool --create-keyring
--gen-key -n
--cap mds ''
Please suggest a procedure for this uninstallation process?
Regards
Gaurav Goyal
On Wed, Aug 3, 2016 at 5:58 PM, Gaurav Goyal
wrote:
> Thanks for your prompt
> response!
>
> Situation is bit different now. Customer want us to remove the ceph
> storage configuration from scratch. Let is openst
Hi All,
I am trying to do a fresh install of Ceph Jewel on my cluster. I went through
all the steps in configuring the network, ssh, password, etc. Now I am at the
stage of running the ceph-deploy commands to install monitors and other nodes.
I am getting the below error when I am deploying
Thanks Somnath,
I'll try moving my testing to master tomorrow to see if that improves the
stability at all.
Bryan
On 8/3/16, 4:50 PM, "Somnath Roy" wrote:
>Probably, it is better to move to latest master and reproduce this
>defect. Lot of stuff has changed between this.
>This is a good test ca
Yes Greg, agreed, I found some corruption during BlueFS replay , could also be
caught in detail if I run fsck() may be..
Will do it , but, in dev environment time consumed during fsck() could be a
challenge (though I have no idea how long it will take per TB of data, never
ran it) considering t
On Wed, Aug 3, 2016 at 3:50 PM, Somnath Roy wrote:
> Probably, it is better to move to latest master and reproduce this defect.
> Lot of stuff has changed between this.
> This is a good test case and I doubt any of us testing by enabling fsck() on
> mount/unmount.
Given that the allocator keeps
Probably, it is better to move to latest master and reproduce this defect. Lot
of stuff has changed between this.
This is a good test case and I doubt any of us testing by enabling fsck() on
mount/unmount.
Thanks & Regards
Somnath
-Original Message-
From: ceph-users [mailto:ceph-users-b
I've been doing some benchmarking of BlueStore in 10.2.2 the last few days
and
have come across a failure that keeps happening after stressing the cluster
fairly heavily. Some of the OSDs started failing and attempts to restart
them
fail to log anything in /var/log/ceph/, so I tried starting them
I'm having a hard time removing an RBD that I no longer need.
# rbd rm /
2016-08-03 15:00:01.085784 7ff9dfc997c0 -1 librbd: image has watchers - not
removing
Removing image: 0% complete...failed.
rbd: error: image still has watchers
This means the image is still open or the client using it crashe
Thanks for your prompt
response!
Situation is bit different now. Customer want us to remove the ceph storage
configuration from scratch. Let is openstack system work without ceph.
Later on install ceph with local disks.
So I need to know a procedure to uninstall ceph and unconfigure it from
open
It¹s probably rbd cache taking effect. If you know all your clients are
well behaved, you could set "rbd cache writethrough until flush" to false,
instead of the default true, but understand the ramification. You could
also just do it during benchmarking.
Warren Wang
From: ceph-users on behal
If I'm understanding your question correctly that you're asking how to actually
remove the SAN osds from ceph, then it doesn't matter what is using the storage
(ie openstack, cephfs, krbd, etc) as the steps are the same.
I'm going to assume that you've already added the new storage/osds to the
Hello David,
Can you help me with steps/Procedure to uninstall Ceph storage from
openstack environment?
Regards
Gaurav Goyal
On Tue, Aug 2, 2016 at 11:57 AM, Gaurav Goyal
wrote:
> Hello David,
>
> Thanks a lot for detailed information!
>
> This is going to help me.
>
>
> Regards
> Gaurav Goya
Hi Tom,
thank you very much for your hint regarding tcp_sack and sysctl network stack
tuning. This pointed me in the right direction.
We had seldom similar issues where under high network load reads stalled on
osds.
Enabling tcp_sack made the situation better for us and some more tuning
compl
Inspecting the ceph-dbg packages under
http://download.ceph.com/debian-jewel/pool/main/c/ceph/ it looks like this
is an ongoing issue and not specific to just 10.2.2. Specifically there
are only 2 ceph-dbg package versions:
ceph-dbg_10.0.2-1trusty_amd64.deb
ceph-dbg_10.0.2-1~bpo80+1_amd64.deb
Th
>
> The problem is that operations can happen concurrently, so the decision
> whether to remove or not to remove an entry is not very easy. We have seen
> before that application initiated multiple uploads of the same part, but
> the one that actually complete the last was not the last to upload (e
For some reason, during the v10.2.2 release,
ceph-dbg_10.0.2-1xenial_amd64.deb did not get transferred to
http://download.ceph.com/debian-jewel/pool/main/c/ceph/
- Ken
On Wed, Aug 3, 2016 at 12:27 PM, J. Ryan Earl wrote:
> Hello,
>
> New to the list. I'm working on performance tuning and testin
Hello,
New to the list. I'm working on performance tuning and testing a new Ceph
cluster built on Ubuntu 16.04 LTS and newest "Jewel" Ceph release. I'm in
the process of collecting stack frames as part of a profiling inspection
using FlameGraph (https://github.com/brendangregg/FlameGraph) to ins
Moving this to ceph-user so the broader community can weigh in.
However, I would recommend you please spell out your question in much
more detail if possible. Using a fragment like this will most likely
not get a response. Thanks.
On Tue, Aug 2, 2016 at 7:55 PM, Leandro wrote:
> Hi.
> after the
On Wed, Aug 3, 2016 at 10:57 AM, Brian Felton wrote:
> I should clarify:
>
> There doesn't seem to be a problem with list_multipart_parts -- upon
> further review, it seems to be doing the right thing. What tipped me off
> is that when one aborts a multipart upload where parts have been uploaded
I should clarify:
There doesn't seem to be a problem with list_multipart_parts -- upon
further review, it seems to be doing the right thing. What tipped me off
is that when one aborts a multipart upload where parts have been uploaded
more than once, the last copy of each part uploaded is successf
On Wed, Aug 3, 2016 at 10:10 AM, Brian Felton wrote:
> This may just be me having a conversation with myself, but maybe this will
> be helpful to someone else.
>
> Having dug and dug and dug through the code, I've come to the following
> realizations:
>
>1. When a multipart upload is complete
This may just be me having a conversation with myself, but maybe this will
be helpful to someone else.
Having dug and dug and dug through the code, I've come to the following
realizations:
1. When a multipart upload is completed, the function
list_multipart_parts in rgw_op.cc is called. Th
On Wed, Aug 3, 2016 at 5:24 PM, Lazuardi Nasution
wrote:
> Hi John,
>
> If I have multi MON, should I put all MON IPs on /etc/fstab?
Yes, you put all your mons in your /etc/fstab entry.
> Is there any
> way to overcome MDS metadata bottleneck when only single MDS active?
Nope, you're going to h
Hi John,
If I have multi MON, should I put all MON IPs on /etc/fstab? Is there any
way to overcome MDS metadata bottleneck when only single MDS active?
In case on loadbalanced file/web servers, which one is better, each server
mount to other replicated/distributed FS (for example via GlusterFS) o
Hi,
I'm looking for example about what to put on /etc/fstab if I want to auto
mount CephFS on failovered multi MDS (only one MDS is active) especially
with Jewel. My target is to build loadbalanced file/web servers with CephFS
backend.
Best regards,
___
Just a reminder, the Ceph Developer Monthly planning meeting is
starting online in approx 15m
http://wiki.ceph.com/Planning
--
Best Regards,
Patrick McGarry
Director Ceph Community || Red Hat
http://ceph.com || http://community.redhat.com
@scuttlemonkey || @ceph
_
On Wed, Aug 3, 2016 at 5:10 PM, Lazuardi Nasution
wrote:
> Hi,
>
> I'm looking for example about what to put on /etc/fstab if I want to auto
> mount CephFS on failovered multi MDS (only one MDS is active) especially
> with Jewel. My target is to build loadbalanced file/web servers with CephFS
> ba
Maybe something like this?
192.168.135.31:6789:/ /cephfs ceph
name=cephfs,secretfile=/etc/ceph/client.cephfs,noatime 0 0
Best regards
Daniel
From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of
Lazuardi Nasution
Sent: Wednesday, August 03, 2016 6:10 PM
On Wed, Aug 3, 2016 at 9:59 AM, Alex Gorbachev wrote:
> On Tue, Aug 2, 2016 at 10:49 PM, Vladislav Bolkhovitin wrote:
>> Alex Gorbachev wrote on 08/02/2016 07:56 AM:
>>> On Tue, Aug 2, 2016 at 9:56 AM, Ilya Dryomov wrote:
On Tue, Aug 2, 2016 at 3:49 PM, Alex Gorbachev
wrote:
> On
On Tue, Aug 2, 2016 at 10:49 PM, Vladislav Bolkhovitin wrote:
> Alex Gorbachev wrote on 08/02/2016 07:56 AM:
>> On Tue, Aug 2, 2016 at 9:56 AM, Ilya Dryomov wrote:
>>> On Tue, Aug 2, 2016 at 3:49 PM, Alex Gorbachev
>>> wrote:
On Mon, Aug 1, 2016 at 11:03 PM, Vladislav Bolkhovitin
wr
Hello,
On Wed, 3 Aug 2016 15:15:21 +0300 Maged Mokhtar wrote:
> Hello,
>
> I would like to build a small cluster with 20 disks to start but in the
> future would like to gradually increase it to maybe 200 disks.
> Is it better to fix the number of PGs in the pool from the beginning or is it
>
Hi List,
i am using Ceph Infernalis and Ubuntu 14.04 Kernel 3.13.
18 Data Server / 3 MON / 3 RBD Clients
I want to use RBD on the Client with image format 2 and Striping.
Is it supported?
I want to create rbd with:
rbd create testrbd -s 2T --image-format=2 --image-feature=striping
--image-featu
Changing the number of PGs is one of the most expensive operations you
can run, and should be avoided as much as possible.
Having said that you should try to avoid having way too many PGs with
very few OSDs, but it's certainly preferable to splitting PGs...
On Wed, Aug 3, 2016 at 1:15 PM, Maged M
Hello,
On Wed, 3 Aug 2016 13:42:50 +0200 Jan Schermer wrote:
> Christian, can you post your values for Power_Loss_Cap_Test on the drive
> which is failing?
>
Sure:
---
175 Power_Loss_Cap_Test 0x0033 001 001 010Pre-fail Always
FAILING_NOW 1 (47 942)
---
Now according to the In
Hello,
I would like to build a small cluster with 20 disks to start but in the future
would like to gradually increase it to maybe 200 disks.
Is it better to fix the number of PGs in the pool from the beginning or is it
better to start with a small number and then gradually change the number of
Christian, can you post your values for Power_Loss_Cap_Test on the drive which
is failing?
Thanks
Jan
> On 03 Aug 2016, at 13:33, Christian Balzer wrote:
>
>
> Hello,
>
> yeah, I was particular interested in the Power_Loss_Cap_Test bit, as it
> seemed to be such an odd thing to fail (given t
Dear all,
Further to my Cephfs testing, I am trying to put quota on the mount I
have done on client end. I am getting error message when querying the same.
ceph-fuse fuse.ceph-fuse 2.8T 5.5G 2.8T 1% /cephfs
# setfattr -n ceph.quota.max_bytes -v 1 /cephfs/test1/
# getfa
Hello,
yeah, I was particular interested in the Power_Loss_Cap_Test bit, as it
seemed to be such an odd thing to fail (given that's not single capacitor).
As for your Reallocated_Sector_Ct, that's really odd and definitely a RMA
worthy issue.
For the record, Intel SSDs use (typically 24) secto
Right, I actually updated to smartmontools 6.5+svn4324, which now
properly supports this drive model. Some of the smart attr names have
changed, and make more sense now (and there are no more "Unknowns"):
ID# ATTRIBUTE_NAME FLAGSVALUE WORST THRESH FAIL RAW_VALUE
5 Reallocated_Sector
I'm a fool, I miscalculated the writes by a factor of 1000 of course :-)
600GB/month is not much for S36xx at all, must be some sort of defect then...
Jan
> On 03 Aug 2016, at 12:15, Jan Schermer wrote:
>
> Make sure you are reading the right attribute and interpreting it right.
> update-smart
Make sure you are reading the right attribute and interpreting it right.
update-smart-drivedb sometimes makes wonders :)
I wonder what isdct tool would say the drive's life expectancy is with this
workload? Are you really writing ~600TB/month??
Jan
> On 03 Aug 2016, at 12:06, Maxime Guyot wro
Hi,
I haven’t had problems with Power_Loss_Cap_Test so far.
Regarding Reallocated_Sector_Ct (SMART ID: 5/05h), you can check the “Available
Reserved Space” (SMART ID: 232/E8h), the data sheet
(http://www.intel.com/content/dam/www/public/us/en/documents/product-specifications/ssd-dc-s3610-spec.
Hi Christian,
Intel drives are good, but apparently not infallible. I'm watching a DC
S3610 480GB die from reallocated sectors.
ID# ATTRIBUTE_NAME FLAGSVALUE WORST THRESH FAIL RAW_VALUE
5 Reallocated_Sector_Ct -O--CK 081 081 000-756
9 Power_On_Hours -O--C
Hi,
> I have never tried it, but gets back to my original question: Why the rack in
> between and not add the hosts directly to the root?
>
> You should add the rack when you want to set the failure domain to racks and
> thus replicate over multiple racks.
>
> In your case you want the failure
Hello,
On Wed, 3 Aug 2016 08:35:49 + Rob Reus wrote:
> Hi Wido,
>
>
> This is indeed something I have tried, and confirmed to work, see the other
> CRUSH map link I have provided in my original email.
>
>
> However, I was wondering if achieving that same goal, but with only 1 root,
> i
> Op 3 augustus 2016 om 10:35 schreef Rob Reus :
>
>
> Hi Wido,
>
>
> This is indeed something I have tried, and confirmed to work, see the other
> CRUSH map link I have provided in my original email.
>
Ah, double e-mails.
>
> However, I was wondering if achieving that same goal, but with
Hi Wido,
This is indeed something I have tried, and confirmed to work, see the other
CRUSH map link I have provided in my original email.
However, I was wondering if achieving that same goal, but with only 1 root, is
possible/feasible.
Thanks!
Van: Wido de
> Op 3 augustus 2016 om 10:08 schreef Rob Reus :
>
>
> Hi all,
>
>
> I built a CRUSH map, with the goal to distinguish between SSD and HDD storage
> machines using only 1 root. The map can be found here:
> http://pastebin.com/VQdB0CE9
>
>
> The issue I am having is this:
>
>
> root@ceph2
Dear All,
I am trying to use CephFS in my setup. I have created a test setup with
01 MON and 03 OSD's.
I have created a MDS server and able to mount it on client using FUSE.
Using admin keyring, I am able to write to cephfs and sub-dir also.
I am experiencing issue when I try to write to cephfs
Hi all,
I built a CRUSH map, with the goal to distinguish between SSD and HDD storage
machines using only 1 root. The map can be found here:
http://pastebin.com/VQdB0CE9
The issue I am having is this:
root@ceph2:~/crush_files# crushtool -i crushmap --test --show-utilization
--rule 0 --num-
Hi all,
I have built a CRUSH map myself, with the goal to distinguish between SSD
storage machines, and HDD storage machines, using a custom type. This with in
mind to have only 1 root (default). The map can be found here:
http://pastebin.com/VQdB0CE9
Now the issue I am seeing is that when I
54 matches
Mail list logo