Hello,
On Wed, 11 May 2016 11:24:29 +0800 Geocast Networks wrote:
> Hi,
>
> If we have 12 SATA disks, each 4TB as storage pool.
> Then how many SSD disks we should have for cache tier usage?
>
That question makes no sense.
Firstly you mentioned earlier that you have 21 of those hosts.
Which
Hi,
If we have 12 SATA disks, each 4TB as storage pool.
Then how many SSD disks we should have for cache tier usage?
thanks.
2016-05-10 16:40 GMT+08:00 Yoann Moulin :
> Hello,
>
> I'd like some advices about the setup of a new ceph cluster. Here the use
> case :
>
>
On Tue, 10 May 2016 17:51:24 +0200 Yoann Moulin wrote:
[snip]
> Journal or cache Storage : 2 x SSD 400GB Intel S3300 DC (no Raid)
> >>>
> >>> These SSDs do not exist according to the Intel site and the only
> >>> references I can find for them are on "no longer available" European
> >>>
Hello,
As far as I know and can tell, you're doing everything that is possible
for having a least impact OSD rebuild/replacement.
If your cluster is still strongly, adversely impacted by this gradual and
throttled approach, how about the following things:
1. Does scrub or deep_scrub also
Hello All,
I´m writing to you because i´m trying to find the way to rebuild a osd disk
in a way to don´t impact the performance of the cluster.
That´s because my applications are very latency sensitive.
1_ I found the way to reuse a OSD ID and don´t rebalance the cluster every
time that I lost a
All,
I am trying to add another OSD to our cluster using ceph-deploy. This is
running Jewel.
I previously set up the other 12 OSDs on a fresh install using the command:
ceph-deploy osd create :/dev/mapper/mpath:/dev/sda
Those are all up and happy. On the systems /dev/sda is an SSD which I have
Thanks Nick. I added it to my ceph.conf. I'm guessing this is an OSD
setting and therefor I should restart my OSDs is that correct?
On Tue, May 10, 2016 at 3:48 PM, Nick Fisk wrote:
>
>
> > -Original Message-
> > From: ceph-users
On Tue, May 10, 2016 at 6:48 AM, Nick Fisk wrote:
>
>
>> -Original Message-
>> From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of
>> Nick Fisk
>> Sent: 10 May 2016 13:30
>> To: 'Eric Eastman'
>> Cc: 'Ceph Users'
I must also add that I just found in the log the following. I don't know if
this has something to do with the problem.
=> ceph-osd.admin.log <==
2016-05-10 18:21:46.060278 7fa8f30cc8c0 0 ceph version 10.1.2
(4a2a6f72640d6b74a3bbd92798bb913ed380dcd4), process ceph-osd, pid 14135
2016-05-10
Hello,
I forgot to say that the nodes are in preboot status. Something seems
strange to me.
root@red-compute:/var/lib/ceph/osd/ceph-1# ceph daemon osd.1 status
{
"cluster_fsid": "9028f4da-0d77-462b-be9b-dbdf7fa57771",
"osd_fsid": "adf9890a-e680-48e4-82c6-e96f4ed56889",
"whoami": 1,
On Thu, 21 Apr 2016, Dan van der Ster wrote:
> On Thu, Apr 21, 2016 at 1:23 PM, Dan van der Ster wrote:
> > Hi cephalapods,
> >
> > In our couple years of operating a large Ceph cluster, every single
> > inconsistency I can recall was caused by a failed read during
> >
Hello,
I just upgraded my cluster to the version 10.1.2 and it worked well for a
while until I saw that systemctl ceph-disk@dev-sdc1.service was failed and
I reruned it.
>From there the OSD stopped working.
This is ubuntu 16.04.
I connected to the IRC looking for help where people pointed me
Hey,
we currently have a problem with our radosgw.
The quota value of a user does not get updated after an admin manually deleted
a bucket (via radosgw-admin). You can only circumvent this if you synced the
user stats before the removal. So there are now users which can not upload new
objects
> -Original Message-
> From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of
> Peter Kerdisle
> Sent: 10 May 2016 14:37
> Cc: ceph-users@lists.ceph.com
> Subject: Re: [ceph-users] Erasure pool performance expectations
>
> To answer my own question it seems that you
> -Original Message-
> From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of
> Nick Fisk
> Sent: 10 May 2016 13:30
> To: 'Eric Eastman'
> Cc: 'Ceph Users'
> Subject: Re: [ceph-users] CephFS + CTDB/Samba - MDS
> -Original Message-
> From: Eric Eastman [mailto:eric.east...@keepertech.com]
> Sent: 09 May 2016 23:09
> To: Nick Fisk
> Cc: Ceph Users
> Subject: Re: [ceph-users] CephFS + CTDB/Samba - MDS session timeout on
> lockfile
>
> On Mon, May 9,
Hello,
On Tue, 10 May 2016 13:14:35 +0200 Yoann Moulin wrote:
> Hello,
>
> >> I'd like some advices about the setup of a new ceph cluster. Here the
> >> use case :
> >>
> >> RadowGW (S3 and maybe swift for hadoop/spark) will be the main usage.
> >> Most of the access will be in read only mode.
Hello.
I have a two node cluster with 4x replicas for all objects distributed between
the two nodes (two copies on each node). I recently converted my OSDs from
BTRFS to XFS (BTRFS was slow) by removing / preparing / activating OSDs on each
node (one at at time) as XFS allowing cluster to
Hello,
we are running ceph 0.94.5 (9764da52395923e0b32908d83a9f7304401fee43) on a
4.5.2 kernel. Our cluster currently consists of 5 nodes, with 6 OSD's each.
An issue has also been filed here (also containg logs, etc.):
http://tracker.ceph.com/issues/15813
Last night we have observed a single
Hello,
On Tue, 10 May 2016 16:50:17 +0800 Geocast Networks wrote:
> Hello Chris,
>
> We don't use SSD as journal.
> each host has one intel E5-2620 CPU which is 6 cores.
That should be enough.
> the networking (both cluster and data networks) is 10Gbps.
>
12 HDDs will barely saturate a 10Gb/s
Hello,
On Tue, 10 May 2016 10:40:08 +0200 Yoann Moulin wrote:
> Hello,
>
> I'd like some advices about the setup of a new ceph cluster. Here the
> use case :
>
> RadowGW (S3 and maybe swift for hadoop/spark) will be the main usage.
> Most of the access will be in read only mode. Write access
> rbd_cache_size = 268435456
Are you sure that you have 256MB per client to waste on RBD cache?
If so, bully for you, but you might find that depending on your use case a
smaller RBD cache but more VM memory (for pagecache, SLAB, etc) could be
more beneficial.
We have changed this value to 64MB.
Hello Chris,
We don't use SSD as journal.
each host has one intel E5-2620 CPU which is 6 cores.
the networking (both cluster and data networks) is 10Gbps.
My further questions include,
(1) osd_mkfs_type = xfs
osd_mkfs_options_xfs = -f
filestore_xattr_use_omap = true
for XFS filesystem, we
Hello,
I'd like some advices about the setup of a new ceph cluster. Here the use case :
RadowGW (S3 and maybe swift for hadoop/spark) will be the main usage. Most of
the access will be in read only mode. Write access will only be done by the
admin to update the datasets.
We might use rbd some
On Tue, 10 May 2016 11:48:07 +0800 Geocast wrote:
Hello,
> We have 21 hosts for ceph OSD servers, each host has 12 SATA disks (4TB
> each), 64GB memory.
No journal SSDs?
What CPU(s) and network?
> ceph version 10.2.0, Ubuntu 16.04 LTS
> The whole cluster is new installed.
>
> Can you help
25 matches
Mail list logo