That is correct, you make a tradeoff between space, performance and
resiliency. By reducing replication from 3 to 2, you will get more space
and likely more performance (less overhead from third copy), but it comes
at the expense of being able to recover your data when there are multiple
failures.

On Fri, Mar 13, 2015 at 4:13 AM, Thomas Foster <[email protected]>
wrote:

> Ok, now if I run a lab and the data is somewhat important but I can bare
> losing the data, couldn't I shrink the pool replica count and that
> increases the amount of storage I can use without using erasure coding?
>
> So for 145TB with a replica of 3 = ~41 TB total in the cluster
>
> But if that same clusters replica was decreased to 2 I could possibly get
> 145TB / 2 - overhead for cluster and get ~65TB in the cluster at one
> time..correct?
>
> Thanks in advance!
>
>
> On Mar 12, 2015 11:53 AM, "Kamil Kuramshin" <[email protected]>
> wrote:
>
>>  For example, here is my confuguration:
>>
>> superuser@admin:~$ ceph df
>> GLOBAL:
>>     SIZE     AVAIL     RAW USED     %RAW USED
>>     242T      209T       20783G          8.38
>> POOLS:
>>     NAME                  ID     USED      %USED     MAX AVAIL
>> OBJECTS
>>     ec_backup-storage     4      9629G      3.88          137T
>> 2465171
>>     cache                 5       136G      0.06        38393M
>> 35036
>>     block-devices         6      1953G      0.79        70202G
>> 500060
>>
>>
>> *ec_backup-storage* - is Erasure Encoded pool, k=2, m=1 (default)
>> *cache* - is replicated pool consisting dedicated 12xSSDx60Gb disks,
>> replica size=3, used as cache tier for EC pool
>> *block-devices* - is replicated pool, replica size=3, using same OSD's
>> that in Erasure Encoded pool
>>
>> On* '**MAX AVAIL**'* column you can see that EC pool currently has
>> *137Tb* of free space, but in same time if we will write to replicated
>> pool there is only *70Tb, *but *both* pools are on the *same* *OSD's. *So
>> using EC pool saves 2 times more effective space in my case!
>>
>> 12.03.2015 17:50, Thomas Foster пишет:
>>
>> Thank you!  That helps alot.
>> On Mar 12, 2015 10:40 AM, "Steve Anthony" <[email protected]> wrote:
>>
>>>  Actually, it's more like 41TB. It's a bad idea to run at near full
>>> capacity (by default past 85%) because you need some space where Ceph can
>>> replicate data as part of its healing process in the event of disk or node
>>> failure. You'll get a health warning when you exceed this ratio.
>>>
>>> You can use erasure coding to increase the amount of data you can store
>>> beyond 41TB, but you'll still need some replicated disk as a caching layer
>>> in front of the erasure coded pool if you're using RBD. See:
>>> http://lists.ceph.com/pipermail/ceph-users-ceph.com/2013-December/036430.html
>>>
>>> As to how much space you can save with erasure coding, that will depend
>>> on if you're using RBD and need a cache layer and the values you set for k
>>> and m (number of data chunks and coding chunks). There's been some
>>> discussion on the list with regards to choosing those values.
>>>
>>> -Steve
>>>
>>> On 03/12/2015 10:07 AM, Thomas Foster wrote:
>>>
>>> I am looking into how I can maximize my space with replication, and I am
>>> trying to understand how I can do that.
>>>
>>>  I have 145TB of space and a replication of 3 for the pool and was
>>> thinking that the max data I can have in the cluster is ~47TB in my cluster
>>> at one time..is that correct?  Or is there a way to get more data into the
>>> cluster with less space using erasure coding?
>>>
>>>  Any help would be greatly appreciated.
>>>
>>>
>>>
>>>
>>> _______________________________________________
>>> ceph-users mailing 
>>> [email protected]http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>>
>>> --
>>> Steve Anthony
>>> LTS HPC Support Specialist
>>> Lehigh [email protected]
>>>
>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> [email protected]
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>>
>>
>> _______________________________________________
>> ceph-users mailing 
>> [email protected]http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> [email protected]
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to