Dear Ceph Users,

Awaiting some suggestion please!



On Wed, Aug 17, 2016 at 11:15 AM, Gaurav Goyal <[email protected]>
wrote:

> Hello Mart,
>
> Thanks a lot for the detailed information!
> Please find my response inline and help me to get more knowledge on it
>
>
> Ceph works best with more hardware. It is not really designed for small
> scale setups. Of course small setups can work for a PoC or testing, but I
> would not advise this for production.
>
> [Gaurav] : We need this setup for PoC or testing.
>
> If you want to proceed however, have a good look the manuals or this
> mailinglist archive and do invest some time to understand the logic and
> workings of ceph before working or ordering hardware
>
> At least you want:
> - 3 monitors, preferable on dedicated servers
> [Gaurav] : With my current setup, can i install MON on Host 1 -->
> Controller + Compute1, Host 2 and Host 3
>
> - Per disk you will be running an ceph-osd instance. So a host with 2
> disks will run 2 osd instances. More OSD process is better performance, but
> also more memory and cpu usage.
>
> [Gaurav] : Understood, That means having 1T x 4 would be better than 2T x
> 2.
>
> - Per default ceph uses a replication factor of 3 (it is possible to set
> this to 2, but is not advised)
> - You can not fill up disks to 100%, also data will not distribute even
> over all disks, expect disks to be filled up (on average) maximum to
> 60-70%. You want to add more disks once you reach this limit.
>
> All on all, with a setup of 3 hosts, with 2x2TB disks, this will result in
> a net data availablity of (3x2x2TBx0.6)/3 = 2.4 TB
>
> [Gaurav] : As this is going to be a test lab environment, can we change
> the configuration to have more capacity rather than redundancy? How can we
> achieve it?
>
> If speed is required, consider SSD's (for data & journals, or only
> journals).
>
> In you email you mention "compute1/2/3", please note, if you use the rbd
> kernel driver, this can interfere with the OSD process and is not advised
> to run OSD and Kernel driver on the same hardware. If you still want to do
> that, split it up using VMs (we have a small testing cluster where we do
> mix compute and storage, there we have the OSDs running in VMs)
>
> [Gaurav] : within my mentioned environment, How can we split rbd kernel
> driver and OSD process? Should it be like rbd kernel driver on controller
> and OSD processes on compute hosts?
>
> Since my host 1 is controller + Compute1, Can you please share the steps
> to split it up using VMs and suggested by you.
>
> Regards
> Gaurav Goyal
>
>
> On Wed, Aug 17, 2016 at 9:28 AM, Mart van Santen <[email protected]>
> wrote:
>
>>
>> Dear Gaurav,
>>
>> Ceph works best with more hardware. It is not really designed for small
>> scale setups. Of course small setups can work for a PoC or testing, but I
>> would not advise this for production.
>>
>> If you want to proceed however, have a good look the manuals or this
>> mailinglist archive and do invest some time to understand the logic and
>> workings of ceph before working or ordering hardware
>>
>> At least you want:
>> - 3 monitors, preferable on dedicated servers
>> - Per disk you will be running an ceph-osd instance. So a host with 2
>> disks will run 2 osd instances. More OSD process is better performance, but
>> also more memory and cpu usage.
>> - Per default ceph uses a replication factor of 3 (it is possible to set
>> this to 2, but is not advised)
>> - You can not fill up disks to 100%, also data will not distribute even
>> over all disks, expect disks to be filled up (on average) maximum to
>> 60-70%. You want to add more disks once you reach this limit.
>>
>> All on all, with a setup of 3 hosts, with 2x2TB disks, this will result
>> in a net data availablity of (3x2x2TBx0.6)/3 = 2.4 TB
>>
>>
>> If speed is required, consider SSD's (for data & journals, or only
>> journals).
>>
>> In you email you mention "compute1/2/3", please note, if you use the rbd
>> kernel driver, this can interfere with the OSD process and is not advised
>> to run OSD and Kernel driver on the same hardware. If you still want to do
>> that, split it up using VMs (we have a small testing cluster where we do
>> mix compute and storage, there we have the OSDs running in VMs)
>>
>> Hope this helps,
>>
>> regards,
>>
>> mart
>>
>>
>>
>>
>> On 08/17/2016 02:21 PM, Gaurav Goyal wrote:
>>
>> Dear Ceph Users,
>>
>> I need your help to redesign my ceph storage network.
>>
>> As suggested in earlier discussions, i must not use SAN storage. So we
>> have decided to removed it.
>>
>> Now we are ordering Local HDDs.
>>
>> My Network would be
>>
>> Host1 --> Controller + Compute1 Host 2--> Compute2 Host 3 --> Compute3
>>
>> Is it right setup for ceph network? For Host1 and Host2 , we are using 1
>> 500GB disk for OS on each host .
>>
>> Should we use same size storage disks 500GB *8 for ceph environment or i
>> can order Disks in size of 2TB for ceph cluster?
>>
>> Making it
>>
>> 2T X 2 on Host1 2T X 2 on Host 2 2T X 2 on Host 3
>>
>> 12TB in total. replication factor 2 should make it 6 TB?
>>
>>
>> _______________________________________________
>> ceph-users mailing 
>> [email protected]http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>> --
>> Mart van Santen
>> Greenhost
>> E: [email protected]
>> T: +31 20 4890444
>> W: https://greenhost.nl
>>
>> A PGP signature can be attached to this e-mail,
>> you need PGP software to verify it.
>> My public key is available in keyserver(s)
>> see: http://tinyurl.com/openpgp-manual
>>
>> PGP Fingerprint: CA85 EB11 2B70 042D AF66  B29A 6437 01A1 10A3 D3A5
>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> [email protected]
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to