Workaround... We're building now a huge computing cluster 140 computing 
DISKLESS nodes and they are pulling to storage a lot of computing data 
concurrently
    User that put job for the cluster - need also access to the same storage 
place (seeking progress & results)

We've built Ceph cluster:
    3 mon nodes (one of them is combined with mds)
    3 osd nodes (each one have 10 osd + ssd for journaling)
    switch 24 ports x 10G
    10 gigabit - for public network
    20 gigabit bonding - between osds
    Ubuntu 12.04.05
    Ceph 0.87.2 - giant
-----------------------------------------------------
Clients has:
    10 gigabit for ceph-connection
    CentOS 6.6 with upgraded kernel 3.19.8 (already running computing cluster)

Surely all nodes, switches and clients were configured to jumbo-frames of 
network

=========================================================================

First test:
    I thought to make big rbd with shareing, but:
          -  RBD supports multiple clients' mapping&mounting but not parallel 
writes ...

Second test:
    NFS over RBD - it's working pretty good, but:
        1. NFS gateway - it's Single-Point-of-Failure
        2. There's no performance scaling of scale-out storage e.g. bottleneck 
(limited with bandwidth of NFS-gateway)

Third test:
    We wanted to try CephFS, because our client is familiar with Lustre, that's 
very near to CephFS capabilities:
           1. I've used my CEPH nodes in the client's role. I've mounted CephFS 
on one of nodes, and ran dd with bs=1M ...
                    - I've got wonderful write performance ~ 1.1 GBytes/s 
(really near to 10Gbit network throughput)

            2. I've connected CentOS client to 10gig public network, mounted 
CephFS, but ...
                    - It was just ~ 250 MBytes/s

            3. I've connected Ubuntu client (non-ceph member) to 10gig public 
network, mounted CephFS, and ...
                    - It was also ~ 260 MBytes/s

            Now I have to know: perhaps ceph-members-nodes have privileged 
access ???

I'm sure you have more ceph deployment experience,
    have you seen this CephFS performance deviations?

Thanks,
Shneur




************************************************************************************
This footnote confirms that this email message has been scanned by
PineApp Mail-SeCure for the presence of malicious code, vandals & computer 
viruses.
************************************************************************************



_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to