Travis,

We run a routed ECMP spine-leaf network architecture with Ceph and have no issues on the network side whatsoever. Each leaf switch has an L2 cidr block inside a common L3 supernet.

We do not currently split cluster_network and public_network. If we did, we'd likely build a separate spine-leaf network with it's own L3 supernet.

A simple IPv4 example:

- ceph-cluster: 10.1.0.0/16
    - cluster-leaf1: 10.1.1.0/24
        - node1: 10.1.1.1/24
        - node2: 10.1.1.2/24
    - cluster-leaf2: 10.1.2.0/24

- ceph-public: 10.2.0.0/16
    - public-leaf1: 10.2.1.0/24
        - node1: 10.2.1.1/24
        - node2: 10.2.1.2/24
    - public-leaf2: 10.2.2.0/24

ceph.conf would be:

cluster_network: 10.1.0.0/255.255.0.0
public_network: 10.2.0.0/255.255.0.0

- Mike Dawson

On 5/28/2014 1:01 PM, Travis Rhoden wrote:
Hi folks,

Does anybody know if there are any issues running Ceph with multiple L2
LAN segements?  I'm picturing a large multi-rack/multi-row deployment
where you may give each rack (or row) it's own L2 segment, then connect
them all with L3/ECMP in a leaf-spine architecture.

I'm wondering how cluster_network (or public_network) in ceph.conf works
in this case.  Does that directive just tell a daemon starting on a
particular node which network to bind to?  Or is a CIDR that has to be
accurate for every OSD and MON in the entire cluster?

Thanks,

  - Travis


_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to