On Oct 7, 2016, at 6:19 PM, Srinivas Naga Kotaru (skotaru) < [email protected]> wrote:
Rajat & Clayton That is really very useful information. Thanks for the data .. Do we need to keep both clusterNetowrk & serviceNetwork same size? Based on your data, we would like to go with 1000 node cluster with 256 pods per node clusterNetworkCIDR: 10.128.0.0/13 hostSubnetLength: 9 serviceNetworkCIDR: 172.30.0.0/13 networkPluginName: "redhat/openshift-ovs-multitenant" is above configuration is right? Want to keep few things in mind · We don’t have control on how many projects clients will be created unless there is a hard limit platform admin can enforce There is a limit that can be enforced - it is documented here: https://docs.openshift.org/latest/admin_guide/managing_projects.html#limit-projects-per-user · We don’t have predictions on how many services will be created in a project. In practice most apps are 1 or 2 pods, and most users tend to create 1 or fewer apps, so estimating one service per pod is reasonable. You can limit the number of services a user can create in a project with limit ranges (see the docs) · Since Openshift now support # pod per node enforcement, we can monitor and tune this number as per our environement Do we have any latest planning guide ? The planning guide should be up to date, although I have not checked in the last few weeks to see about 3.3z -- *Srinivas Kotaru* *From: *Rajat Chopra <[email protected]> *Date: *Friday, October 7, 2016 at 4:56 PM *To: *"[email protected]" <[email protected]> *Cc: *Srinivas Naga Kotaru <[email protected]>, dev < [email protected]> *Subject: *Re: NetworkCIDR for big cluster Check this for the new defaults: https://docs.openshift.com/container-platform/3.3/install_config/configuring_sdn.html#configuring-the-pod-network-on-masters The default cluster subnet mask is now 14, but the important number to look at is hostSubnetLength. The default now is 9, which means one can have 512 pods and 512 nodes. To have 1000 nodes, either keep hostSubnetLength 8 (256 pods) or put the cidr mask at 13. For a cidr mask of 12, and a hostSubnetLength of 9, you will get 512 pods on each of 2048 nodes. /Rajat On Oct 7, 2016 4:08 PM, "Clayton Coleman" <[email protected]> wrote: > > Pod limits are now directly settable, so no longer hard coded. > > It sounds like you would plan for 100k pods which a 15 or 14 cidr for pods would cover. I would recommend leaving some headroom there. 14 might be safer if in the future you want to continue to grow the cluster - our target max cluster size is ~ 300k pods, so if you expect a significant bump plan for 14 ahead of time. > > In practice, most large clusters will be dense, and be dense due to services. I would plan for 1 pod per service just to be safe, so match the cidr from pods. > > There are multi tenant network settings that might impact how many pods per node are available that I think default for >>100 pods per node, but please double check. > > On Oct 6, 2016, at 5:24 PM, Srinivas Naga Kotaru (skotaru) < [email protected]> wrote: > >> Yes, we are using multi-tenant SDN plugin and our nodes are 8cpu x 64 GB capacity. Planning to host more than 100 pods per node. With the latest OpenShift release, I hope we don’t have old hard limits on cluster size and # pods on per node >> >> >> >> I don’t have an accurate answer for # services per project, which we can’t control at platform side. >> >> >> >> Based on abov facts can you recommend a decent CIDR for pods & service? >> >> >> >> -- *>>* *>> Srinivas Kotaru* >> >> *>>* *>> From:* "[email protected]" <[email protected]> *>> Date:* Thursday, October 6, 2016 at 1:50 PM *>> To:* Srinivas Naga Kotaru <[email protected]> *>> Cc:* dev <[email protected]> *>> Subject:* Re: NetworkCIDR for big cluster >> >> >> >> >> >> >> On Oct 6, 2016, at 1:23 PM, Srinivas Naga Kotaru (skotaru) < [email protected]> wrote: >>> >>> Hi >>> >>> >>> >>> We ‘re building 3 big clusters, 1 specific to each data center. growth expected to 1000 nodes each cluster over the time. >>> >>> >>> >>> Questions: >>> >>> >>> >>> 1. >>> >>> # egrep 'clusterNetworkCIDR|serviceNetworkCIDR' /etc/origin/master/master-config.yaml >>> >>> >>> >>> clusterNetworkCIDR: 10.1.0.0/16 >>> >>> serviceNetworkCIDR: 172.30.0.0/16 >>> >>> >>> >>> above are default subnet values. These defaults will be sufficient for 1k clsuter size? >> >> >> >> That would be 64k pods and services. If you expect densities in excess of 100 pods per node you should probably plan for a larger cidr for the cluster network and if you are using the multi tenant plugin you'll also want to plan around how many unique projects you will have. >> >> >> >> So inputs are: >> >> # of projects >> >> Expected number of services per project (simple apps or complex apps per project) >> >> Expected pod densities per node >> >> >> >> >>> >>> >>> >>> 2. if answer is ‘no’ can we change once cluster is build with new CIDR values depending on growth? ( I heard it is not possible once cluster was build) >> >> >> >> >> >> It is possible to grow it if you include the original cidr in the expanded cidr. However, it can be disruptive so we recommend planning up front. I would recommend leaving the possibility of a larger block open by expanding the mask if you are unsure about exact limits. >> >> >> >> >>> >>> 3. if answer is ‘no’ what is the right CIDR’s for 1k cluster size? >>> >>> >>> >>> >>> >>> -- *>>>* *>>> Srinivas Kotaru* >>> >>> _______________________________________________ >>> dev mailing list >>> [email protected] <[email protected]> >>> http://lists.openshift.redhat.com/openshiftmm/listinfo/dev <http://lists.openshift.redhat.com/openshiftmm/listinfo/dev> > > > _______________________________________________ > dev mailing list > [email protected] <[email protected]> > http://lists.openshift.redhat.com/openshiftmm/listinfo/dev <http://lists.openshift.redhat.com/openshiftmm/listinfo/dev> >
_______________________________________________ dev mailing list [email protected] http://lists.openshift.redhat.com/openshiftmm/listinfo/dev
