Let me clarify - I did not say that you have to "label" nodes and masters.
I was suggesting to tag nodes and masters, the way you tag a cloud resource via AWS console or AWS CLI. I meant - AWS tag not openshift labels. The reason you have volumes created in another zone is because - your AWS account has nodes in more than one zone, possibly not part of Openshift cluster. But when you are requesting a dynamic provisioned volume - Openshift considers all nodes it can find and accordingly it "randomly" selects a zone among zone it discovered. But if you were to use AWS Console or CLI to tag all nodes(including master) in your cluster with "KubernetesCluster" : "cluster_id" then it will only select tagged nodes and hence it will pick zone in which Openshift cluster did not exist. On Fri, Jan 5, 2018 at 11:48 PM, Marc Boorshtein <[email protected]> wrote: > how do i label a master? When i create PVCs it switches between 1c and > 1a. look on the master I see: > > Creating volume for PVC "wtf3"; chose zone="us-east-1c" from > zones=["us-east-1a" "us-east-1c"] > > Where did us-east-1c come from??? > > On Fri, Jan 5, 2018 at 11:07 PM Hemant Kumar <[email protected]> wrote: > >> Both nodes and masters. The tag information is picked from master >> itself(Where controller-manager is running) and then openshift uses same >> value to find all nodes in the cluster. >> >> >> >> >> On Fri, Jan 5, 2018 at 10:26 PM, Marc Boorshtein <[email protected]> >> wrote: >> >>> node and masters? or just nodes? (sounded like just nodes from the docs) >>> >>> On Fri, Jan 5, 2018 at 9:16 PM Hemant Kumar <[email protected]> wrote: >>> >>>> Make sure that you configure ALL instances in the cluster with tag >>>> "KubernetesCluster": "value". The value of the tag for key >>>> "KubernetesCluster" should be same for all instances in the cluster. You >>>> can choose any string you want for value. >>>> >>>> You will probably have to restart openshift controller-manager after >>>> the change at very minimum. >>>> >>>> >>>> >>>> On Fri, Jan 5, 2018 at 8:21 PM, Marc Boorshtein <[email protected]> >>>> wrote: >>>> >>>>> Hello, >>>>> >>>>> I have a brand new Origin 3.6 running on AWS, the master and all nodes >>>>> are in us-east-1a but whenever I try to have AWS create a new volume, it >>>>> puts it in us-east-1c so then no one can access it and all my nodes go >>>>> into >>>>> a permanent pending state because NoVolumeZoneConflict. Looking at >>>>> aws.conf it states us-east-1a. What am I missing? >>>>> >>>>> Thanks >>>>> >>>>> _______________________________________________ >>>>> users mailing list >>>>> [email protected] >>>>> http://lists.openshift.redhat.com/openshiftmm/listinfo/users >>>>> >>>>> >>>> >>
_______________________________________________ users mailing list [email protected] http://lists.openshift.redhat.com/openshiftmm/listinfo/users
