tokers commented on a change in pull request #479: URL: https://github.com/apache/apisix-ingress-controller/pull/479#discussion_r637878378
########## File path: docs/en/latest/practices/the-hard-way.md ########## @@ -0,0 +1,826 @@ +--- +title: APISIX Ingress Controller the Hard Way +--- + +<!-- +# +# Licensed to the Apache Software Foundation (ASF) under one or more +# contributor license agreements. See the NOTICE file distributed with +# this work for additional information regarding copyright ownership. +# The ASF licenses this file to You under the Apache License, Version 2.0 +# (the "License"); you may not use this file except in compliance with +# the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# +--> + +In this tutorial, we will install APISIX and APISIX Ingress Controller in Kubernetes from native yaml. + +## Prerequisites + +If you don't have a Kubernetes cluster to use, we recommend you to use [KiND](https://kind.sigs.k8s.io/docs/user/quick-start/) to create a local Kubernetes cluster. + +```bash +kubectl create ns apisix +``` + +In this tutorial, all our operations will be performed at namespace `apisix`. + +## ETCD Installation + +Here, we will deploy a single-node ETCD cluster without authentication inside the Kubernetes cluster. + +In this case, we assume you have a storage provisioner. If you are using KiND, a local path provisioner will be created automatically. If you don't have a storage provisioner or don't want to use persistence volume, you could use an `emptyDir` as volume. + +```yaml +# etcd-headless.yaml +apiVersion: v1 +kind: Service +metadata: + name: etcd-headless + labels: + app.kubernetes.io/name: etcd + annotations: + service.alpha.kubernetes.io/tolerate-unready-endpoints: "true" +spec: + type: ClusterIP + clusterIP: None + ports: + - name: "client" + port: 2379 + targetPort: client + - name: "peer" + port: 2380 + targetPort: peer + selector: + app.kubernetes.io/name: etcd +--- +# etcd.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: etcd + labels: + app.kubernetes.io/name: etcd +spec: + selector: + matchLabels: + app.kubernetes.io/name: etcd + serviceName: etcd-headless + podManagementPolicy: Parallel + replicas: 1 + updateStrategy: + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: etcd + spec: + securityContext: + fsGroup: 1001 + runAsUser: 1001 + containers: + - name: etcd + image: docker.io/bitnami/etcd:3.4.14-debian-10-r0 + imagePullPolicy: "IfNotPresent" + # command: + # - /scripts/setup.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: MY_POD_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: MY_POD_NAME + valueFrom: + fieldRef: + fieldPath: metadata.name + - name: ETCDCTL_API + value: "3" + - name: ETCD_NAME + value: "$(MY_POD_NAME)" + - name: ETCD_DATA_DIR + value: /etcd/data + - name: ETCD_ADVERTISE_CLIENT_URLS + value: "http://$(MY_POD_NAME).etcd-headless.apisix.svc.cluster.local:2379" + - name: ETCD_LISTEN_CLIENT_URLS + value: "http://0.0.0.0:2379" + - name: ETCD_INITIAL_ADVERTISE_PEER_URLS + value: "http://$(MY_POD_NAME).etcd-headless.apisix.svc.cluster.local:2380" + - name: ETCD_LISTEN_PEER_URLS + value: "http://0.0.0.0:2380" + - name: ALLOW_NONE_AUTHENTICATION + value: "yes" + ports: + - name: client + containerPort: 2379 + - name: peer + containerPort: 2380 + volumeMounts: + - name: data + mountPath: /etcd + volumeClaimTemplates: + - metadata: + name: data + spec: + accessModes: + - "ReadWriteOnce" + resources: + requests: + storage: "8Gi" +``` + +Apply these two yaml files to Kubernetes, wait few seconds, etcd installation should be successful. We could run a health check to ensure that. + +```bash +$ kubectl -n apisix exec -it etcd-0 -- etcdctl endpoint health +127.0.0.1:2379 is healthy: successfully committed proposal: took = 1.741883ms +``` + +Please notice that this etcd installation is quite simple and lack of many necessary production features, it should only be used for learning case. If you want to deploy a production-ready etcd, please refer [bitnami/etcd](https://bitnami.com/stack/etcd/helm). Review comment: ```suggestion Please notice that this etcd installation is quite simple and lack of many necessary production features, it should only be used for learning case. If you want to deploy a production-ready etcd, please refer to [bitnami/etcd](https://bitnami.com/stack/etcd/helm). ``` ########## File path: docs/en/latest/practices/the-hard-way.md ########## @@ -0,0 +1,826 @@ +--- +title: APISIX Ingress Controller the Hard Way +--- + +<!-- +# +# Licensed to the Apache Software Foundation (ASF) under one or more +# contributor license agreements. See the NOTICE file distributed with +# this work for additional information regarding copyright ownership. +# The ASF licenses this file to You under the Apache License, Version 2.0 +# (the "License"); you may not use this file except in compliance with +# the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# +--> + +In this tutorial, we will install APISIX and APISIX Ingress Controller in Kubernetes from native yaml. + +## Prerequisites + +If you don't have a Kubernetes cluster to use, we recommend you to use [KiND](https://kind.sigs.k8s.io/docs/user/quick-start/) to create a local Kubernetes cluster. + +```bash +kubectl create ns apisix +``` + +In this tutorial, all our operations will be performed at namespace `apisix`. + +## ETCD Installation + +Here, we will deploy a single-node ETCD cluster without authentication inside the Kubernetes cluster. + +In this case, we assume you have a storage provisioner. If you are using KiND, a local path provisioner will be created automatically. If you don't have a storage provisioner or don't want to use persistence volume, you could use an `emptyDir` as volume. + +```yaml +# etcd-headless.yaml +apiVersion: v1 +kind: Service +metadata: + name: etcd-headless + labels: + app.kubernetes.io/name: etcd + annotations: + service.alpha.kubernetes.io/tolerate-unready-endpoints: "true" +spec: + type: ClusterIP + clusterIP: None + ports: + - name: "client" + port: 2379 + targetPort: client + - name: "peer" + port: 2380 + targetPort: peer + selector: + app.kubernetes.io/name: etcd +--- +# etcd.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: etcd + labels: + app.kubernetes.io/name: etcd +spec: + selector: + matchLabels: + app.kubernetes.io/name: etcd + serviceName: etcd-headless + podManagementPolicy: Parallel + replicas: 1 + updateStrategy: + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: etcd + spec: + securityContext: + fsGroup: 1001 + runAsUser: 1001 + containers: + - name: etcd + image: docker.io/bitnami/etcd:3.4.14-debian-10-r0 + imagePullPolicy: "IfNotPresent" + # command: + # - /scripts/setup.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: MY_POD_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: MY_POD_NAME + valueFrom: + fieldRef: + fieldPath: metadata.name + - name: ETCDCTL_API + value: "3" + - name: ETCD_NAME + value: "$(MY_POD_NAME)" + - name: ETCD_DATA_DIR + value: /etcd/data + - name: ETCD_ADVERTISE_CLIENT_URLS + value: "http://$(MY_POD_NAME).etcd-headless.apisix.svc.cluster.local:2379" + - name: ETCD_LISTEN_CLIENT_URLS + value: "http://0.0.0.0:2379" + - name: ETCD_INITIAL_ADVERTISE_PEER_URLS + value: "http://$(MY_POD_NAME).etcd-headless.apisix.svc.cluster.local:2380" + - name: ETCD_LISTEN_PEER_URLS + value: "http://0.0.0.0:2380" + - name: ALLOW_NONE_AUTHENTICATION + value: "yes" + ports: + - name: client + containerPort: 2379 + - name: peer + containerPort: 2380 + volumeMounts: + - name: data + mountPath: /etcd + volumeClaimTemplates: + - metadata: + name: data + spec: + accessModes: + - "ReadWriteOnce" + resources: + requests: + storage: "8Gi" +``` + +Apply these two yaml files to Kubernetes, wait few seconds, etcd installation should be successful. We could run a health check to ensure that. + +```bash +$ kubectl -n apisix exec -it etcd-0 -- etcdctl endpoint health +127.0.0.1:2379 is healthy: successfully committed proposal: took = 1.741883ms +``` + +Please notice that this etcd installation is quite simple and lack of many necessary production features, it should only be used for learning case. If you want to deploy a production-ready etcd, please refer [bitnami/etcd](https://bitnami.com/stack/etcd/helm). + +## APISIX Installation + +Create a config file for our APISIX. We are going to deploy APISIX version 2.5. + +Note that the APISIX ingress controller needs to communicate with the APISIX admin API, so we set `apisix.allow_admin` to `0.0.0.0/0` for test. + +```yaml +apisix: + node_listen: 9080 # APISIX listening port + enable_heartbeat: true + enable_admin: true + enable_admin_cors: true + enable_debug: false + enable_dev_mode: false # Sets nginx worker_processes to 1 if set to true + enable_reuseport: true # Enable nginx SO_REUSEPORT switch if set to true. + enable_ipv6: true + config_center: etcd # etcd: use etcd to store the config value + + allow_admin: # http://nginx.org/en/docs/http/ngx_http_access_module.html#allow + - 0.0.0.0/0 + port_admin: 9180 + + # Default token when use API to call for Admin API. + # *NOTE*: Highly recommended to modify this value to protect APISIX's Admin API. + # Disabling this configuration item means that the Admin API does not + # require any authentication. + admin_key: + # admin: can everything for configuration data + - name: "admin" + key: edd1c9f034335f136f87ad84b625c8f1 + role: admin + # viewer: only can view configuration data + - name: "viewer" + key: 4054f7cf07e344346cd3f287985e76a2 + role: viewer + # dns_resolver: + # - 127.0.0.1 + dns_resolver_valid: 30 + resolver_timeout: 5 + +nginx_config: # config for render the template to genarate nginx.conf + error_log: "/dev/stderr" + error_log_level: "warn" # warn,error + worker_rlimit_nofile: 20480 # the number of files a worker process can open, should be larger than worker_connections + event: + worker_connections: 10620 + http: + access_log: "/dev/stdout" + keepalive_timeout: 60s # timeout during which a keep-alive client connection will stay open on the server side. + client_header_timeout: 60s # timeout for reading client request header, then 408 (Request Time-out) error is returned to the client + client_body_timeout: 60s # timeout for reading client request body, then 408 (Request Time-out) error is returned to the client + send_timeout: 10s # timeout for transmitting a response to the client.then the connection is closed + underscores_in_headers: "on" # default enables the use of underscores in client request header fields + real_ip_header: "X-Real-IP" # http://nginx.org/en/docs/http/ngx_http_realip_module.html#real_ip_header + real_ip_from: # http://nginx.org/en/docs/http/ngx_http_realip_module.html#set_real_ip_from + - 127.0.0.1 + - 'unix:' + +etcd: + host: + - "http://etcd-headless.apisix.svc.cluster.local:2379" + prefix: "/apisix" # apisix configurations prefix + timeout: 30 # seconds +plugins: # plugin list + - api-breaker + - authz-keycloak + - basic-auth + - batch-requests + - consumer-restriction + - cors + - echo + - fault-injection + - grpc-transcode + - hmac-auth + - http-logger + - ip-restriction + - jwt-auth + - kafka-logger + - key-auth + - limit-conn + - limit-count + - limit-req + - node-status + - openid-connect + - prometheus + - proxy-cache + - proxy-mirror + - proxy-rewrite + - redirect + - referer-restriction + - request-id + - request-validation + - response-rewrite + - serverless-post-function + - serverless-pre-function + - sls-logger + - syslog + - tcp-logger + - udp-logger + - uri-blocker + - wolf-rbac + - zipkin + - traffic-split +stream_plugins: + - mqtt-proxy +``` + +Please make sure `etcd.host` matches the headless service we created at first. In our case, it's `http://etcd-headless.apisix.svc.cluster.local:2379`. + +In this config, we defined an access key with the `admin` name under the `apisix.admin_key` section. This key is our API key, will be used to control APISIX later. + +Save this as `config.yaml`, then run `kubectl -n apisix create cm apisix-conf --from-file ./config.yaml` to create configmap. Later we will mount this configmap into APISIX deployment. + +```yaml +apiVersion: apps/v1 +kind: Deployment +metadata: + name: apisix + labels: + app.kubernetes.io/name: apisix +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: apisix + template: + metadata: + labels: + app.kubernetes.io/name: apisix + spec: + containers: + - name: apisix + image: "apache/apisix:2.5-alpine" + imagePullPolicy: IfNotPresent + ports: + - name: http + containerPort: 9080 + protocol: TCP + - name: tls + containerPort: 9443 + protocol: TCP + - name: admin + containerPort: 9180 + protocol: TCP + readinessProbe: + failureThreshold: 6 + initialDelaySeconds: 10 + periodSeconds: 10 + successThreshold: 1 + tcpSocket: + port: 9080 + timeoutSeconds: 1 + lifecycle: + preStop: + exec: + command: + - /bin/sh + - -c + - "sleep 30" + volumeMounts: + - mountPath: /usr/local/apisix/conf/config.yaml + name: apisix-config + subPath: config.yaml + resources: {} + volumes: + - configMap: + name: apisix-conf + name: apisix-config +``` + +Now, APISIX should be ready to use. Use `kubectl get pods -n apisix -l app.kubernetes.io/name=apisix -o name` to list APISIX pod name. Here we assume the pod name is `apisix-7644966c4d-cl4k6`. + +Let's have a check: + +```bash +kubectl -n apisix exec -it apisix-7644966c4d-cl4k6 -- curl http://127.0.0.1:9080 +``` + +If you are using Linux or macOS, run the command below in bash: + +```bash +kubectl -n apisix exec -it $(kubectl get pods -n apisix -l app.kubernetes.io/name=apisix -o name) -- curl http://127.0.0.1:9080 +``` + +If APISIX works properly, it should output: `{"error_msg":"404 Route Not Found"}`. Because we haven't defined any route yet. + +## HTTPBIN service + +Before configuring the APISIX, we need to create a test service. We use [kennethreitz/httpbin] here. We put this httpbin service in `demo` namespace. Review comment: Incomplete hyperlink definition? ########## File path: docs/en/latest/practices/the-hard-way.md ########## @@ -0,0 +1,826 @@ +--- +title: APISIX Ingress Controller the Hard Way +--- + +<!-- +# +# Licensed to the Apache Software Foundation (ASF) under one or more +# contributor license agreements. See the NOTICE file distributed with +# this work for additional information regarding copyright ownership. +# The ASF licenses this file to You under the Apache License, Version 2.0 +# (the "License"); you may not use this file except in compliance with +# the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# +--> + +In this tutorial, we will install APISIX and APISIX Ingress Controller in Kubernetes from native yaml. + +## Prerequisites + +If you don't have a Kubernetes cluster to use, we recommend you to use [KiND](https://kind.sigs.k8s.io/docs/user/quick-start/) to create a local Kubernetes cluster. + +```bash +kubectl create ns apisix +``` + +In this tutorial, all our operations will be performed at namespace `apisix`. + +## ETCD Installation + +Here, we will deploy a single-node ETCD cluster without authentication inside the Kubernetes cluster. + +In this case, we assume you have a storage provisioner. If you are using KiND, a local path provisioner will be created automatically. If you don't have a storage provisioner or don't want to use persistence volume, you could use an `emptyDir` as volume. + +```yaml +# etcd-headless.yaml +apiVersion: v1 +kind: Service +metadata: + name: etcd-headless + labels: + app.kubernetes.io/name: etcd + annotations: + service.alpha.kubernetes.io/tolerate-unready-endpoints: "true" +spec: + type: ClusterIP + clusterIP: None + ports: + - name: "client" + port: 2379 + targetPort: client + - name: "peer" + port: 2380 + targetPort: peer + selector: + app.kubernetes.io/name: etcd +--- +# etcd.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: etcd + labels: + app.kubernetes.io/name: etcd +spec: + selector: + matchLabels: + app.kubernetes.io/name: etcd + serviceName: etcd-headless + podManagementPolicy: Parallel + replicas: 1 + updateStrategy: + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: etcd + spec: + securityContext: + fsGroup: 1001 + runAsUser: 1001 + containers: + - name: etcd + image: docker.io/bitnami/etcd:3.4.14-debian-10-r0 + imagePullPolicy: "IfNotPresent" + # command: + # - /scripts/setup.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: MY_POD_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: MY_POD_NAME + valueFrom: + fieldRef: + fieldPath: metadata.name + - name: ETCDCTL_API + value: "3" + - name: ETCD_NAME + value: "$(MY_POD_NAME)" + - name: ETCD_DATA_DIR + value: /etcd/data + - name: ETCD_ADVERTISE_CLIENT_URLS + value: "http://$(MY_POD_NAME).etcd-headless.apisix.svc.cluster.local:2379" + - name: ETCD_LISTEN_CLIENT_URLS + value: "http://0.0.0.0:2379" + - name: ETCD_INITIAL_ADVERTISE_PEER_URLS + value: "http://$(MY_POD_NAME).etcd-headless.apisix.svc.cluster.local:2380" + - name: ETCD_LISTEN_PEER_URLS + value: "http://0.0.0.0:2380" + - name: ALLOW_NONE_AUTHENTICATION + value: "yes" + ports: + - name: client + containerPort: 2379 + - name: peer + containerPort: 2380 + volumeMounts: + - name: data + mountPath: /etcd + volumeClaimTemplates: + - metadata: + name: data + spec: + accessModes: + - "ReadWriteOnce" + resources: + requests: + storage: "8Gi" +``` + +Apply these two yaml files to Kubernetes, wait few seconds, etcd installation should be successful. We could run a health check to ensure that. + +```bash +$ kubectl -n apisix exec -it etcd-0 -- etcdctl endpoint health +127.0.0.1:2379 is healthy: successfully committed proposal: took = 1.741883ms +``` + +Please notice that this etcd installation is quite simple and lack of many necessary production features, it should only be used for learning case. If you want to deploy a production-ready etcd, please refer [bitnami/etcd](https://bitnami.com/stack/etcd/helm). + +## APISIX Installation + +Create a config file for our APISIX. We are going to deploy APISIX version 2.5. + +Note that the APISIX ingress controller needs to communicate with the APISIX admin API, so we set `apisix.allow_admin` to `0.0.0.0/0` for test. + +```yaml +apisix: + node_listen: 9080 # APISIX listening port + enable_heartbeat: true + enable_admin: true + enable_admin_cors: true + enable_debug: false + enable_dev_mode: false # Sets nginx worker_processes to 1 if set to true + enable_reuseport: true # Enable nginx SO_REUSEPORT switch if set to true. + enable_ipv6: true + config_center: etcd # etcd: use etcd to store the config value + + allow_admin: # http://nginx.org/en/docs/http/ngx_http_access_module.html#allow + - 0.0.0.0/0 + port_admin: 9180 + + # Default token when use API to call for Admin API. + # *NOTE*: Highly recommended to modify this value to protect APISIX's Admin API. + # Disabling this configuration item means that the Admin API does not + # require any authentication. + admin_key: + # admin: can everything for configuration data + - name: "admin" + key: edd1c9f034335f136f87ad84b625c8f1 + role: admin + # viewer: only can view configuration data + - name: "viewer" + key: 4054f7cf07e344346cd3f287985e76a2 + role: viewer + # dns_resolver: + # - 127.0.0.1 + dns_resolver_valid: 30 + resolver_timeout: 5 + +nginx_config: # config for render the template to genarate nginx.conf + error_log: "/dev/stderr" + error_log_level: "warn" # warn,error + worker_rlimit_nofile: 20480 # the number of files a worker process can open, should be larger than worker_connections + event: + worker_connections: 10620 + http: + access_log: "/dev/stdout" + keepalive_timeout: 60s # timeout during which a keep-alive client connection will stay open on the server side. + client_header_timeout: 60s # timeout for reading client request header, then 408 (Request Time-out) error is returned to the client + client_body_timeout: 60s # timeout for reading client request body, then 408 (Request Time-out) error is returned to the client + send_timeout: 10s # timeout for transmitting a response to the client.then the connection is closed + underscores_in_headers: "on" # default enables the use of underscores in client request header fields + real_ip_header: "X-Real-IP" # http://nginx.org/en/docs/http/ngx_http_realip_module.html#real_ip_header + real_ip_from: # http://nginx.org/en/docs/http/ngx_http_realip_module.html#set_real_ip_from + - 127.0.0.1 + - 'unix:' + +etcd: + host: + - "http://etcd-headless.apisix.svc.cluster.local:2379" + prefix: "/apisix" # apisix configurations prefix + timeout: 30 # seconds +plugins: # plugin list + - api-breaker + - authz-keycloak + - basic-auth + - batch-requests + - consumer-restriction + - cors + - echo + - fault-injection + - grpc-transcode + - hmac-auth + - http-logger + - ip-restriction + - jwt-auth + - kafka-logger + - key-auth + - limit-conn + - limit-count + - limit-req + - node-status + - openid-connect + - prometheus + - proxy-cache + - proxy-mirror + - proxy-rewrite + - redirect + - referer-restriction + - request-id + - request-validation + - response-rewrite + - serverless-post-function + - serverless-pre-function + - sls-logger + - syslog + - tcp-logger + - udp-logger + - uri-blocker + - wolf-rbac + - zipkin + - traffic-split +stream_plugins: + - mqtt-proxy +``` + +Please make sure `etcd.host` matches the headless service we created at first. In our case, it's `http://etcd-headless.apisix.svc.cluster.local:2379`. + +In this config, we defined an access key with the `admin` name under the `apisix.admin_key` section. This key is our API key, will be used to control APISIX later. + +Save this as `config.yaml`, then run `kubectl -n apisix create cm apisix-conf --from-file ./config.yaml` to create configmap. Later we will mount this configmap into APISIX deployment. + +```yaml +apiVersion: apps/v1 +kind: Deployment +metadata: + name: apisix + labels: + app.kubernetes.io/name: apisix +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: apisix + template: + metadata: + labels: + app.kubernetes.io/name: apisix + spec: + containers: + - name: apisix + image: "apache/apisix:2.5-alpine" + imagePullPolicy: IfNotPresent + ports: + - name: http + containerPort: 9080 + protocol: TCP + - name: tls + containerPort: 9443 + protocol: TCP + - name: admin + containerPort: 9180 + protocol: TCP + readinessProbe: + failureThreshold: 6 + initialDelaySeconds: 10 + periodSeconds: 10 + successThreshold: 1 + tcpSocket: + port: 9080 + timeoutSeconds: 1 + lifecycle: + preStop: + exec: + command: + - /bin/sh + - -c + - "sleep 30" + volumeMounts: + - mountPath: /usr/local/apisix/conf/config.yaml + name: apisix-config + subPath: config.yaml + resources: {} + volumes: + - configMap: + name: apisix-conf + name: apisix-config +``` + +Now, APISIX should be ready to use. Use `kubectl get pods -n apisix -l app.kubernetes.io/name=apisix -o name` to list APISIX pod name. Here we assume the pod name is `apisix-7644966c4d-cl4k6`. + +Let's have a check: + +```bash +kubectl -n apisix exec -it apisix-7644966c4d-cl4k6 -- curl http://127.0.0.1:9080 +``` + +If you are using Linux or macOS, run the command below in bash: + +```bash +kubectl -n apisix exec -it $(kubectl get pods -n apisix -l app.kubernetes.io/name=apisix -o name) -- curl http://127.0.0.1:9080 +``` + +If APISIX works properly, it should output: `{"error_msg":"404 Route Not Found"}`. Because we haven't defined any route yet. + +## HTTPBIN service + +Before configuring the APISIX, we need to create a test service. We use [kennethreitz/httpbin] here. We put this httpbin service in `demo` namespace. + +```bash +kubectl create ns demo +kubectl -n demo run httpbin --image-pull-policy=IfNotPresent --image kennethreitz/httpbin --port 80 +kubectl -n demo expose pod httpbin --port 80 +``` + +After the httpbin service started, we should be able to access it inside the APISIX pod via service. + +```bash +kubectl -n apisix exec -it apisix-7644966c4d-cl4k6 -- curl http://httpbin.demo/get +``` + +This should output the request's query parameters, for example: + +```json +{ + "args": {}, + "headers": { + "Accept": "*/*", + "Host": "httpbin.demo", + "User-Agent": "curl/7.67.0" + }, + "origin": "172.17.0.1", + "url": "http://httpbin.demo/get" +} +``` + +To read more, please refer to [Getting Started](https://apisix.apache.org/docs/apisix/getting-started). + +## Define Route + +Now, we can define the route for APISIX. + +Assuming we want to route all traffic which URI has `/httpbin` prefix and the request contains `Host: httpbin.org` header. + +Please notice that the admin port is `9180`. + +```bash +kubectl -n apisix exec -it apisix-7644966c4d-cl4k6 -- curl "http://127.0.0.1:9180/apisix/admin/routes/1" -H "X-API-KEY: edd1c9f034335f136f87ad84b625c8f1" -X PUT -d ' +{ + "uri": "/*", + "host": "httpbin.org", + "upstream": { + "type": "roundrobin", + "nodes": { + "httpbin.demo:80": 1 + } + } +}' +``` + +The output would be like this: + +```json +{"action":"set","node":{"key":"\/apisix\/routes\/1","value":{"status":1,"create_time":1621408897,"upstream":{"pass_host":"pass","type":"roundrobin","hash_on":"vars","nodes":{"httpbin.demo:80":1},"scheme":"http"},"update_time":1621408897,"priority":0,"host":"httpbin.org","id":"1","uri":"\/*"}}} +``` + +We could check route rules by `GET /apisix/admin/routes`: + +```bash +kubectl -n apisix exec -it apisix-7644966c4d-cl4k6 -- curl "http://127.0.0.1:9180/apisix/admin/routes/1" -H "X-API-KEY: edd1c9f034335f136f87ad84b625c8f1" +``` + +It should output like this: + +```json +{"action":"get","node":{"key":"\/apisix\/routes\/1","value":{"upstream":{"pass_host":"pass","type":"roundrobin","scheme":"http","hash_on":"vars","nodes":{"httpbin.demo:80":1}},"id":"1","create_time":1621408897,"update_time":1621408897,"host":"httpbin.org","priority":0,"status":1,"uri":"\/*"}},"count":"1"} +``` + +Now, we can test the routing rule: + +```bash +kubectl -n apisix exec -it apisix-7644966c4d-cl4k6 -- curl "http://127.0.0.1:9080/get" -H 'Host: httpbin.org' +``` + +It will output like: + +```json +{ + "args": {}, + "headers": { + "Accept": "*/*", + "Host": "httpbin.org", + "User-Agent": "curl/7.67.0", + "X-Forwarded-Host": "httpbin.org" + }, + "origin": "127.0.0.1", + "url": "http://httpbin.org/get" +} +``` + +## Install APISIX Ingress Controller + +APISIX ingress controller can help you manage your configurations declaratively by using Kubernetes resources. Here we will install version 0.5.0. + +Currently, the APISIX ingress controller supports both official Ingress resource or APISIX's CustomResourceDefinitions, which includes ApisixRoute and ApisixUpstream. + +Before installing the APISIX controller, we need to create a service account and the corresponding ClusterRole to ensure that the APISIX ingress controller has sufficient permissions to access required resources. + +Here is an example config from [apisix-helm-chart](https://github.com/apache/apisix-helm-chart): + +```yaml +apiVersion: v1 +kind: ServiceAccount +metadata: + name: apisix-ingress-controller +--- +apiVersion: rbac.authorization.k8s.io/v1 +kind: ClusterRole +metadata: + name: apisix-clusterrole +rules: + - apiGroups: + - "" + resources: + - configmaps + - endpoints + - persistentvolumeclaims + - pods + - replicationcontrollers + - replicationcontrollers/scale + - serviceaccounts + - services + - secrets + verbs: + - get + - list + - watch + - apiGroups: + - "" + resources: + - bindings + - events + - limitranges + - namespaces/status + - pods/log + - pods/status + - replicationcontrollers/status + - resourcequotas + - resourcequotas/status + verbs: + - get + - list + - watch + - apiGroups: + - "" + resources: + - namespaces + verbs: + - get + - list + - watch + - apiGroups: + - apps + resources: + - controllerrevisions + - daemonsets + - deployments + - deployments/scale + - replicasets + - replicasets/scale + - statefulsets + - statefulsets/scale + verbs: + - get + - list + - watch + - apiGroups: + - autoscaling + resources: + - horizontalpodautoscalers + verbs: + - get + - list + - watch + - apiGroups: + - batch + resources: + - cronjobs + - jobs + verbs: + - get + - list + - watch + - apiGroups: + - extensions + resources: + - daemonsets + - deployments + - deployments/scale + - ingresses + - networkpolicies + - replicasets + - replicasets/scale + - replicationcontrollers/scale + verbs: + - get + - list + - watch + - apiGroups: + - policy + resources: + - poddisruptionbudgets + verbs: + - get + - list + - watch + - apiGroups: + - networking.k8s.io + resources: + - ingresses + - networkpolicies + verbs: + - get + - list + - watch + - apiGroups: + - metrics.k8s.io + resources: + - pods + verbs: + - get + - list + - watch + - apiGroups: + - apisix.apache.org + resources: + - apisixroutes + - apisixupstreams + - apisixservices + - apisixtlses + verbs: + - get + - list + - watch + - apiGroups: + - coordination.k8s.io + resources: + - leases + verbs: + - '*' +--- +apiVersion: rbac.authorization.k8s.io/v1 +kind: ClusterRoleBinding +metadata: + name: apisix-clusterrolebinding +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: ClusterRole + name: apisix-clusterrole +subjects: + - kind: ServiceAccount + name: apisix-ingress-controller + namespace: apisix +``` + +Then, we need to create ApisixRoute CRD: + +```yaml + +apiVersion: apiextensions.k8s.io/v1beta1 +kind: CustomResourceDefinition +metadata: + name: apisixroutes.apisix.apache.org +spec: + group: apisix.apache.org + versions: + - name: v1 + served: true + storage: false + - name: v2alpha1 + served: true + storage: true + scope: Namespaced + names: + plural: apisixroutes + singular: apisixroute + kind: ApisixRoute + shortNames: + - ar +--- +apiVersion: apiextensions.k8s.io/v1beta1 +kind: CustomResourceDefinition +metadata: + name: apisixtlses.apisix.apache.org +spec: + group: apisix.apache.org + versions: + - name: v1 + served: true + storage: true + scope: Namespaced + names: + plural: apisixtlses + singular: apisixtls + kind: ApisixTls + shortNames: + - atls +--- +apiVersion: apiextensions.k8s.io/v1beta1 +kind: CustomResourceDefinition +metadata: + name: apisixupstreams.apisix.apache.org +spec: + group: apisix.apache.org + versions: + - name: v1 + served: true + storage: true + scope: Namespaced + names: + plural: apisixupstreams + singular: apisixupstream + kind: ApisixUpstream + shortNames: + - au +``` + +To make the ingress controller works properly with APISIX, we need to create a config file containing the APISIX admin API URL and API key as below: + +```yaml +apiVersion: v1 +data: + config.yaml: | + # log options + log_level: "debug" + log_output: "stderr" + http_listen: ":8080" + enable_profiling: true + kubernetes: + kubeconfig: "" + resync_interval: "30s" + app_namespaces: + - "*" + ingress_class: "apisix" + ingress_version: "networking/v1" + apisix_route_version: "apisix.apache.org/v1" + apisix: + base_url: "http://apisix-admin.apisix:9180/apisix/admin" + admin_key: "edd1c9f034335f136f87ad84b625c8f1" +kind: ConfigMap +metadata: + name: apisix-configmap + labels: + app.kubernetes.io/name: ingress-controller +``` + +Because the ingress controller needs to access APISIX admin API, we need to create a service for APISIX. + +```yaml +apiVersion: v1 +kind: Service +metadata: + name: apisix-admin + labels: + app.kubernetes.io/name: apisix +spec: + type: ClusterIP + ports: + - name: apisix-admin + port: 9180 + targetPort: 9180 + protocol: TCP + selector: + app.kubernetes.io/name: apisix +``` + +Because currently APISIX ingress controller doesn't 100% compatible with APISIX, we need to delete the previously created route in case of some data structure mismatch. + +```bash +kubectl -n apisix exec -it $(kngp -l app.kubernetes.io/name=apisix -o name) -- curl "http://127.0.0.1:9180/apisix/admin/routes/1" -X DELETE -H "X-API-KEY: edd1c9f034335f136f87ad84b625c8f1" +``` + +After these configurations, we could deploy the ingress controller now. + +```yaml +apiVersion: apps/v1 +kind: Deployment +metadata: + name: apisix-ingress-controller + labels: + app.kubernetes.io/name: ingress-controller +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: ingress-controller + template: + metadata: + labels: + app.kubernetes.io/name: ingress-controller + spec: + serviceAccountName: apisix-ingress-controller + volumes: + - name: configuration + configMap: + name: apisix-configmap + items: + - key: config.yaml + path: config.yaml + containers: + - name: ingress-controller + command: + - /ingress-apisix/apisix-ingress-controller + - ingress + - --config-path + - /ingress-apisix/conf/config.yaml + image: "apache/apisix-ingress-controller:0.5.0" + imagePullPolicy: IfNotPresent + ports: + - name: http + containerPort: 8080 + protocol: TCP + livenessProbe: + httpGet: + path: /healthz + port: 8080 + readinessProbe: + httpGet: + path: /healthz + port: 8080 + resources: + {} + volumeMounts: + - mountPath: /ingress-apisix/conf + name: configuration +``` + +In this deployment, we mount the configmap created above as a config file, and tell Kubernetes to use the service account `apisix-ingress-controller`. + +After the ingress controller status is converted to `Running`, we could create an ApisixRoute resource and observe its behaviors. + +Here is an example ApisixRoute: + +```yaml +apiVersion: apisix.apache.org/v1 Review comment: Use ApisixRoute v2alpha1, we'll deprecate v1 in the near future. ########## File path: docs/en/latest/practices/the-hard-way.md ########## @@ -0,0 +1,826 @@ +--- +title: APISIX Ingress Controller the Hard Way +--- + +<!-- +# +# Licensed to the Apache Software Foundation (ASF) under one or more +# contributor license agreements. See the NOTICE file distributed with +# this work for additional information regarding copyright ownership. +# The ASF licenses this file to You under the Apache License, Version 2.0 +# (the "License"); you may not use this file except in compliance with +# the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# +--> + +In this tutorial, we will install APISIX and APISIX Ingress Controller in Kubernetes from native yaml. + +## Prerequisites + +If you don't have a Kubernetes cluster to use, we recommend you to use [KiND](https://kind.sigs.k8s.io/docs/user/quick-start/) to create a local Kubernetes cluster. + +```bash +kubectl create ns apisix +``` + +In this tutorial, all our operations will be performed at namespace `apisix`. + +## ETCD Installation + +Here, we will deploy a single-node ETCD cluster without authentication inside the Kubernetes cluster. + +In this case, we assume you have a storage provisioner. If you are using KiND, a local path provisioner will be created automatically. If you don't have a storage provisioner or don't want to use persistence volume, you could use an `emptyDir` as volume. + +```yaml +# etcd-headless.yaml +apiVersion: v1 +kind: Service +metadata: + name: etcd-headless + labels: + app.kubernetes.io/name: etcd + annotations: + service.alpha.kubernetes.io/tolerate-unready-endpoints: "true" +spec: + type: ClusterIP + clusterIP: None + ports: + - name: "client" + port: 2379 + targetPort: client + - name: "peer" + port: 2380 + targetPort: peer + selector: + app.kubernetes.io/name: etcd +--- +# etcd.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: etcd + labels: + app.kubernetes.io/name: etcd +spec: + selector: + matchLabels: + app.kubernetes.io/name: etcd + serviceName: etcd-headless + podManagementPolicy: Parallel + replicas: 1 + updateStrategy: + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: etcd + spec: + securityContext: + fsGroup: 1001 + runAsUser: 1001 + containers: + - name: etcd + image: docker.io/bitnami/etcd:3.4.14-debian-10-r0 + imagePullPolicy: "IfNotPresent" + # command: + # - /scripts/setup.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: MY_POD_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: MY_POD_NAME + valueFrom: + fieldRef: + fieldPath: metadata.name + - name: ETCDCTL_API + value: "3" + - name: ETCD_NAME + value: "$(MY_POD_NAME)" + - name: ETCD_DATA_DIR + value: /etcd/data + - name: ETCD_ADVERTISE_CLIENT_URLS + value: "http://$(MY_POD_NAME).etcd-headless.apisix.svc.cluster.local:2379" + - name: ETCD_LISTEN_CLIENT_URLS + value: "http://0.0.0.0:2379" + - name: ETCD_INITIAL_ADVERTISE_PEER_URLS + value: "http://$(MY_POD_NAME).etcd-headless.apisix.svc.cluster.local:2380" + - name: ETCD_LISTEN_PEER_URLS + value: "http://0.0.0.0:2380" + - name: ALLOW_NONE_AUTHENTICATION + value: "yes" + ports: + - name: client + containerPort: 2379 + - name: peer + containerPort: 2380 + volumeMounts: + - name: data + mountPath: /etcd + volumeClaimTemplates: + - metadata: + name: data + spec: + accessModes: + - "ReadWriteOnce" + resources: + requests: + storage: "8Gi" +``` + +Apply these two yaml files to Kubernetes, wait few seconds, etcd installation should be successful. We could run a health check to ensure that. + +```bash +$ kubectl -n apisix exec -it etcd-0 -- etcdctl endpoint health +127.0.0.1:2379 is healthy: successfully committed proposal: took = 1.741883ms +``` + +Please notice that this etcd installation is quite simple and lack of many necessary production features, it should only be used for learning case. If you want to deploy a production-ready etcd, please refer [bitnami/etcd](https://bitnami.com/stack/etcd/helm). + +## APISIX Installation + +Create a config file for our APISIX. We are going to deploy APISIX version 2.5. + +Note that the APISIX ingress controller needs to communicate with the APISIX admin API, so we set `apisix.allow_admin` to `0.0.0.0/0` for test. + +```yaml +apisix: + node_listen: 9080 # APISIX listening port + enable_heartbeat: true + enable_admin: true + enable_admin_cors: true + enable_debug: false + enable_dev_mode: false # Sets nginx worker_processes to 1 if set to true + enable_reuseport: true # Enable nginx SO_REUSEPORT switch if set to true. + enable_ipv6: true + config_center: etcd # etcd: use etcd to store the config value + + allow_admin: # http://nginx.org/en/docs/http/ngx_http_access_module.html#allow + - 0.0.0.0/0 + port_admin: 9180 + + # Default token when use API to call for Admin API. + # *NOTE*: Highly recommended to modify this value to protect APISIX's Admin API. + # Disabling this configuration item means that the Admin API does not + # require any authentication. + admin_key: + # admin: can everything for configuration data + - name: "admin" + key: edd1c9f034335f136f87ad84b625c8f1 + role: admin + # viewer: only can view configuration data + - name: "viewer" + key: 4054f7cf07e344346cd3f287985e76a2 + role: viewer + # dns_resolver: + # - 127.0.0.1 + dns_resolver_valid: 30 + resolver_timeout: 5 + +nginx_config: # config for render the template to genarate nginx.conf + error_log: "/dev/stderr" + error_log_level: "warn" # warn,error + worker_rlimit_nofile: 20480 # the number of files a worker process can open, should be larger than worker_connections + event: + worker_connections: 10620 + http: + access_log: "/dev/stdout" + keepalive_timeout: 60s # timeout during which a keep-alive client connection will stay open on the server side. + client_header_timeout: 60s # timeout for reading client request header, then 408 (Request Time-out) error is returned to the client + client_body_timeout: 60s # timeout for reading client request body, then 408 (Request Time-out) error is returned to the client + send_timeout: 10s # timeout for transmitting a response to the client.then the connection is closed + underscores_in_headers: "on" # default enables the use of underscores in client request header fields + real_ip_header: "X-Real-IP" # http://nginx.org/en/docs/http/ngx_http_realip_module.html#real_ip_header + real_ip_from: # http://nginx.org/en/docs/http/ngx_http_realip_module.html#set_real_ip_from + - 127.0.0.1 + - 'unix:' + +etcd: + host: + - "http://etcd-headless.apisix.svc.cluster.local:2379" + prefix: "/apisix" # apisix configurations prefix + timeout: 30 # seconds +plugins: # plugin list + - api-breaker + - authz-keycloak + - basic-auth + - batch-requests + - consumer-restriction + - cors + - echo + - fault-injection + - grpc-transcode + - hmac-auth + - http-logger + - ip-restriction + - jwt-auth + - kafka-logger + - key-auth + - limit-conn + - limit-count + - limit-req + - node-status + - openid-connect + - prometheus + - proxy-cache + - proxy-mirror + - proxy-rewrite + - redirect + - referer-restriction + - request-id + - request-validation + - response-rewrite + - serverless-post-function + - serverless-pre-function + - sls-logger + - syslog + - tcp-logger + - udp-logger + - uri-blocker + - wolf-rbac + - zipkin + - traffic-split +stream_plugins: + - mqtt-proxy +``` + +Please make sure `etcd.host` matches the headless service we created at first. In our case, it's `http://etcd-headless.apisix.svc.cluster.local:2379`. + +In this config, we defined an access key with the `admin` name under the `apisix.admin_key` section. This key is our API key, will be used to control APISIX later. + +Save this as `config.yaml`, then run `kubectl -n apisix create cm apisix-conf --from-file ./config.yaml` to create configmap. Later we will mount this configmap into APISIX deployment. + +```yaml +apiVersion: apps/v1 +kind: Deployment +metadata: + name: apisix + labels: + app.kubernetes.io/name: apisix +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: apisix + template: + metadata: + labels: + app.kubernetes.io/name: apisix + spec: + containers: + - name: apisix + image: "apache/apisix:2.5-alpine" + imagePullPolicy: IfNotPresent + ports: + - name: http + containerPort: 9080 + protocol: TCP + - name: tls + containerPort: 9443 + protocol: TCP + - name: admin + containerPort: 9180 + protocol: TCP + readinessProbe: + failureThreshold: 6 + initialDelaySeconds: 10 + periodSeconds: 10 + successThreshold: 1 + tcpSocket: + port: 9080 + timeoutSeconds: 1 + lifecycle: + preStop: + exec: + command: + - /bin/sh + - -c + - "sleep 30" + volumeMounts: + - mountPath: /usr/local/apisix/conf/config.yaml + name: apisix-config + subPath: config.yaml + resources: {} + volumes: + - configMap: + name: apisix-conf + name: apisix-config +``` + +Now, APISIX should be ready to use. Use `kubectl get pods -n apisix -l app.kubernetes.io/name=apisix -o name` to list APISIX pod name. Here we assume the pod name is `apisix-7644966c4d-cl4k6`. + +Let's have a check: + +```bash +kubectl -n apisix exec -it apisix-7644966c4d-cl4k6 -- curl http://127.0.0.1:9080 +``` + +If you are using Linux or macOS, run the command below in bash: + +```bash +kubectl -n apisix exec -it $(kubectl get pods -n apisix -l app.kubernetes.io/name=apisix -o name) -- curl http://127.0.0.1:9080 +``` + +If APISIX works properly, it should output: `{"error_msg":"404 Route Not Found"}`. Because we haven't defined any route yet. + +## HTTPBIN service + +Before configuring the APISIX, we need to create a test service. We use [kennethreitz/httpbin] here. We put this httpbin service in `demo` namespace. + +```bash +kubectl create ns demo +kubectl -n demo run httpbin --image-pull-policy=IfNotPresent --image kennethreitz/httpbin --port 80 +kubectl -n demo expose pod httpbin --port 80 +``` + +After the httpbin service started, we should be able to access it inside the APISIX pod via service. + +```bash +kubectl -n apisix exec -it apisix-7644966c4d-cl4k6 -- curl http://httpbin.demo/get +``` + +This should output the request's query parameters, for example: + +```json +{ + "args": {}, + "headers": { + "Accept": "*/*", + "Host": "httpbin.demo", + "User-Agent": "curl/7.67.0" + }, + "origin": "172.17.0.1", + "url": "http://httpbin.demo/get" +} +``` + +To read more, please refer to [Getting Started](https://apisix.apache.org/docs/apisix/getting-started). + +## Define Route + +Now, we can define the route for APISIX. + +Assuming we want to route all traffic which URI has `/httpbin` prefix and the request contains `Host: httpbin.org` header. + +Please notice that the admin port is `9180`. + +```bash +kubectl -n apisix exec -it apisix-7644966c4d-cl4k6 -- curl "http://127.0.0.1:9180/apisix/admin/routes/1" -H "X-API-KEY: edd1c9f034335f136f87ad84b625c8f1" -X PUT -d ' +{ + "uri": "/*", + "host": "httpbin.org", + "upstream": { + "type": "roundrobin", + "nodes": { + "httpbin.demo:80": 1 + } + } +}' +``` + +The output would be like this: + +```json +{"action":"set","node":{"key":"\/apisix\/routes\/1","value":{"status":1,"create_time":1621408897,"upstream":{"pass_host":"pass","type":"roundrobin","hash_on":"vars","nodes":{"httpbin.demo:80":1},"scheme":"http"},"update_time":1621408897,"priority":0,"host":"httpbin.org","id":"1","uri":"\/*"}}} +``` + +We could check route rules by `GET /apisix/admin/routes`: + +```bash +kubectl -n apisix exec -it apisix-7644966c4d-cl4k6 -- curl "http://127.0.0.1:9180/apisix/admin/routes/1" -H "X-API-KEY: edd1c9f034335f136f87ad84b625c8f1" +``` + +It should output like this: + +```json +{"action":"get","node":{"key":"\/apisix\/routes\/1","value":{"upstream":{"pass_host":"pass","type":"roundrobin","scheme":"http","hash_on":"vars","nodes":{"httpbin.demo:80":1}},"id":"1","create_time":1621408897,"update_time":1621408897,"host":"httpbin.org","priority":0,"status":1,"uri":"\/*"}},"count":"1"} +``` + +Now, we can test the routing rule: + +```bash +kubectl -n apisix exec -it apisix-7644966c4d-cl4k6 -- curl "http://127.0.0.1:9080/get" -H 'Host: httpbin.org' +``` + +It will output like: + +```json +{ + "args": {}, + "headers": { + "Accept": "*/*", + "Host": "httpbin.org", + "User-Agent": "curl/7.67.0", + "X-Forwarded-Host": "httpbin.org" + }, + "origin": "127.0.0.1", + "url": "http://httpbin.org/get" +} +``` + +## Install APISIX Ingress Controller + +APISIX ingress controller can help you manage your configurations declaratively by using Kubernetes resources. Here we will install version 0.5.0. + +Currently, the APISIX ingress controller supports both official Ingress resource or APISIX's CustomResourceDefinitions, which includes ApisixRoute and ApisixUpstream. + +Before installing the APISIX controller, we need to create a service account and the corresponding ClusterRole to ensure that the APISIX ingress controller has sufficient permissions to access required resources. + +Here is an example config from [apisix-helm-chart](https://github.com/apache/apisix-helm-chart): + +```yaml +apiVersion: v1 +kind: ServiceAccount +metadata: + name: apisix-ingress-controller +--- +apiVersion: rbac.authorization.k8s.io/v1 +kind: ClusterRole +metadata: + name: apisix-clusterrole +rules: + - apiGroups: + - "" + resources: + - configmaps + - endpoints + - persistentvolumeclaims + - pods + - replicationcontrollers + - replicationcontrollers/scale + - serviceaccounts + - services + - secrets + verbs: + - get + - list + - watch + - apiGroups: + - "" + resources: + - bindings + - events + - limitranges + - namespaces/status + - pods/log + - pods/status + - replicationcontrollers/status + - resourcequotas + - resourcequotas/status + verbs: + - get + - list + - watch + - apiGroups: + - "" + resources: + - namespaces + verbs: + - get + - list + - watch + - apiGroups: + - apps + resources: + - controllerrevisions + - daemonsets + - deployments + - deployments/scale + - replicasets + - replicasets/scale + - statefulsets + - statefulsets/scale + verbs: + - get + - list + - watch + - apiGroups: + - autoscaling + resources: + - horizontalpodautoscalers + verbs: + - get + - list + - watch + - apiGroups: + - batch + resources: + - cronjobs + - jobs + verbs: + - get + - list + - watch + - apiGroups: + - extensions + resources: + - daemonsets + - deployments + - deployments/scale + - ingresses + - networkpolicies + - replicasets + - replicasets/scale + - replicationcontrollers/scale + verbs: + - get + - list + - watch + - apiGroups: + - policy + resources: + - poddisruptionbudgets + verbs: + - get + - list + - watch + - apiGroups: + - networking.k8s.io + resources: + - ingresses + - networkpolicies + verbs: + - get + - list + - watch + - apiGroups: + - metrics.k8s.io + resources: + - pods + verbs: + - get + - list + - watch + - apiGroups: + - apisix.apache.org + resources: + - apisixroutes + - apisixupstreams + - apisixservices + - apisixtlses + verbs: + - get + - list + - watch + - apiGroups: + - coordination.k8s.io + resources: + - leases + verbs: + - '*' +--- +apiVersion: rbac.authorization.k8s.io/v1 +kind: ClusterRoleBinding +metadata: + name: apisix-clusterrolebinding +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: ClusterRole + name: apisix-clusterrole +subjects: + - kind: ServiceAccount + name: apisix-ingress-controller + namespace: apisix +``` + +Then, we need to create ApisixRoute CRD: + +```yaml + +apiVersion: apiextensions.k8s.io/v1beta1 +kind: CustomResourceDefinition +metadata: + name: apisixroutes.apisix.apache.org +spec: + group: apisix.apache.org + versions: + - name: v1 + served: true + storage: false + - name: v2alpha1 + served: true + storage: true + scope: Namespaced + names: + plural: apisixroutes + singular: apisixroute + kind: ApisixRoute + shortNames: + - ar +--- +apiVersion: apiextensions.k8s.io/v1beta1 +kind: CustomResourceDefinition +metadata: + name: apisixtlses.apisix.apache.org +spec: + group: apisix.apache.org + versions: + - name: v1 + served: true + storage: true + scope: Namespaced + names: + plural: apisixtlses + singular: apisixtls + kind: ApisixTls + shortNames: + - atls +--- +apiVersion: apiextensions.k8s.io/v1beta1 +kind: CustomResourceDefinition +metadata: + name: apisixupstreams.apisix.apache.org +spec: + group: apisix.apache.org + versions: + - name: v1 + served: true + storage: true + scope: Namespaced + names: + plural: apisixupstreams + singular: apisixupstream + kind: ApisixUpstream + shortNames: + - au +``` Review comment: Add a note that the above is not the complete CRD resource list, see `samples/deploy/crd/v1beta1` for details. ########## File path: docs/en/latest/practices/the-hard-way.md ########## @@ -0,0 +1,826 @@ +--- +title: APISIX Ingress Controller the Hard Way +--- + +<!-- +# +# Licensed to the Apache Software Foundation (ASF) under one or more +# contributor license agreements. See the NOTICE file distributed with +# this work for additional information regarding copyright ownership. +# The ASF licenses this file to You under the Apache License, Version 2.0 +# (the "License"); you may not use this file except in compliance with +# the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# +--> + +In this tutorial, we will install APISIX and APISIX Ingress Controller in Kubernetes from native yaml. + +## Prerequisites + +If you don't have a Kubernetes cluster to use, we recommend you to use [KiND](https://kind.sigs.k8s.io/docs/user/quick-start/) to create a local Kubernetes cluster. + +```bash +kubectl create ns apisix +``` + +In this tutorial, all our operations will be performed at namespace `apisix`. + +## ETCD Installation + +Here, we will deploy a single-node ETCD cluster without authentication inside the Kubernetes cluster. + +In this case, we assume you have a storage provisioner. If you are using KiND, a local path provisioner will be created automatically. If you don't have a storage provisioner or don't want to use persistence volume, you could use an `emptyDir` as volume. + +```yaml +# etcd-headless.yaml +apiVersion: v1 +kind: Service +metadata: + name: etcd-headless + labels: + app.kubernetes.io/name: etcd + annotations: + service.alpha.kubernetes.io/tolerate-unready-endpoints: "true" +spec: + type: ClusterIP + clusterIP: None + ports: + - name: "client" + port: 2379 + targetPort: client + - name: "peer" + port: 2380 + targetPort: peer + selector: + app.kubernetes.io/name: etcd +--- +# etcd.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: etcd + labels: + app.kubernetes.io/name: etcd +spec: + selector: + matchLabels: + app.kubernetes.io/name: etcd + serviceName: etcd-headless + podManagementPolicy: Parallel + replicas: 1 + updateStrategy: + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: etcd + spec: + securityContext: + fsGroup: 1001 + runAsUser: 1001 + containers: + - name: etcd + image: docker.io/bitnami/etcd:3.4.14-debian-10-r0 + imagePullPolicy: "IfNotPresent" + # command: + # - /scripts/setup.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: MY_POD_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: MY_POD_NAME + valueFrom: + fieldRef: + fieldPath: metadata.name + - name: ETCDCTL_API + value: "3" + - name: ETCD_NAME + value: "$(MY_POD_NAME)" + - name: ETCD_DATA_DIR + value: /etcd/data + - name: ETCD_ADVERTISE_CLIENT_URLS + value: "http://$(MY_POD_NAME).etcd-headless.apisix.svc.cluster.local:2379" + - name: ETCD_LISTEN_CLIENT_URLS + value: "http://0.0.0.0:2379" + - name: ETCD_INITIAL_ADVERTISE_PEER_URLS + value: "http://$(MY_POD_NAME).etcd-headless.apisix.svc.cluster.local:2380" + - name: ETCD_LISTEN_PEER_URLS + value: "http://0.0.0.0:2380" + - name: ALLOW_NONE_AUTHENTICATION + value: "yes" + ports: + - name: client + containerPort: 2379 + - name: peer + containerPort: 2380 + volumeMounts: + - name: data + mountPath: /etcd + volumeClaimTemplates: + - metadata: + name: data + spec: + accessModes: + - "ReadWriteOnce" + resources: + requests: + storage: "8Gi" +``` + +Apply these two yaml files to Kubernetes, wait few seconds, etcd installation should be successful. We could run a health check to ensure that. + +```bash +$ kubectl -n apisix exec -it etcd-0 -- etcdctl endpoint health +127.0.0.1:2379 is healthy: successfully committed proposal: took = 1.741883ms +``` + +Please notice that this etcd installation is quite simple and lack of many necessary production features, it should only be used for learning case. If you want to deploy a production-ready etcd, please refer [bitnami/etcd](https://bitnami.com/stack/etcd/helm). + +## APISIX Installation + +Create a config file for our APISIX. We are going to deploy APISIX version 2.5. + +Note that the APISIX ingress controller needs to communicate with the APISIX admin API, so we set `apisix.allow_admin` to `0.0.0.0/0` for test. + +```yaml +apisix: + node_listen: 9080 # APISIX listening port + enable_heartbeat: true + enable_admin: true + enable_admin_cors: true + enable_debug: false + enable_dev_mode: false # Sets nginx worker_processes to 1 if set to true + enable_reuseport: true # Enable nginx SO_REUSEPORT switch if set to true. + enable_ipv6: true + config_center: etcd # etcd: use etcd to store the config value + + allow_admin: # http://nginx.org/en/docs/http/ngx_http_access_module.html#allow + - 0.0.0.0/0 + port_admin: 9180 + + # Default token when use API to call for Admin API. + # *NOTE*: Highly recommended to modify this value to protect APISIX's Admin API. + # Disabling this configuration item means that the Admin API does not + # require any authentication. + admin_key: + # admin: can everything for configuration data + - name: "admin" + key: edd1c9f034335f136f87ad84b625c8f1 + role: admin + # viewer: only can view configuration data + - name: "viewer" + key: 4054f7cf07e344346cd3f287985e76a2 + role: viewer + # dns_resolver: + # - 127.0.0.1 + dns_resolver_valid: 30 + resolver_timeout: 5 + +nginx_config: # config for render the template to genarate nginx.conf + error_log: "/dev/stderr" + error_log_level: "warn" # warn,error + worker_rlimit_nofile: 20480 # the number of files a worker process can open, should be larger than worker_connections + event: + worker_connections: 10620 + http: + access_log: "/dev/stdout" + keepalive_timeout: 60s # timeout during which a keep-alive client connection will stay open on the server side. + client_header_timeout: 60s # timeout for reading client request header, then 408 (Request Time-out) error is returned to the client + client_body_timeout: 60s # timeout for reading client request body, then 408 (Request Time-out) error is returned to the client + send_timeout: 10s # timeout for transmitting a response to the client.then the connection is closed + underscores_in_headers: "on" # default enables the use of underscores in client request header fields + real_ip_header: "X-Real-IP" # http://nginx.org/en/docs/http/ngx_http_realip_module.html#real_ip_header + real_ip_from: # http://nginx.org/en/docs/http/ngx_http_realip_module.html#set_real_ip_from + - 127.0.0.1 + - 'unix:' + +etcd: + host: + - "http://etcd-headless.apisix.svc.cluster.local:2379" + prefix: "/apisix" # apisix configurations prefix + timeout: 30 # seconds +plugins: # plugin list + - api-breaker + - authz-keycloak + - basic-auth + - batch-requests + - consumer-restriction + - cors + - echo + - fault-injection + - grpc-transcode + - hmac-auth + - http-logger + - ip-restriction + - jwt-auth + - kafka-logger + - key-auth + - limit-conn + - limit-count + - limit-req + - node-status + - openid-connect + - prometheus + - proxy-cache + - proxy-mirror + - proxy-rewrite + - redirect + - referer-restriction + - request-id + - request-validation + - response-rewrite + - serverless-post-function + - serverless-pre-function + - sls-logger + - syslog + - tcp-logger + - udp-logger + - uri-blocker + - wolf-rbac + - zipkin + - traffic-split +stream_plugins: + - mqtt-proxy +``` + +Please make sure `etcd.host` matches the headless service we created at first. In our case, it's `http://etcd-headless.apisix.svc.cluster.local:2379`. + +In this config, we defined an access key with the `admin` name under the `apisix.admin_key` section. This key is our API key, will be used to control APISIX later. Review comment: Add some security hints about changing this admin key in the real world. ########## File path: docs/en/latest/practices/the-hard-way.md ########## @@ -0,0 +1,826 @@ +--- +title: APISIX Ingress Controller the Hard Way +--- + +<!-- +# +# Licensed to the Apache Software Foundation (ASF) under one or more +# contributor license agreements. See the NOTICE file distributed with +# this work for additional information regarding copyright ownership. +# The ASF licenses this file to You under the Apache License, Version 2.0 +# (the "License"); you may not use this file except in compliance with +# the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# +--> + +In this tutorial, we will install APISIX and APISIX Ingress Controller in Kubernetes from native yaml. + +## Prerequisites + +If you don't have a Kubernetes cluster to use, we recommend you to use [KiND](https://kind.sigs.k8s.io/docs/user/quick-start/) to create a local Kubernetes cluster. + +```bash +kubectl create ns apisix +``` + +In this tutorial, all our operations will be performed at namespace `apisix`. + +## ETCD Installation + +Here, we will deploy a single-node ETCD cluster without authentication inside the Kubernetes cluster. + +In this case, we assume you have a storage provisioner. If you are using KiND, a local path provisioner will be created automatically. If you don't have a storage provisioner or don't want to use persistence volume, you could use an `emptyDir` as volume. + +```yaml +# etcd-headless.yaml +apiVersion: v1 +kind: Service +metadata: + name: etcd-headless + labels: + app.kubernetes.io/name: etcd + annotations: + service.alpha.kubernetes.io/tolerate-unready-endpoints: "true" +spec: + type: ClusterIP + clusterIP: None + ports: + - name: "client" + port: 2379 + targetPort: client + - name: "peer" + port: 2380 + targetPort: peer + selector: + app.kubernetes.io/name: etcd +--- +# etcd.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: etcd + labels: + app.kubernetes.io/name: etcd +spec: + selector: + matchLabels: + app.kubernetes.io/name: etcd + serviceName: etcd-headless + podManagementPolicy: Parallel + replicas: 1 + updateStrategy: + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: etcd + spec: + securityContext: + fsGroup: 1001 + runAsUser: 1001 + containers: + - name: etcd + image: docker.io/bitnami/etcd:3.4.14-debian-10-r0 + imagePullPolicy: "IfNotPresent" + # command: + # - /scripts/setup.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: MY_POD_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: MY_POD_NAME + valueFrom: + fieldRef: + fieldPath: metadata.name + - name: ETCDCTL_API + value: "3" + - name: ETCD_NAME + value: "$(MY_POD_NAME)" + - name: ETCD_DATA_DIR + value: /etcd/data + - name: ETCD_ADVERTISE_CLIENT_URLS + value: "http://$(MY_POD_NAME).etcd-headless.apisix.svc.cluster.local:2379" + - name: ETCD_LISTEN_CLIENT_URLS + value: "http://0.0.0.0:2379" + - name: ETCD_INITIAL_ADVERTISE_PEER_URLS + value: "http://$(MY_POD_NAME).etcd-headless.apisix.svc.cluster.local:2380" + - name: ETCD_LISTEN_PEER_URLS + value: "http://0.0.0.0:2380" + - name: ALLOW_NONE_AUTHENTICATION + value: "yes" + ports: + - name: client + containerPort: 2379 + - name: peer + containerPort: 2380 + volumeMounts: + - name: data + mountPath: /etcd + volumeClaimTemplates: + - metadata: + name: data + spec: + accessModes: + - "ReadWriteOnce" + resources: + requests: + storage: "8Gi" +``` + +Apply these two yaml files to Kubernetes, wait few seconds, etcd installation should be successful. We could run a health check to ensure that. + +```bash +$ kubectl -n apisix exec -it etcd-0 -- etcdctl endpoint health +127.0.0.1:2379 is healthy: successfully committed proposal: took = 1.741883ms +``` + +Please notice that this etcd installation is quite simple and lack of many necessary production features, it should only be used for learning case. If you want to deploy a production-ready etcd, please refer [bitnami/etcd](https://bitnami.com/stack/etcd/helm). + +## APISIX Installation + +Create a config file for our APISIX. We are going to deploy APISIX version 2.5. Review comment: We may add a hyperlink about APISIX 2.5 ########## File path: docs/en/latest/practices/the-hard-way.md ########## @@ -0,0 +1,826 @@ +--- +title: APISIX Ingress Controller the Hard Way +--- + +<!-- +# +# Licensed to the Apache Software Foundation (ASF) under one or more +# contributor license agreements. See the NOTICE file distributed with +# this work for additional information regarding copyright ownership. +# The ASF licenses this file to You under the Apache License, Version 2.0 +# (the "License"); you may not use this file except in compliance with +# the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# +--> + +In this tutorial, we will install APISIX and APISIX Ingress Controller in Kubernetes from native yaml. + +## Prerequisites + +If you don't have a Kubernetes cluster to use, we recommend you to use [KiND](https://kind.sigs.k8s.io/docs/user/quick-start/) to create a local Kubernetes cluster. + +```bash +kubectl create ns apisix +``` + +In this tutorial, all our operations will be performed at namespace `apisix`. + +## ETCD Installation + +Here, we will deploy a single-node ETCD cluster without authentication inside the Kubernetes cluster. + +In this case, we assume you have a storage provisioner. If you are using KiND, a local path provisioner will be created automatically. If you don't have a storage provisioner or don't want to use persistence volume, you could use an `emptyDir` as volume. + +```yaml +# etcd-headless.yaml +apiVersion: v1 +kind: Service +metadata: + name: etcd-headless + labels: + app.kubernetes.io/name: etcd + annotations: + service.alpha.kubernetes.io/tolerate-unready-endpoints: "true" +spec: + type: ClusterIP + clusterIP: None + ports: + - name: "client" + port: 2379 + targetPort: client + - name: "peer" + port: 2380 + targetPort: peer + selector: + app.kubernetes.io/name: etcd +--- +# etcd.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: etcd + labels: + app.kubernetes.io/name: etcd +spec: + selector: + matchLabels: + app.kubernetes.io/name: etcd + serviceName: etcd-headless + podManagementPolicy: Parallel + replicas: 1 + updateStrategy: + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: etcd + spec: + securityContext: + fsGroup: 1001 + runAsUser: 1001 + containers: + - name: etcd + image: docker.io/bitnami/etcd:3.4.14-debian-10-r0 + imagePullPolicy: "IfNotPresent" + # command: + # - /scripts/setup.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: MY_POD_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: MY_POD_NAME + valueFrom: + fieldRef: + fieldPath: metadata.name + - name: ETCDCTL_API + value: "3" + - name: ETCD_NAME + value: "$(MY_POD_NAME)" + - name: ETCD_DATA_DIR + value: /etcd/data + - name: ETCD_ADVERTISE_CLIENT_URLS + value: "http://$(MY_POD_NAME).etcd-headless.apisix.svc.cluster.local:2379" + - name: ETCD_LISTEN_CLIENT_URLS + value: "http://0.0.0.0:2379" + - name: ETCD_INITIAL_ADVERTISE_PEER_URLS + value: "http://$(MY_POD_NAME).etcd-headless.apisix.svc.cluster.local:2380" + - name: ETCD_LISTEN_PEER_URLS + value: "http://0.0.0.0:2380" + - name: ALLOW_NONE_AUTHENTICATION + value: "yes" + ports: + - name: client + containerPort: 2379 + - name: peer + containerPort: 2380 + volumeMounts: + - name: data + mountPath: /etcd + volumeClaimTemplates: + - metadata: + name: data + spec: + accessModes: + - "ReadWriteOnce" + resources: + requests: + storage: "8Gi" +``` + +Apply these two yaml files to Kubernetes, wait few seconds, etcd installation should be successful. We could run a health check to ensure that. + +```bash +$ kubectl -n apisix exec -it etcd-0 -- etcdctl endpoint health +127.0.0.1:2379 is healthy: successfully committed proposal: took = 1.741883ms +``` + +Please notice that this etcd installation is quite simple and lack of many necessary production features, it should only be used for learning case. If you want to deploy a production-ready etcd, please refer [bitnami/etcd](https://bitnami.com/stack/etcd/helm). + +## APISIX Installation + +Create a config file for our APISIX. We are going to deploy APISIX version 2.5. + +Note that the APISIX ingress controller needs to communicate with the APISIX admin API, so we set `apisix.allow_admin` to `0.0.0.0/0` for test. + +```yaml +apisix: + node_listen: 9080 # APISIX listening port + enable_heartbeat: true + enable_admin: true + enable_admin_cors: true + enable_debug: false + enable_dev_mode: false # Sets nginx worker_processes to 1 if set to true + enable_reuseport: true # Enable nginx SO_REUSEPORT switch if set to true. + enable_ipv6: true + config_center: etcd # etcd: use etcd to store the config value + + allow_admin: # http://nginx.org/en/docs/http/ngx_http_access_module.html#allow + - 0.0.0.0/0 + port_admin: 9180 + + # Default token when use API to call for Admin API. + # *NOTE*: Highly recommended to modify this value to protect APISIX's Admin API. + # Disabling this configuration item means that the Admin API does not + # require any authentication. + admin_key: + # admin: can everything for configuration data + - name: "admin" + key: edd1c9f034335f136f87ad84b625c8f1 + role: admin + # viewer: only can view configuration data + - name: "viewer" + key: 4054f7cf07e344346cd3f287985e76a2 + role: viewer + # dns_resolver: + # - 127.0.0.1 + dns_resolver_valid: 30 + resolver_timeout: 5 + +nginx_config: # config for render the template to genarate nginx.conf + error_log: "/dev/stderr" + error_log_level: "warn" # warn,error + worker_rlimit_nofile: 20480 # the number of files a worker process can open, should be larger than worker_connections + event: + worker_connections: 10620 + http: + access_log: "/dev/stdout" + keepalive_timeout: 60s # timeout during which a keep-alive client connection will stay open on the server side. + client_header_timeout: 60s # timeout for reading client request header, then 408 (Request Time-out) error is returned to the client + client_body_timeout: 60s # timeout for reading client request body, then 408 (Request Time-out) error is returned to the client + send_timeout: 10s # timeout for transmitting a response to the client.then the connection is closed + underscores_in_headers: "on" # default enables the use of underscores in client request header fields + real_ip_header: "X-Real-IP" # http://nginx.org/en/docs/http/ngx_http_realip_module.html#real_ip_header + real_ip_from: # http://nginx.org/en/docs/http/ngx_http_realip_module.html#set_real_ip_from + - 127.0.0.1 + - 'unix:' + +etcd: + host: + - "http://etcd-headless.apisix.svc.cluster.local:2379" + prefix: "/apisix" # apisix configurations prefix + timeout: 30 # seconds +plugins: # plugin list + - api-breaker + - authz-keycloak + - basic-auth + - batch-requests + - consumer-restriction + - cors + - echo + - fault-injection + - grpc-transcode + - hmac-auth + - http-logger + - ip-restriction + - jwt-auth + - kafka-logger + - key-auth + - limit-conn + - limit-count + - limit-req + - node-status + - openid-connect + - prometheus + - proxy-cache + - proxy-mirror + - proxy-rewrite + - redirect + - referer-restriction + - request-id + - request-validation + - response-rewrite + - serverless-post-function + - serverless-pre-function + - sls-logger + - syslog + - tcp-logger + - udp-logger + - uri-blocker + - wolf-rbac + - zipkin + - traffic-split +stream_plugins: + - mqtt-proxy +``` + +Please make sure `etcd.host` matches the headless service we created at first. In our case, it's `http://etcd-headless.apisix.svc.cluster.local:2379`. + +In this config, we defined an access key with the `admin` name under the `apisix.admin_key` section. This key is our API key, will be used to control APISIX later. + +Save this as `config.yaml`, then run `kubectl -n apisix create cm apisix-conf --from-file ./config.yaml` to create configmap. Later we will mount this configmap into APISIX deployment. + +```yaml +apiVersion: apps/v1 +kind: Deployment +metadata: + name: apisix + labels: + app.kubernetes.io/name: apisix +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: apisix + template: + metadata: + labels: + app.kubernetes.io/name: apisix + spec: + containers: + - name: apisix + image: "apache/apisix:2.5-alpine" + imagePullPolicy: IfNotPresent + ports: + - name: http + containerPort: 9080 + protocol: TCP + - name: tls + containerPort: 9443 + protocol: TCP + - name: admin + containerPort: 9180 + protocol: TCP + readinessProbe: + failureThreshold: 6 + initialDelaySeconds: 10 + periodSeconds: 10 + successThreshold: 1 + tcpSocket: + port: 9080 + timeoutSeconds: 1 + lifecycle: + preStop: + exec: + command: + - /bin/sh + - -c + - "sleep 30" + volumeMounts: + - mountPath: /usr/local/apisix/conf/config.yaml + name: apisix-config + subPath: config.yaml + resources: {} + volumes: + - configMap: + name: apisix-conf + name: apisix-config +``` + +Now, APISIX should be ready to use. Use `kubectl get pods -n apisix -l app.kubernetes.io/name=apisix -o name` to list APISIX pod name. Here we assume the pod name is `apisix-7644966c4d-cl4k6`. + +Let's have a check: + +```bash +kubectl -n apisix exec -it apisix-7644966c4d-cl4k6 -- curl http://127.0.0.1:9080 +``` + +If you are using Linux or macOS, run the command below in bash: + +```bash +kubectl -n apisix exec -it $(kubectl get pods -n apisix -l app.kubernetes.io/name=apisix -o name) -- curl http://127.0.0.1:9080 +``` + +If APISIX works properly, it should output: `{"error_msg":"404 Route Not Found"}`. Because we haven't defined any route yet. + +## HTTPBIN service + +Before configuring the APISIX, we need to create a test service. We use [kennethreitz/httpbin] here. We put this httpbin service in `demo` namespace. + +```bash +kubectl create ns demo +kubectl -n demo run httpbin --image-pull-policy=IfNotPresent --image kennethreitz/httpbin --port 80 +kubectl -n demo expose pod httpbin --port 80 +``` + +After the httpbin service started, we should be able to access it inside the APISIX pod via service. + +```bash +kubectl -n apisix exec -it apisix-7644966c4d-cl4k6 -- curl http://httpbin.demo/get +``` + +This should output the request's query parameters, for example: + +```json +{ + "args": {}, + "headers": { + "Accept": "*/*", + "Host": "httpbin.demo", + "User-Agent": "curl/7.67.0" + }, + "origin": "172.17.0.1", + "url": "http://httpbin.demo/get" +} +``` + +To read more, please refer to [Getting Started](https://apisix.apache.org/docs/apisix/getting-started). + +## Define Route + +Now, we can define the route for APISIX. Review comment: ```suggestion Now, we can define the route for proxying HTTPBIN through APISIX. ``` ########## File path: docs/en/latest/practices/the-hard-way.md ########## @@ -0,0 +1,826 @@ +--- +title: APISIX Ingress Controller the Hard Way +--- + +<!-- +# +# Licensed to the Apache Software Foundation (ASF) under one or more +# contributor license agreements. See the NOTICE file distributed with +# this work for additional information regarding copyright ownership. +# The ASF licenses this file to You under the Apache License, Version 2.0 +# (the "License"); you may not use this file except in compliance with +# the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# +--> + +In this tutorial, we will install APISIX and APISIX Ingress Controller in Kubernetes from native yaml. + +## Prerequisites + +If you don't have a Kubernetes cluster to use, we recommend you to use [KiND](https://kind.sigs.k8s.io/docs/user/quick-start/) to create a local Kubernetes cluster. + +```bash +kubectl create ns apisix +``` + +In this tutorial, all our operations will be performed at namespace `apisix`. + +## ETCD Installation + +Here, we will deploy a single-node ETCD cluster without authentication inside the Kubernetes cluster. + +In this case, we assume you have a storage provisioner. If you are using KiND, a local path provisioner will be created automatically. If you don't have a storage provisioner or don't want to use persistence volume, you could use an `emptyDir` as volume. + +```yaml +# etcd-headless.yaml +apiVersion: v1 +kind: Service +metadata: + name: etcd-headless + labels: + app.kubernetes.io/name: etcd + annotations: + service.alpha.kubernetes.io/tolerate-unready-endpoints: "true" +spec: + type: ClusterIP + clusterIP: None + ports: + - name: "client" + port: 2379 + targetPort: client + - name: "peer" + port: 2380 + targetPort: peer + selector: + app.kubernetes.io/name: etcd +--- +# etcd.yaml +apiVersion: apps/v1 +kind: StatefulSet +metadata: + name: etcd + labels: + app.kubernetes.io/name: etcd +spec: + selector: + matchLabels: + app.kubernetes.io/name: etcd + serviceName: etcd-headless + podManagementPolicy: Parallel + replicas: 1 + updateStrategy: + type: RollingUpdate + template: + metadata: + labels: + app.kubernetes.io/name: etcd + spec: + securityContext: + fsGroup: 1001 + runAsUser: 1001 + containers: + - name: etcd + image: docker.io/bitnami/etcd:3.4.14-debian-10-r0 + imagePullPolicy: "IfNotPresent" + # command: + # - /scripts/setup.sh + env: + - name: BITNAMI_DEBUG + value: "false" + - name: MY_POD_IP + valueFrom: + fieldRef: + fieldPath: status.podIP + - name: MY_POD_NAME + valueFrom: + fieldRef: + fieldPath: metadata.name + - name: ETCDCTL_API + value: "3" + - name: ETCD_NAME + value: "$(MY_POD_NAME)" + - name: ETCD_DATA_DIR + value: /etcd/data + - name: ETCD_ADVERTISE_CLIENT_URLS + value: "http://$(MY_POD_NAME).etcd-headless.apisix.svc.cluster.local:2379" + - name: ETCD_LISTEN_CLIENT_URLS + value: "http://0.0.0.0:2379" + - name: ETCD_INITIAL_ADVERTISE_PEER_URLS + value: "http://$(MY_POD_NAME).etcd-headless.apisix.svc.cluster.local:2380" + - name: ETCD_LISTEN_PEER_URLS + value: "http://0.0.0.0:2380" + - name: ALLOW_NONE_AUTHENTICATION + value: "yes" + ports: + - name: client + containerPort: 2379 + - name: peer + containerPort: 2380 + volumeMounts: + - name: data + mountPath: /etcd + volumeClaimTemplates: + - metadata: + name: data + spec: + accessModes: + - "ReadWriteOnce" + resources: + requests: + storage: "8Gi" +``` + +Apply these two yaml files to Kubernetes, wait few seconds, etcd installation should be successful. We could run a health check to ensure that. + +```bash +$ kubectl -n apisix exec -it etcd-0 -- etcdctl endpoint health +127.0.0.1:2379 is healthy: successfully committed proposal: took = 1.741883ms +``` + +Please notice that this etcd installation is quite simple and lack of many necessary production features, it should only be used for learning case. If you want to deploy a production-ready etcd, please refer [bitnami/etcd](https://bitnami.com/stack/etcd/helm). + +## APISIX Installation + +Create a config file for our APISIX. We are going to deploy APISIX version 2.5. + +Note that the APISIX ingress controller needs to communicate with the APISIX admin API, so we set `apisix.allow_admin` to `0.0.0.0/0` for test. + +```yaml +apisix: + node_listen: 9080 # APISIX listening port + enable_heartbeat: true + enable_admin: true + enable_admin_cors: true + enable_debug: false + enable_dev_mode: false # Sets nginx worker_processes to 1 if set to true + enable_reuseport: true # Enable nginx SO_REUSEPORT switch if set to true. + enable_ipv6: true + config_center: etcd # etcd: use etcd to store the config value + + allow_admin: # http://nginx.org/en/docs/http/ngx_http_access_module.html#allow + - 0.0.0.0/0 + port_admin: 9180 + + # Default token when use API to call for Admin API. + # *NOTE*: Highly recommended to modify this value to protect APISIX's Admin API. + # Disabling this configuration item means that the Admin API does not + # require any authentication. + admin_key: + # admin: can everything for configuration data + - name: "admin" + key: edd1c9f034335f136f87ad84b625c8f1 + role: admin + # viewer: only can view configuration data + - name: "viewer" + key: 4054f7cf07e344346cd3f287985e76a2 + role: viewer + # dns_resolver: + # - 127.0.0.1 + dns_resolver_valid: 30 + resolver_timeout: 5 + +nginx_config: # config for render the template to genarate nginx.conf + error_log: "/dev/stderr" + error_log_level: "warn" # warn,error + worker_rlimit_nofile: 20480 # the number of files a worker process can open, should be larger than worker_connections + event: + worker_connections: 10620 + http: + access_log: "/dev/stdout" + keepalive_timeout: 60s # timeout during which a keep-alive client connection will stay open on the server side. + client_header_timeout: 60s # timeout for reading client request header, then 408 (Request Time-out) error is returned to the client + client_body_timeout: 60s # timeout for reading client request body, then 408 (Request Time-out) error is returned to the client + send_timeout: 10s # timeout for transmitting a response to the client.then the connection is closed + underscores_in_headers: "on" # default enables the use of underscores in client request header fields + real_ip_header: "X-Real-IP" # http://nginx.org/en/docs/http/ngx_http_realip_module.html#real_ip_header + real_ip_from: # http://nginx.org/en/docs/http/ngx_http_realip_module.html#set_real_ip_from + - 127.0.0.1 + - 'unix:' + +etcd: + host: + - "http://etcd-headless.apisix.svc.cluster.local:2379" + prefix: "/apisix" # apisix configurations prefix + timeout: 30 # seconds +plugins: # plugin list + - api-breaker + - authz-keycloak + - basic-auth + - batch-requests + - consumer-restriction + - cors + - echo + - fault-injection + - grpc-transcode + - hmac-auth + - http-logger + - ip-restriction + - jwt-auth + - kafka-logger + - key-auth + - limit-conn + - limit-count + - limit-req + - node-status + - openid-connect + - prometheus + - proxy-cache + - proxy-mirror + - proxy-rewrite + - redirect + - referer-restriction + - request-id + - request-validation + - response-rewrite + - serverless-post-function + - serverless-pre-function + - sls-logger + - syslog + - tcp-logger + - udp-logger + - uri-blocker + - wolf-rbac + - zipkin + - traffic-split +stream_plugins: + - mqtt-proxy +``` + +Please make sure `etcd.host` matches the headless service we created at first. In our case, it's `http://etcd-headless.apisix.svc.cluster.local:2379`. + +In this config, we defined an access key with the `admin` name under the `apisix.admin_key` section. This key is our API key, will be used to control APISIX later. + +Save this as `config.yaml`, then run `kubectl -n apisix create cm apisix-conf --from-file ./config.yaml` to create configmap. Later we will mount this configmap into APISIX deployment. + +```yaml +apiVersion: apps/v1 +kind: Deployment +metadata: + name: apisix + labels: + app.kubernetes.io/name: apisix +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: apisix + template: + metadata: + labels: + app.kubernetes.io/name: apisix + spec: + containers: + - name: apisix + image: "apache/apisix:2.5-alpine" + imagePullPolicy: IfNotPresent + ports: + - name: http + containerPort: 9080 + protocol: TCP + - name: tls + containerPort: 9443 + protocol: TCP + - name: admin + containerPort: 9180 + protocol: TCP + readinessProbe: + failureThreshold: 6 + initialDelaySeconds: 10 + periodSeconds: 10 + successThreshold: 1 + tcpSocket: + port: 9080 + timeoutSeconds: 1 + lifecycle: + preStop: + exec: + command: + - /bin/sh + - -c + - "sleep 30" + volumeMounts: + - mountPath: /usr/local/apisix/conf/config.yaml + name: apisix-config + subPath: config.yaml + resources: {} + volumes: + - configMap: + name: apisix-conf + name: apisix-config +``` + +Now, APISIX should be ready to use. Use `kubectl get pods -n apisix -l app.kubernetes.io/name=apisix -o name` to list APISIX pod name. Here we assume the pod name is `apisix-7644966c4d-cl4k6`. + +Let's have a check: + +```bash +kubectl -n apisix exec -it apisix-7644966c4d-cl4k6 -- curl http://127.0.0.1:9080 +``` + +If you are using Linux or macOS, run the command below in bash: + +```bash +kubectl -n apisix exec -it $(kubectl get pods -n apisix -l app.kubernetes.io/name=apisix -o name) -- curl http://127.0.0.1:9080 +``` + +If APISIX works properly, it should output: `{"error_msg":"404 Route Not Found"}`. Because we haven't defined any route yet. + +## HTTPBIN service + +Before configuring the APISIX, we need to create a test service. We use [kennethreitz/httpbin] here. We put this httpbin service in `demo` namespace. + +```bash +kubectl create ns demo +kubectl -n demo run httpbin --image-pull-policy=IfNotPresent --image kennethreitz/httpbin --port 80 +kubectl -n demo expose pod httpbin --port 80 +``` + +After the httpbin service started, we should be able to access it inside the APISIX pod via service. + +```bash +kubectl -n apisix exec -it apisix-7644966c4d-cl4k6 -- curl http://httpbin.demo/get +``` + +This should output the request's query parameters, for example: + +```json +{ + "args": {}, + "headers": { + "Accept": "*/*", + "Host": "httpbin.demo", + "User-Agent": "curl/7.67.0" + }, + "origin": "172.17.0.1", + "url": "http://httpbin.demo/get" +} +``` + +To read more, please refer to [Getting Started](https://apisix.apache.org/docs/apisix/getting-started). + +## Define Route + +Now, we can define the route for APISIX. + +Assuming we want to route all traffic which URI has `/httpbin` prefix and the request contains `Host: httpbin.org` header. + +Please notice that the admin port is `9180`. + +```bash +kubectl -n apisix exec -it apisix-7644966c4d-cl4k6 -- curl "http://127.0.0.1:9180/apisix/admin/routes/1" -H "X-API-KEY: edd1c9f034335f136f87ad84b625c8f1" -X PUT -d ' +{ + "uri": "/*", + "host": "httpbin.org", + "upstream": { + "type": "roundrobin", + "nodes": { + "httpbin.demo:80": 1 + } + } +}' +``` + +The output would be like this: + +```json +{"action":"set","node":{"key":"\/apisix\/routes\/1","value":{"status":1,"create_time":1621408897,"upstream":{"pass_host":"pass","type":"roundrobin","hash_on":"vars","nodes":{"httpbin.demo:80":1},"scheme":"http"},"update_time":1621408897,"priority":0,"host":"httpbin.org","id":"1","uri":"\/*"}}} +``` + +We could check route rules by `GET /apisix/admin/routes`: + +```bash +kubectl -n apisix exec -it apisix-7644966c4d-cl4k6 -- curl "http://127.0.0.1:9180/apisix/admin/routes/1" -H "X-API-KEY: edd1c9f034335f136f87ad84b625c8f1" +``` + +It should output like this: + +```json +{"action":"get","node":{"key":"\/apisix\/routes\/1","value":{"upstream":{"pass_host":"pass","type":"roundrobin","scheme":"http","hash_on":"vars","nodes":{"httpbin.demo:80":1}},"id":"1","create_time":1621408897,"update_time":1621408897,"host":"httpbin.org","priority":0,"status":1,"uri":"\/*"}},"count":"1"} +``` + +Now, we can test the routing rule: + +```bash +kubectl -n apisix exec -it apisix-7644966c4d-cl4k6 -- curl "http://127.0.0.1:9080/get" -H 'Host: httpbin.org' +``` + +It will output like: + +```json +{ + "args": {}, + "headers": { + "Accept": "*/*", + "Host": "httpbin.org", + "User-Agent": "curl/7.67.0", + "X-Forwarded-Host": "httpbin.org" + }, + "origin": "127.0.0.1", + "url": "http://httpbin.org/get" +} +``` + +## Install APISIX Ingress Controller + +APISIX ingress controller can help you manage your configurations declaratively by using Kubernetes resources. Here we will install version 0.5.0. + +Currently, the APISIX ingress controller supports both official Ingress resource or APISIX's CustomResourceDefinitions, which includes ApisixRoute and ApisixUpstream. + +Before installing the APISIX controller, we need to create a service account and the corresponding ClusterRole to ensure that the APISIX ingress controller has sufficient permissions to access required resources. + +Here is an example config from [apisix-helm-chart](https://github.com/apache/apisix-helm-chart): + +```yaml +apiVersion: v1 +kind: ServiceAccount +metadata: + name: apisix-ingress-controller +--- +apiVersion: rbac.authorization.k8s.io/v1 +kind: ClusterRole +metadata: + name: apisix-clusterrole +rules: + - apiGroups: + - "" + resources: + - configmaps + - endpoints + - persistentvolumeclaims + - pods + - replicationcontrollers + - replicationcontrollers/scale + - serviceaccounts + - services + - secrets + verbs: + - get + - list + - watch + - apiGroups: + - "" + resources: + - bindings + - events + - limitranges + - namespaces/status + - pods/log + - pods/status + - replicationcontrollers/status + - resourcequotas + - resourcequotas/status + verbs: + - get + - list + - watch + - apiGroups: + - "" + resources: + - namespaces + verbs: + - get + - list + - watch + - apiGroups: + - apps + resources: + - controllerrevisions + - daemonsets + - deployments + - deployments/scale + - replicasets + - replicasets/scale + - statefulsets + - statefulsets/scale + verbs: + - get + - list + - watch + - apiGroups: + - autoscaling + resources: + - horizontalpodautoscalers + verbs: + - get + - list + - watch + - apiGroups: + - batch + resources: + - cronjobs + - jobs + verbs: + - get + - list + - watch + - apiGroups: + - extensions + resources: + - daemonsets + - deployments + - deployments/scale + - ingresses + - networkpolicies + - replicasets + - replicasets/scale + - replicationcontrollers/scale + verbs: + - get + - list + - watch + - apiGroups: + - policy + resources: + - poddisruptionbudgets + verbs: + - get + - list + - watch + - apiGroups: + - networking.k8s.io + resources: + - ingresses + - networkpolicies + verbs: + - get + - list + - watch + - apiGroups: + - metrics.k8s.io + resources: + - pods + verbs: + - get + - list + - watch + - apiGroups: + - apisix.apache.org + resources: + - apisixroutes + - apisixupstreams + - apisixservices + - apisixtlses + verbs: + - get + - list + - watch + - apiGroups: + - coordination.k8s.io + resources: + - leases + verbs: + - '*' +--- +apiVersion: rbac.authorization.k8s.io/v1 +kind: ClusterRoleBinding +metadata: + name: apisix-clusterrolebinding +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: ClusterRole + name: apisix-clusterrole +subjects: + - kind: ServiceAccount + name: apisix-ingress-controller + namespace: apisix +``` + +Then, we need to create ApisixRoute CRD: + +```yaml + +apiVersion: apiextensions.k8s.io/v1beta1 +kind: CustomResourceDefinition +metadata: + name: apisixroutes.apisix.apache.org +spec: + group: apisix.apache.org + versions: + - name: v1 + served: true + storage: false + - name: v2alpha1 + served: true + storage: true + scope: Namespaced + names: + plural: apisixroutes + singular: apisixroute + kind: ApisixRoute + shortNames: + - ar +--- +apiVersion: apiextensions.k8s.io/v1beta1 +kind: CustomResourceDefinition +metadata: + name: apisixtlses.apisix.apache.org +spec: + group: apisix.apache.org + versions: + - name: v1 + served: true + storage: true + scope: Namespaced + names: + plural: apisixtlses + singular: apisixtls + kind: ApisixTls + shortNames: + - atls +--- +apiVersion: apiextensions.k8s.io/v1beta1 +kind: CustomResourceDefinition +metadata: + name: apisixupstreams.apisix.apache.org +spec: + group: apisix.apache.org + versions: + - name: v1 + served: true + storage: true + scope: Namespaced + names: + plural: apisixupstreams + singular: apisixupstream + kind: ApisixUpstream + shortNames: + - au +``` + +To make the ingress controller works properly with APISIX, we need to create a config file containing the APISIX admin API URL and API key as below: + +```yaml +apiVersion: v1 +data: + config.yaml: | + # log options + log_level: "debug" + log_output: "stderr" + http_listen: ":8080" + enable_profiling: true + kubernetes: + kubeconfig: "" + resync_interval: "30s" + app_namespaces: + - "*" + ingress_class: "apisix" + ingress_version: "networking/v1" + apisix_route_version: "apisix.apache.org/v1" + apisix: + base_url: "http://apisix-admin.apisix:9180/apisix/admin" + admin_key: "edd1c9f034335f136f87ad84b625c8f1" +kind: ConfigMap +metadata: + name: apisix-configmap + labels: + app.kubernetes.io/name: ingress-controller +``` Review comment: Add a note that if you want to learn all the configuration items, see [conf/config-default.yaml](./conf/config-default.yaml) for details. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [email protected]
