diff --git a/Makefile b/Makefile index 80bd46d938..9e23b0d921 100644 --- a/Makefile +++ b/Makefile @@ -49,7 +49,7 @@ TELEKUBE_APP_TAG := $(GRAVITY_TAG) WORMHOLE_APP_TAG := $(GRAVITY_TAG) LOGGING_APP_TAG ?= 6.0.7 MONITORING_APP_TAG ?= 6.0.17 -DNS_APP_TAG = 6.1.2 +DNS_APP_TAG = 6.1.3 BANDWAGON_TAG ?= 6.0.1 RBAC_APP_TAG := $(GRAVITY_TAG) # IMPORTANT: When updating tiller version, DO NOT FORGET to bump TILLER_APP_TAG as well! diff --git a/assets/dns-app/resources/dns.yaml b/assets/dns-app/resources/dns.yaml index 597e6040e2..3a2d30b4d7 100644 --- a/assets/dns-app/resources/dns.yaml +++ b/assets/dns-app/resources/dns.yaml @@ -199,7 +199,7 @@ spec: matchExpressions: - key: k8s-app operator: In - values: ["kube-dns"] + values: ["kube-dns-worker"] topologyKey: kubernetes.io/hostname containers: - name: coredns @@ -421,7 +421,7 @@ data: # If using small nodes, "nodesPerReplica" should dominate. # Parameters based on kubernetes defaults: https://github.com/kubernetes/kubernetes/blob/f65f868aa041e108a73a014df360427c6e05e493/cluster/addons/dns-horizontal-autoscaler/dns-horizontal-autoscaler.yaml # linear: '{"coresPerReplica":256,"includeUnschedulableNodes":true,"nodesPerReplica":16,"preventSinglePointFailure":true}' - # Use ladder scaling that emulates the linear scale, because we want to scale down to 0 replicas when no workers are + # Use ladder scaling that emulates the linear scale, because we want to scale down to 0 replicas when no workers are # joined to the cluster. The linear scaler doesn't support 0 replicas. ladder: |- {