With v2.9 the Kubermatic chart wont contain any CustomResourceDefinitions. Upgrading the existing kubermatic installation with new charts would result in all CRD’s being deleted.
For this purpose we wrote a migration script. The script will:
The script is located inside the Kubermatic helm chart & must be executed before the chart upgrade: https://github.com/kubermatic/kubermatic-installer/blob/release/v2.9/charts/kubermatic/migrate/migrate-kubermatic-chart.sh
Afterwards, the CRDs must be installed with kubectl
apply -f charts/kubermatic/crd/.
Ark 0.10 requires significant changes to the chart configuation in your
values.yaml. Please consult the
the chart to learn more about the new configuration structure. For exsiting backups, Heptio provides a script to migrate them.
Consult the official upgrade guide for more information.
When updating an existing cert-manager, make sure to manually label the
kubectl label namespace cert-manager certmanager.k8s.io/disable-validation=true
Without this label, the new cert-manager will not be able to create its own PKI and the pods will crashloop because
webhook-ca secrets cannot be found.
The new version 0.17 has made significant changes to the metric names it provides. Consult the official guidelines to learn more about the new names and adjust your recording and alerting rules as needed. Note that Kubermatic does not install the recording rules to keep the old metric names intact, so you will notice gaps in the Grafana charts after you updated.
Kubermatic 2.9 replaces the old fluentd chart with fluentbit in order to improve performance and reduce resource usage of the
logging stack. When updating an existing installation, make sure to
delete --purge the fluentd chart, so you do not end up
with two log shippers in your cluster.
helm --tiller-namespace kubermatic delete --purge fluentd
Until Kubermatic v2.9 all OpenStack nodes got assigned a floating IP. With v2.9 this behaviour changes, as floating IP’s are now optional by default. Within the “Add Node” dialog, the user can specific if a floating IP should be assigned or not.
If the assignment of floating IP’s is a requirement to ensure Node-> API server communication, the assignment can be enforced within the datacenters.yaml:
loodse-hamburg-1: location: Hamburg seed: europe-west3-c country: DE provider: openstack spec: openstack: auth_url: https://some-keystone:5000/v3 availability_zone: hamburg-1 region: hamburg dns_servers: - "220.127.116.11" - "18.104.22.168" images: ubuntu: "Ubuntu 18.04 LTS - 2018-08-10" centos: "" coreos: "" # Enforce the assignment for floating IP's for nodes of this datacenter enforce_floating_ip: true
Disabled by default. Can be enabled by setting the feature flag:
#Feature flag kubermatic.controller.featureGates="VerticalPodAutoscaler=true"
This will instruct the kubermatic cluster controller to deploy VerticalPodAutoscaler resources for all control plane components.
The VerticalPodAutoscaler will then make sure that those pod will receive resource requests according to the actual usage.
Getting the VPA resources for a cluster: For example:
kubectl -n cluster-xxxxxx get vpa
If the VerticalPodAutoscaler notices a difference by 20% between the current usage and the specified resource request, the pod will be deleted, so it gets recreated by the controller(ReplicaSet, StatefulSet). More details on the VerticalPodAutoscaler can be found in the official repository: https://github.com/kubernetes/autoscaler/tree/master/vertical-pod-autoscaler#vertical-pod-autoscaler
API server downtime In case the VPA needs to scale up the API server and the cluster only has 1 replica, the API server won’t be available for a short timeframe.
New pods cannot be scheduled In case the VPA deletes a Pod, the new Pod might be rescheduled in case the cluster has not enough resources available for fulfil the pods resource request.