Skip to content

Improving HA configuration for Knative workloads

Published on: 2023-06-13 ,  Revised on: 2024-01-17

Improving HA configuration for Knative workloads

Author: Matthias Weßendorf, Senior Principal Software Engineer @ Red Hat

In this blog post you will learn how to use the Knative Operator to maintain a fine-grain configuration for high availablitly of Knative workloads.

The Knative Operator gives you a declarative API to describe your Knative Serving and Eventing installation. The spec field has several properties to define the desired behavior.

A convenient global configuration for high availability

Take a look at the following installation manifest:

apiVersion: operator.knative.dev/v1beta1
kind: KnativeEventing
metadata:
  name: knative-eventing
  namespace: knative-eventing
spec:
  high-availability:
    replicas: 3

This configures Knative Eventing in the knative-eventing namespace, and defines that all workloads, managed by the Operator, do require a replica set of 3 pods. Let's have a look:

kubectl get deployments.apps -n knative-eventing 
NAME                    READY   UP-TO-DATE   AVAILABLE   AGE
eventing-controller     3/3     3            3           58s
eventing-webhook        3/3     3            3           57s
imc-controller          3/3     3            3           53s
imc-dispatcher          3/3     3            3           52s
mt-broker-controller    3/3     3            3           51s
mt-broker-filter        3/3     3            3           51s
mt-broker-ingress       3/3     3            3           51s

For each workload we do see exactly three deployements. Now, take a detailed look at the above shell snippet. You will notice that for the InMemoryChannel we do have 6 deployments: 3 for each, the controller and the dispatcher data-plane. This is not always what you want, since the InMemoryChannel is more often used as a tool during development, while in production scenarios other worksloads, like the Knative Broker or Knative Channel implementations for Apache Kafka are being used.

Fine tuning the HA configuration with workload overrides

Now here is where the workloads fields comes into place. The workloads field allows administrator to perform a more fine grain tuning of each workload, managed by the Knative Operator. Details on the configuration options can be found in the documentation.

Take a look at the modified manifest:

apiVersion: operator.knative.dev/v1beta1
kind: KnativeEventing
metadata:
  name: knative-eventing
  namespace: knative-eventing
spec:
  high-availability:
    replicas: 3
  workloads:
  - name: imc-controller
    replicas: 1  
  - name: imc-dispatcher
    replicas: 1  

For the imc-controller and imc-dispatcher we have now done an override of the global default, and have reduced it to exactly one deployment for each:

kubectl get deployments.apps -n knative-eventing
NAME                    READY   UP-TO-DATE   AVAILABLE   AGE
eventing-controller     3/3     3            3           9m31s
eventing-webhook        3/3     3            3           9m30s
imc-controller          1/1     1            1           9m26s
imc-dispatcher          1/1     1            1           9m25s
mt-broker-controller    3/3     3            3           9m24s
mt-broker-filter        3/3     3            3           9m24s
mt-broker-ingress       3/3     3            3           9m24s

We could even set the replica to 0, if the system would not use any InMemoryChannel in this example.

Conclusion

While the high-availability offers a nice and easy to use API for defining high availability configuration for the workloads of Knative, the workloads is a nice option to adjust and optimize the installation. While the workloads is a little more complex to configure, since it allows a per workload override for production systems it is recommended to use this to adjust each workload on its exacts needs, rather than relying on global defaults, which may look promising only in the beginning.

We use analytics and cookies to understand site traffic. Information about your use of our site is shared with Google for that purpose. Learn more.

× OK