This guide assumes you have a basic understanding of the Prometheus resource and have read the getting started.

Besides the Prometheus and ServiceMonitor resource the Prometheus Operator also introduces the Alertmanager. It allows declaratively describing an Alertmanager cluster. Before diving into deploying an Alertmanager cluster, it is important to understand the contract between Prometheus and Alertmanager.

The Alertmanager's features include:

  • Deduplicating alerts fired by Prometheus
  • Silencing alerts
  • Route and send grouped notifications via providers (PagerDuty, OpsGenie, …)

Prometheus' configuration includes so called rule files, which contain the alerting rules. When an alerting rule triggers it fires that alert against all Alertmanager instances, on every rule evaluation interval. The Alertmanager instances communicate to each other which notifications have already been sent out. You can read more about why these systems have been designed this way in the High Availability scheme description.

Let's create an example Alertmanager cluster, with three instances.

kind: Alertmanager
  name: example
  replicas: 3

The Alertmanager instances will not be able to start up, unless a valid configuration is given. This is an example configuration, that does not actually do anything as it sends notifications against a non existent webhook, but will allow the Alertmanager to start up. Read more about how to configure the Alertmanager on the upstream documentation.

  resolve_timeout: 5m
  group_by: ['job']
  group_wait: 30s
  group_interval: 5m
  repeat_interval: 12h
  receiver: 'webhook'
- name: 'webhook'
  - url: 'http://alertmanagerwh:30500/'

Save the above alertmanager config in a file called alertmanager.yaml and create a secret from it using kubectl.

The Alertmanager instance needs an special naming on the secret resource in order to be able to grab it. It looks for alertmanager-{ALERTMANAGER_NAME} pattern, in this example the name of the Alertmanager is example so the secret name should be alertmanager-example and the name of the config file alertmanager.yaml

$ kubectl create secret generic alertmanager-example --from-file=alertmanager.yaml

Note that altermanagers configurations can use templates(.tmpl files), these can be added on the secret along with the alertmanager.yaml config file, for example:

apiVersion: v1
kind: Secret
  name: alertmanager-example
  alertmanager.yaml: {BASE64_CONFIG}
  template_1.tmpl: {BASE64_TEMPLATE_1}
  template_2.tmpl: {BASE64_TEMPLATE_2}

Templates will be placed on the same path as the configuration, in order to be able to load the templates, the configuration (alertmanager.yaml) should point to them:

- '*.tmpl'

Once created this Secret is mounted by Alertmanager Pods created through the Alertmanager object.

To be able to view the web UI, expose it via a Service. A simple way to do this is to use a Service of type NodePort.

apiVersion: v1
kind: Service
  name: alertmanager-example
  type: NodePort
  - name: web
    nodePort: 30903
    port: 9093
    protocol: TCP
    targetPort: web
    alertmanager: example

Once created it allows the web UI to be accessible via a node's IP and the port 30903.

Now this is a fully functional highly available Alertmanager cluster, but it does not get any alerts fired against it. Let's setup Prometheus instances that will actually fire alerts to our alertmanagers.

kind: Prometheus
  name: example
  replicas: 2
    - namespace: default
      name: alertmanager-example
      port: web
      team: frontend
      memory: 400Mi
      role: prometheus-rulefiles
      prometheus: example

The above configuration specifies a Prometheus that finds all of the alertmanagers behind the Service we just created. The name and port fields under alertmanagers, should match those of our Service to allow this to occur.

Prometheus rule files are held in ConfigMaps. The ConfigMaps to mount rule files from are selected with a label selector field called ruleSelector in the Prometheus object, as seen above. All top level files that end with the .rules extension will be loaded.

The best practice is to label the ConfigMaps containing rule files with role: prometheus-rulefiles as well as the name of the Prometheus object, prometheus: example in this case.

kind: ConfigMap
apiVersion: v1
  name: prometheus-example-rules
    role: prometheus-rulefiles
    prometheus: example
  example.rules.yaml: |+
    - name: ./example.rules
      - alert: ExampleAlert
        expr: vector(1)

That example ConfigMap always immediately triggers an alert, which is only for demonstration purposes. To validate that everything is working properly have a look at each of the Prometheus web UIs.

To be able to view the web UI without a Service, kubectl's proxy functionality can be used.


kubectl proxy --port=8001

Then the web UI of each Prometheus instance can be viewed, they both have a firing alert called ExampleAlert, as defined in the loaded alerting rules.

  • http://localhost:8001/api/v1/proxy/namespaces/default/pods/prometheus-example-0:9090/alerts
  • http://localhost:8001/api/v1/proxy/namespaces/default/pods/prometheus-example-1:9090/alerts

Looking at the status page for "Runtime & Build Information" on the Prometheus web UI shows the discovered and active Alertmanagers that the Prometheus instance will fire alerts against.

  • http://localhost:8001/api/v1/proxy/namespaces/default/pods/prometheus-example-0:9090/status
  • http://localhost:8001/api/v1/proxy/namespaces/default/pods/prometheus-example-1:9090/status

These show three discovered Alertmanagers.

Heading to the Alertmanager web UI now shows one active alert, although all Prometheus instances are firing it. Configuring the Alertmanager further allows custom alert routing, grouping and notification mechanisms.