2018-09-12 15:44:30 +00:00
---
layout: "docs"
page_title: "Running Consul - Kubernetes"
sidebar_current: "docs-platform-k8s-run"
description: |-
Consul can run directly on Kubernetes, both in server or client mode. For pure-Kubernetes workloads, this enables Consul to also exist purely within Kubernetes. For heterogeneous workloads, Consul agents can join a server running inside or outside of Kubernetes.
---
# Running Consul on Kubernetes
Consul can run directly on Kubernetes, both in server or client mode.
For pure-Kubernetes workloads, this enables Consul to also exist purely
within Kubernetes. For heterogeneous workloads, Consul agents can join
a server running inside or outside of Kubernetes.
This page starts with a large how-to section for various specific tasks.
To learn more about the general architecture of Consul on Kubernetes, scroll
2019-10-14 15:40:35 +00:00
down to the [architecture ](/docs/platform/k8s/run.html#architecture ) section. If you would like to get hands-on experience testing Consul on Kubernetes, try the step-by-step beginner tutorial with an accompanying video in the [Minikube with Consul guide ](https://learn.hashicorp.com/consul/getting-started-k8s/minikube?utm_source=consul.io&utm_medium=docs )
2018-09-12 15:44:30 +00:00
## Helm Chart
The recommended way to run Consul on Kubernetes is via the
2018-09-13 21:45:40 +00:00
[Helm chart ](/docs/platform/k8s/helm.html ). This will install and configure
2018-09-12 15:44:30 +00:00
all the necessary components to run Consul. The configuration enables you
to run just a server cluster, just a client cluster, or both. Using the Helm
2018-09-13 21:45:40 +00:00
chart, you can have a full Consul deployment up and running in seconds.
2018-09-12 15:44:30 +00:00
2018-09-13 21:45:40 +00:00
While the Helm chart exposes dozens of useful configurations and automatically
2018-09-12 15:44:30 +00:00
sets up complex resources, it **does not automatically operate Consul.**
2019-01-28 22:25:56 +00:00
You are still responsible for learning how to monitor, backup,
2018-09-12 15:44:30 +00:00
upgrade, etc. the Consul cluster.
2018-09-13 21:45:40 +00:00
The Helm chart has no required configuration and will install a Consul
2018-09-12 15:44:30 +00:00
cluster with sane defaults out of the box. Prior to going to production,
it is highly recommended that you
[learn about the configuration options ](/docs/platform/k8s/helm.html#configuration-values- ).
2019-01-08 01:46:44 +00:00
~> **Security Warning:** By default, the chart will install an insecure configuration
of Consul. This provides a less complicated out-of-box experience for new users,
but is not appropriate for a production setup. It is highly recommended to use
a properly secured Kubernetes cluster or make sure that you understand and enable
2019-01-16 22:42:12 +00:00
the [recommended security features ](/docs/internals/security.html ). Currently,
2019-01-08 01:46:44 +00:00
some of these features are not supported in the Helm chart and require additional
manual configuration.
2018-09-12 15:44:30 +00:00
## How-To
### Installing Consul
To install Consul, clone the consul-helm repository, checkout the latest release, and install
Consul. You can run `helm install` with the `--dry-run` flag to see the
resources it would configure. In a production environment, you should always
use the `--dry-run` flag prior to making any changes to the Consul cluster
via Helm.
```sh
# Clone the chart repo
$ git clone https://github.com/hashicorp/consul-helm.git
$ cd consul-helm
# Checkout a tagged version
$ git checkout v0.1.0
# Run Helm
$ helm install --name consul ./
...
```
_That's it._ The Helm chart does everything to setup a recommended
Consul-on-Kubernetes deployment.
In a couple minutes, a Consul cluster will be formed and a leader
elected and every node will have a running Consul agent.
The defaults will install both server and client agents. To install
only one or the other, see the
[chart configuration values ](/docs/platform/k8s/helm.html#configuration-values- ).
### Viewing the Consul UI
The Consul UI is enabled by default when using the Helm chart.
For security reasons, it isn't exposed via a Service by default so you must
use `kubectl port-forward` to visit the UI. Once the port is forwarded as
shown below, navigate your browser to `http://localhost:8500` .
```
$ kubectl port-forward consul-server-0 8500:8500
...
```
The UI can also be exposed via a Kubernetes Service. To do this, configure
the [`ui.service` chart values ](/docs/platform/k8s/helm.html#v-ui-service ).
### Joining an Existing Consul Cluster
If you have a Consul cluster already running, you can configure your
Kubernetes nodes to join this existing cluster.
```yaml
global:
enabled: false
client:
enabled: true
join:
2018-11-12 20:22:18 +00:00
- "provider=my-cloud config=val ..."
2018-09-12 15:44:30 +00:00
```
The `values.yaml` file to configure the Helm chart sets the proper
configuration to join an existing cluster.
The `global.enabled` value first disables all chart components by default
so that each component is opt-in. This allows us to _only_ setup the client
agents. We then opt-in to the client agents by setting `client.enabled` to
`true` .
Next, `client.join` is set to an array of valid
[`-retry-join` values ](/docs/agent/options.html#retry-join ). In the
example above, a fake [cloud auto-join ](/docs/agent/cloud-auto-join.html )
value is specified. This should be set to resolve to the proper addresses of
your existing Consul cluster.
-> **Networking:** Note that for the Kubernetes nodes to join an existing
cluster, the nodes (and specifically the agent pods) must be able to connect
to all other server and client agents inside and _outside_ of Kubernetes.
If this isn't possible, consider running the Kubernetes agents as a separate
DC or adopting Enterprise for
[network segments ](/docs/enterprise/network-segments/index.html ).
### Accessing the Consul HTTP API
The Consul HTTP API should be accessed by communicating to the local agent
running on the same node. While technically any listening agent (client or
server) can respond to the HTTP API, communicating with the local agent
has important caching behavior, and allows you to use the simpler
[`/agent` endpoints for services and checks ](/api/agent.html ).
For Consul installed via the Helm chart, a client agent is installed on
each Kubernetes node. This is explained in the [architecture ](/docs/platform/k8s/run.html#client-agents )
section. To access the agent, you may use the
[downward API ](https://kubernetes.io/docs/tasks/inject-data-application/downward-api-volume-expose-pod-information/ ).
2018-10-01 17:35:09 +00:00
An example pod specification is shown below. In addition to pods, anything
with a pod template can also access the downward API and can therefore also
access Consul: StatefulSets, Deployments, Jobs, etc.
2018-09-12 15:44:30 +00:00
```yaml
apiVersion: v1
kind: Pod
metadata:
name: consul-example
spec:
containers:
- name: example
image: "consul:latest"
env:
- name: HOST_IP
valueFrom:
fieldRef:
fieldPath: status.hostIP
command:
- "/bin/sh"
- "-ec"
- |
export CONSUL_HTTP_ADDR="${HOST_IP}:8500"
consul kv put hello world
restartPolicy: Never
```
2018-10-01 17:35:09 +00:00
An example `Deployment` is also shown below to show how the host IP can
be accessed from nested pod specifications:
```yaml
apiVersion: apps/v1
kind: Deployment
metadata:
name: consul-example-deployment
spec:
replicas: 1
selector:
matchLabels:
app: consul-example
template:
metadata:
labels:
app: consul-example
spec:
containers:
- name: example
image: "consul:latest"
env:
- name: HOST_IP
valueFrom:
fieldRef:
fieldPath: status.hostIP
command:
- "/bin/sh"
- "-ec"
- |
export CONSUL_HTTP_ADDR="${HOST_IP}:8500"
consul kv put hello world
```
2018-09-12 15:44:30 +00:00
### Upgrading Consul on Kubernetes
To upgrade Consul on Kubernetes, we follow the same pattern as
[generally upgrading Consul ](/docs/upgrading.html ), except we can use
the Helm chart to step through a rolling deploy. It is important to understand
how to [generally upgrade Consul ](/docs/upgrading.html ) before reading this
section.
Upgrading Consul on Kubernetes will follow the same pattern: each server
will be updated one-by-one. After that is successful, the clients will
be updated in batches.
#### Upgrading Consul Servers
To initiate the upgrade, change the `server.image` value to the
desired Consul version. For illustrative purposes, the example below will
use `consul:123.456` . Also set the `server.updatePartition` value
_equal to the number of server replicas_:
```yaml
server:
image: "consul:123.456"
replicas: 3
updatePartition: 3
```
The `updatePartition` value controls how many instances of the server
cluster are updated. Only instances with an index _greater than_ the
`updatePartition` value are updated (zero-indexed). Therefore, by setting
it equal to replicas, none should update yet.
Next, run the upgrade. You should run this with `--dry-run` first to verify
the changes that will be sent to the Kubernetes cluster.
```
$ helm upgrade consul ./
...
```
This should cause no changes (although the resource will be updated). If
everything is stable, begin by decreasing the `updatePartition` value by one,
and running `helm upgrade` again. This should cause the first Consul server
to be stopped and restarted with the new image.
Wait until the Consul server cluster is healthy again (30s to a few minutes)
then decrease `updatePartition` and upgrade again. Continue until
`updatePartition` is `0` . At this point, you may remove the
`updatePartition` configuration. Your server upgrade is complete.
#### Upgrading Consul Clients
With the servers upgraded, it is time to upgrade the clients. To upgrade
the clients, set the `client.image` value to the desired Consul version.
Then, run `helm upgrade` . This will upgrade the clients in batches, waiting
until the clients come up healthy before continuing.
2019-12-03 19:14:25 +00:00
### Using Existing Persistent Volume Claims (PVCs)
The only way to use a pre-created PVC is to name them in the format Kubernetes expects:
```
data-< kubernetes namespace > -< helm release name > -consul-server-< ordinal >
```
The Kubernetes namespace you are installing into, helm release name, and ordinal
must match between your Consul servers and your pre-created PVCs. You only
need as many PVCs as you have Consul servers. For example, given a Kubernetes
namespace of "vault" and a release name of "consul" and 5 servers, you would need
to create PVCs with the following names:
```
data-vault-consul-consul-server-0
data-vault-consul-consul-server-1
data-vault-consul-consul-server-2
data-vault-consul-consul-server-3
data-vault-consul-consul-server-4
```
If you are using your own storage, you'll need to configure a storage class. See the
documentation for configuring storage classes [here ](https://kubernetes.io/docs/concepts/storage/storage-classes/ ).
2018-09-12 15:44:30 +00:00
## Architecture
We recommend running Consul on Kubernetes with the same
[general architecture ](/docs/internals/architecture.html )
as running it anywhere else. There are some benefits Kubernetes can provide
that eases operating a Consul cluster and we document those below. The standard
2019-05-15 15:49:41 +00:00
[production deployment guide ](https://learn.hashicorp.com/consul/datacenter-deploy/deployment-guide ) is still an
2018-09-12 15:44:30 +00:00
important read even if running Consul within Kubernetes.
Each section below will outline the different components of running Consul
on Kubernetes and an overview of the resources that are used within the
Kubernetes cluster.
### Server Agents
The server agents are run as a **StatefulSet** , using persistent volume
claims to store the server state. This also ensures that the
[node ID ](/docs/agent/options.html#_node_id ) is persisted so that servers
can be rescheduled onto new IP addresses without causing issues. The server agents
are configured with
[anti-affinity ](https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#affinity-and-anti-affinity )
rules so that they are placed on different nodes. A readiness probe is
configured that marks the pod as ready only when it has established a leader.
A **Service** is registered to represent the servers and expose the various
ports. The DNS address of this service is used to join the servers to each
other without requiring any other access to the Kubernetes cluster. The
service is configured to publish non-ready endpoints so that it can be used
for joining during bootstrap and upgrades.
Additionally, a **PodDisruptionBudget** is configured so the Consul server
cluster maintains quorum during voluntary operational events. The maximum
unavailable is `(n/2)-1` where `n` is the number of server agents.
2018-11-02 01:57:50 +00:00
-> **Note:** Kubernetes and Helm do not delete Persistent Volumes or Persistent
Volume Claims when a
[StatefulSet is deleted ](https://kubernetes.io/docs/concepts/workloads/controllers/statefulset/#stable-storage ),
so this must done manually when removing servers.
2018-09-12 15:44:30 +00:00
### Client Agents
The client agents are run as a **DaemonSet** . This places one agent
(within its own pod) on each Kubernetes node.
The clients expose the Consul HTTP API via a static port (default 8500)
bound to the host port. This enables all other pods on the node to connect
to the node-local agent using the host IP that can be retrieved via the
Kubernetes downward API. See
[accessing the Consul HTTP API ](/docs/platform/k8s/run.html#accessing-the-consul-http-api )
for an example.
There is a major limitation to this: there is no way to bind to a local-only
host port. Therefore, any other node can connect to the agent. This should be
considered for security. For a properly production-secured agent with TLS
and ACLs, this is safe.
Some people prefer to run **Consul agent per pod** architectures, since this
2019-04-16 22:03:03 +00:00
makes it easy to register the pod as a service. However, this turns
2018-09-12 15:44:30 +00:00
a pod into a "node" in Consul and also causes an explosion of resource usage
since every pod needs a Consul agent. We recommend instead running an
agent (in a dedicated pod) per node, via the DaemonSet. This maintains the
node equivalence in Consul. Service registration should be handled via the
catalog syncing feature with Services rather than pods.
-> **Note:** Due to a limitation of anti-affinity rules with DaemonSets,
2019-01-28 22:25:56 +00:00
a client-mode agent runs alongside server-mode agents in Kubernetes. This
2018-09-12 15:44:30 +00:00
duplication wastes some resources, but otherwise functions perfectly fine.