3.9 KiB
layout | page_title | sidebar_current | description |
---|---|---|---|
intro | Registering Health Checks | gettingstarted-checks | We've now seen how simple it is to run Consul, add nodes and services, and query those nodes and services. In this section, we will continue our tour by adding health checks to both nodes and services. Health checks are a critical component of service discovery that prevents using services that are unhealthy. |
Health Checks
We've now seen how simple it is to run Consul, add nodes and services, and query those nodes and services. In this section, we will continue our tour by adding health checks to both nodes and services. Health checks are a critical component of service discovery that prevents using services that are unhealthy.
This page will build upon the previous page and assumes you have a two node cluster running.
Defining Checks
Similar to a service, a check can be registered either by providing a check definition or by making the appropriate calls to the HTTP API.
We will use the check definition because, just like with services, definitions are the most common way to set up checks.
Let's create two definition files in the Consul configuration directory of the second node:
$ echo '{"check": {"name": "ping", "script": "ping -c1 google.com >/dev/null", "interval": "30s"}}' >/etc/consul.d/ping.json
$ echo '{"service": {"name": "web", "tags": ["rails"], "port": 80,
"check": {"script": "curl localhost:80 >/dev/null 2>&1", "interval": "10s"}}}' >/etc/consul.d/web.json
The first definition adds a host-level check named "ping". This check runs
on a 30 second interval, invoking ping -c1 google.com
. If the command
exits with a non-zero exit code, then the node will be flagged unhealthy.
This is the contract for any script
-based health check.
The second command modifies the service named web
, adding a check that sends a
request every 10 seconds via curl to verify that the web server is accessible.
As with the host-level health check, if the script exits with a non-zero exit code,
the service will be flagged unhealthy.
Now restart the second agent or send it a SIGHUP
signal. You should now see the
following log lines:
==> Starting Consul agent...
...
[INFO] agent: Synced service 'web'
[INFO] agent: Synced check 'service:web'
[INFO] agent: Synced check 'ping'
[WARN] Check 'service:web' is now critical
The first few log lines indicate that the agent has synced the new
definitions. The last line indicates that the check we added for
the web
service is critical. This is because we're not actually running
a web server, so the curl test is failing!
Checking Health Status
Now that we've added some simple checks, we can use the HTTP API to inspect them. First, we can look for any failing checks using this command (note, this can be run on either node):
$ curl http://localhost:8500/v1/health/state/critical
[{"Node":"agent-two","CheckID":"service:web","Name":"Service 'web' check","Status":"critical","Notes":"","ServiceID":"web","ServiceName":"web"}]
We can see that there is only a single check, our web
service check, in the
critical
state.
Additionally, we can attempt to query the web service using DNS. Consul will not return any results since the service is unhealthy:
dig @127.0.0.1 -p 8600 web.service.consul
...
;; QUESTION SECTION:
;web.service.consul. IN A
In this section, you learned how easy it is to add health checks. Check definitions
can be updated by changing configuration files and sending a SIGHUP
to the agent.
Alternatively the HTTP API can be used to add, remove, and modify checks dynamically.
The API allows for a "dead man's switch" or TTL-based check.
TTL checks can be used to integrate an application more tightly with Consul, enabling
business logic to be evaluated as part of assessing the state of a check.