DC/OS includes a comprehensive metrics service, providing metrics from DC/OS cluster hosts, containers, and applications running on those hosts. It collects, tags, and transmits metrics from every node, container, and application in your DC/OS cluster. Metrics are available via the DC/OS metrics API, allowing for easy integration with a wide range of monitoring solutions.
Types of Metrics
DC/OS collects four types of metrics as follows:
- System: Metrics about each node in the DC/OS cluster.
- Component: Metrics about the components which make up DC/OS.
- Container: Metrics about
cgroupallocations from tasks running in the DC/OS Universal Container Runtime or Docker Engine runtime.
- Application: Metrics emitted from any application running on the Universal Container Runtime.
Operations on Metrics
You can perform the following operations on Metrics:
- Adding Metrics: You can add custom metrics to the DC/OS metrics service from your applications. DC/OS metrics listens for
StatsDmetrics from every application running with the Mesos containerizer. A
StatsD serveris exposed for each container, which allows you to tag all metrics by origin. Its address is available to the application by injecting the standard environment variables
Alternatively, you can serve metrics in Prometheus format by exposing an endpoint which serves metrics. Notify the metrics service that your task is serving Prometheus metrics by adding a label to the port which is serving metrics. Any port with the label
DCOS_METRICS_FORMAT=prometheus will be scraped at the rate of one per minute for the lifetime of the task.
A task may expose multiple ports which serve Prometheus metrics. Metrics collected in this manner will be tagged with the originating task’s name and its framework name.
For more information, read the documentation on how to add custom metrics into the DC/OS metrics API using Python.
Tagging Metrics: The tags are automatically added to the metrics in order to identify, support easy drill-down, filter, and group metrics data. The tags are not limited to the following list:
- Container identification (for all container and application metrics): container_id, executor_id, framework_id, framework_name
- Application identification (e.g., for Marathon apps): application_name
- System identification: agent_id
Example: An administrator can identify the group of metrics based on the following tags:
agent_id- Detects a situation which is system-specific (e.g., a faulty disk).
framework_id- Detects which Mesos frameworks are using the most resources in the system.
container_id- Shows the same information except with per-container granularity.
Forwarding Metrics: After you have collected the metrics from the applications and hosts, it can be forwarded to various metrics stores. Configure Telegraf plugin to forward the metrics data from the cluster.
Monitoring and Visualizing Metrics: It is recommended to install the DC/OS Monitoring service to monitor and visualize metrics in your DC/OS cluster. See the service documentation for instructions on how to install and use the service.
Metrics in DC/OS 1.12 and newer versions use Telegraf to collect and process data. Telegraf provides a plugin-driven architecture. The custom DC/OS plugins provide metrics on the performance of DC/OS workloads and DC/OS itself. Telegraf provides metrics from DC/OS cluster hosts, containers running on those hosts, and from applications running on DC/OS using the
StatsD process. It is natively integrated with DC/OS. By default, it exposes metrics in Prometheus format from
port 61091 on each node, and in JSON format through the DC/OS Metrics API. Telegraf is included in the DC/OS distribution and runs on every host in the cluster.
Telegraf collects application and custom metrics through the
dcos_statsd plugin. A dedicated
StatsD server is started for each new task. Any metrics received by the
StatsD server are tagged with the task name and its service name. The address of the server is provided by environment variables (
The metrics collected by
dcos_statsd are gathered every 30 seconds. The task must run for at least 30 seconds to ensure a task’s metrics are gathered.
For more information about the list of metrics that are automatically collected by DC/OS, read Metrics Reference documentation.
Use the following troubleshooting guidelines to resolve errors:
- You can collect metrics about Telegraf’s own performance by enabling the
- You can check the status of the Telegraf
systemdunit by running
systemctl status dcos-telegraf.
- Logs are available from journald via
journalctl -u dcos-telegraf.
Metrics Plugin Architecture
How DC/OS collects and publishes metrics…Read More
Metrics Quick Start
Getting Started with metrics in DC/OS…Read More
Monitoring Mesos with Telegraf…Read More
Export DC/OS Metrics to Datadog
Sending DC/OS metrics to Datadog…Read More
Using the Metrics API…Read More
Understanding metrics collected by DC/OS…Read More