title: “Metrics” nav-parent_id: monitoring nav-pos: 1

Flink exposes a metric system that allows gathering and exposing metrics to external systems.

  • This will be replaced by the TOC {:toc}

Registering metrics

You can access the metric system from any user function that extends [RichFunction]({{ site.baseurl }}/dev/api_concepts.html#rich-functions) by calling getRuntimeContext().getMetricGroup(). This method returns a MetricGroup object on which you can create and register new metrics.

Metric types

Flink supports Counters, Gauges, Histograms and Meters.

Counter

A Counter is used to count something. The current value can be in- or decremented using inc()/inc(long n) or dec()/dec(long n). You can create and register a Counter by calling counter(String name) on a MetricGroup.

public class MyMapper extends RichMapFunction<String, String> { private transient Counter counter;

@Override public void open(Configuration config) { this.counter = getRuntimeContext() .getMetricGroup() .counter(“myCounter”); }

@Override public String map(String value) throws Exception { this.counter.inc(); return value; } }

{% endhighlight %}

class MyMapper extends RichMapFunction[String,String] { @transient private var counter: Counter = _

override def open(parameters: Configuration): Unit = { counter = getRuntimeContext() .getMetricGroup() .counter(“myCounter”) }

override def map(value: String): String = { counter.inc() value } }

{% endhighlight %}

Alternatively you can also use your own Counter implementation:

public class MyMapper extends RichMapFunction<String, String> { private transient Counter counter;

@Override public void open(Configuration config) { this.counter = getRuntimeContext() .getMetricGroup() .counter(“myCustomCounter”, new CustomCounter()); }

@Override public String map(String value) throws Exception { this.counter.inc(); return value; } }

{% endhighlight %}

class MyMapper extends RichMapFunction[String,String] { @transient private var counter: Counter = _

override def open(parameters: Configuration): Unit = { counter = getRuntimeContext() .getMetricGroup() .counter(“myCustomCounter”, new CustomCounter()) }

override def map(value: String): String = { counter.inc() value } }

{% endhighlight %}

Gauge

A Gauge provides a value of any type on demand. In order to use a Gauge you must first create a class that implements the org.apache.flink.metrics.Gauge interface. There is no restriction for the type of the returned value. You can register a gauge by calling gauge(String name, Gauge gauge) on a MetricGroup.

public class MyMapper extends RichMapFunction<String, String> { private transient int valueToExpose = 0;

@Override public void open(Configuration config) { getRuntimeContext() .getMetricGroup() .gauge(“MyGauge”, new Gauge() { @Override public Integer getValue() { return valueToExpose; } }); }

@Override public String map(String value) throws Exception { valueToExpose++; return value; } }

{% endhighlight %}

new class MyMapper extends RichMapFunction[String,String] { @transient private var valueToExpose = 0

override def open(parameters: Configuration): Unit = { getRuntimeContext() .getMetricGroup() .gauge[Int, ScalaGauge[Int]](“MyGauge”, ScalaGauge[Int]( () => valueToExpose ) ) }

override def map(value: String): String = { valueToExpose += 1 value } }

{% endhighlight %}

Note that reporters will turn the exposed object into a String, which means that a meaningful toString() implementation is required.

Histogram

A Histogram measures the distribution of long values. You can register one by calling histogram(String name, Histogram histogram) on a MetricGroup.

@Override public void open(Configuration config) { this.histogram = getRuntimeContext() .getMetricGroup() .histogram(“myHistogram”, new MyHistogram()); }

@Override public Long map(Long value) throws Exception { this.histogram.update(value); return value; } } {% endhighlight %}

class MyMapper extends RichMapFunction[Long,Long] { @transient private var histogram: Histogram = _

override def open(parameters: Configuration): Unit = { histogram = getRuntimeContext() .getMetricGroup() .histogram(“myHistogram”, new MyHistogram()) }

override def map(value: Long): Long = { histogram.update(value) value } }

{% endhighlight %}

Flink does not provide a default implementation for Histogram, but offers a {% gh_link flink-metrics/flink-metrics-dropwizard/src/main/java/org/apache/flink/dropwizard/metrics/DropwizardHistogramWrapper.java “Wrapper” %} that allows usage of Codahale/DropWizard histograms. To use this wrapper add the following dependency in your pom.xml: {% highlight xml %} com.alibaba.blink flink-metrics-dropwizard {{site.version}} {% endhighlight %}

You can then register a Codahale/DropWizard histogram like this:

@Override public void open(Configuration config) { com.codahale.metrics.Histogram dropwizardHistogram = new com.codahale.metrics.Histogram(new SlidingWindowReservoir(500));

this.histogram = getRuntimeContext()
  .getMetricGroup()
  .histogram("myHistogram", new DropwizardHistogramWrapper(dropwizardHistogram));

}

@Override public Long map(Long value) throws Exception { this.histogram.update(value); return value; } } {% endhighlight %}

class MyMapper extends RichMapFunction[Long, Long] { @transient private var histogram: Histogram = _

override def open(config: Configuration): Unit = { com.codahale.metrics.Histogram dropwizardHistogram = new com.codahale.metrics.Histogram(new SlidingWindowReservoir(500))

histogram = getRuntimeContext()
  .getMetricGroup()
  .histogram("myHistogram", new DropwizardHistogramWrapper(dropwizardHistogram))

}

override def map(value: Long): Long = { histogram.update(value) value } }

{% endhighlight %}

Meter

A Meter measures an average throughput. An occurrence of an event can be registered with the markEvent() method. Occurrence of multiple events at the same time can be registered with markEvent(long n) method. You can register a meter by calling meter(String name, Meter meter) on a MetricGroup.

@Override public void open(Configuration config) { this.meter = getRuntimeContext() .getMetricGroup() .meter(“myMeter”, new MyMeter()); }

@Override public Long map(Long value) throws Exception { this.meter.markEvent(); return value; } } {% endhighlight %}

class MyMapper extends RichMapFunction[Long,Long] { @transient private var meter: Meter = _

override def open(config: Configuration): Unit = { meter = getRuntimeContext() .getMetricGroup() .meter(“myMeter”, new MyMeter()) }

override def map(value: Long): Long = { meter.markEvent() value } }

{% endhighlight %}

Flink offers a {% gh_link flink-metrics/flink-metrics-dropwizard/src/main/java/org/apache/flink/dropwizard/metrics/DropwizardMeterWrapper.java “Wrapper” %} that allows usage of Codahale/DropWizard meters. To use this wrapper add the following dependency in your pom.xml: {% highlight xml %} com.alibaba.blink flink-metrics-dropwizard {{site.version}} {% endhighlight %}

You can then register a Codahale/DropWizard meter like this:

@Override public void open(Configuration config) { com.codahale.metrics.Meter dropwizardMeter = new com.codahale.metrics.Meter();

this.meter = getRuntimeContext()
  .getMetricGroup()
  .meter("myMeter", new DropwizardMeterWrapper(dropwizardMeter));

}

@Override public Long map(Long value) throws Exception { this.meter.markEvent(); return value; } } {% endhighlight %}

class MyMapper extends RichMapFunction[Long,Long] { @transient private var meter: Meter = _

override def open(config: Configuration): Unit = { com.codahale.metrics.Meter dropwizardMeter = new com.codahale.metrics.Meter()

meter = getRuntimeContext()
  .getMetricGroup()
  .meter("myMeter", new DropwizardMeterWrapper(dropwizardMeter))

}

override def map(value: Long): Long = { meter.markEvent() value } }

{% endhighlight %}

Scope

Every metric is assigned an identifier and a set of key-value pairs under which the metric will be reported.

THe identifier is based on 3 components: the user-defined name when registering the metric, an optional user-defined scope and a system-provided scope. For example, if A.B is the system scope, C.D the user scope and E the name, then the identifier for the metric will be A.B.C.D.E.

You can configure which delimiter to use for the identifier (default: .) by setting the metrics.scope.delimiter key in conf/flink-conf.yaml.

User Scope

You can define a user scope by calling MetricGroup#addGroup(String name), MetricGroup#addGroup(int name) or Metric#addGroup(String key, String value). These methods affect what MetricGroup#getMetricIdentifier and MetricGroup#getScopeComponents return.

counter = getRuntimeContext() .getMetricGroup() .addGroup(“MyMetrics”) .counter(“myCounter”);

counter = getRuntimeContext() .getMetricGroup() .addGroup(“MyMetricsKey”, “MyMetricsValue”) .counter(“myCounter”);

{% endhighlight %}

counter = getRuntimeContext() .getMetricGroup() .addGroup(“MyMetrics”) .counter(“myCounter”)

counter = getRuntimeContext() .getMetricGroup() .addGroup(“MyMetricsKey”, “MyMetricsValue”) .counter(“myCounter”)

{% endhighlight %}

System Scope

The system scope contains context information about the metric, for example in which task it was registered or what job that task belongs to.

Which context information should be included can be configured by setting the following keys in conf/flink-conf.yaml. Each of these keys expect a format string that may contain constants (e.g. “taskmanager”) and variables (e.g. “<task_id>”) which will be replaced at runtime.

  • metrics.scope.jm
    • Default: <host>.jobmanager
    • Applied to all metrics that were scoped to a job manager.
  • metrics.scope.jm.job
    • Default: <host>.jobmanager.<job_name>
    • Applied to all metrics that were scoped to a job manager and job.
  • metrics.scope.tm
    • Default: <host>.taskmanager.<tm_id>
    • Applied to all metrics that were scoped to a task manager.
  • metrics.scope.tm.job
    • Default: <host>.taskmanager.<tm_id>.<job_name>
    • Applied to all metrics that were scoped to a task manager and job.
  • metrics.scope.task
    • Default: <host>.taskmanager.<tm_id>.<job_name>.<task_name>.<subtask_index>
    • Applied to all metrics that were scoped to a task.
  • metrics.scope.operator
    • Default: <host>.taskmanager.<tm_id>.<job_name>.<operator_name>.<subtask_index>
    • Applied to all metrics that were scoped to an operator.

There are no restrictions on the number or order of variables. Variables are case sensitive.

The default scope for operator metrics will result in an identifier akin to localhost.taskmanager.1234.MyJob.MyOperator.0.MyMetric

If you also want to include the task name but omit the task manager information you can specify the following format:

metrics.scope.operator: <host>.<job_name>.<task_name>.<operator_name>.<subtask_index>

This could create the identifier localhost.MyJob.MySource_->_MyOperator.MyOperator.0.MyMetric.

Note that for this format string an identifier clash can occur should the same job be run multiple times concurrently, which can lead to inconsistent metric data. As such it is advised to either use format strings that provide a certain degree of uniqueness by including IDs (e.g <job_id>) or by assigning unique names to jobs and operators.

List of all Variables

  • JobManager: <host>
  • TaskManager: <host>, <tm_id>
  • Job: <job_id>, <job_name>
  • Task: <task_id>, <task_name>, <task_attempt_id>, <task_attempt_num>, <subtask_index>
  • Operator: <operator_id>,<operator_name>

Important: For the Batch API, <operator_id> is always equal to <task_id>.

User Variables

You can define a user variable by calling MetricGroup#addGroup(String key, String value). This method affects what MetricGroup#getMetricIdentifier, MetricGroup#getScopeComponents and MetricGroup#getAllVariables() returns.

Important: User variables cannot be used in scope formats.

counter = getRuntimeContext() .getMetricGroup() .addGroup(“MyMetricsKey”, “MyMetricsValue”) .counter(“myCounter”);

{% endhighlight %}

counter = getRuntimeContext() .getMetricGroup() .addGroup(“MyMetricsKey”, “MyMetricsValue”) .counter(“myCounter”)

{% endhighlight %}

Reporter

Metrics can be exposed to an external system by configuring one or several reporters in conf/flink-conf.yaml. These reporters will be instantiated on each job and task manager when they are started.

  • metrics.reporter.<name>.<config>: Generic setting <config> for the reporter named <name>.
  • metrics.reporter.<name>.class: The reporter class to use for the reporter named <name>.
  • metrics.reporter.<name>.interval: The reporter interval to use for the reporter named <name>.
  • metrics.reporter.<name>.scope.delimiter: The delimiter to use for the identifier (default value use metrics.scope.delimiter) for the reporter named <name>.
  • metrics.reporters: (optional) A comma-separated include list of reporter names. By default all configured reporters will be used.

All reporters must at least have the class property, some allow specifying a reporting interval. Below, we will list more settings specific to each reporter.

Example reporter configuration that specifies multiple reporters:

{% highlight yaml %} metrics.reporters: my_jmx_reporter,my_other_reporter

metrics.reporter.my_jmx_reporter.class: org.apache.flink.metrics.jmx.JMXReporter metrics.reporter.my_jmx_reporter.port: 9020-9040

metrics.reporter.my_other_reporter.class: org.apache.flink.metrics.graphite.GraphiteReporter metrics.reporter.my_other_reporter.host: 192.168.1.1 metrics.reporter.my_other_reporter.port: 10000

{% endhighlight %}

Important: The jar containing the reporter must be accessible when Flink is started by placing it in the /lib folder.

You can write your own Reporter by implementing the org.apache.flink.metrics.reporter.MetricReporter interface. If the Reporter should send out reports regularly you have to implement the Scheduled interface as well.

The following sections list the supported reporters.

JMX (org.apache.flink.metrics.jmx.JMXReporter)

You don't have to include an additional dependency since the JMX reporter is available by default but not activated.

Parameters:

  • port - (optional) the port on which JMX listens for connections. In order to be able to run several instancesof the reporter on one host (e.g. when one TaskManager is colocated with the JobManager) it is advisable to use a port range like 9250-9260. When a range is specified the actual port is shown in the relevant job or task manager log. If this setting is set Flink will start an extra JMX connector for the given port/range. Metrics are always available on the default local JMX interface.

Example configuration:

{% highlight yaml %}

metrics.reporter.jmx.class: org.apache.flink.metrics.jmx.JMXReporter metrics.reporter.jmx.port: 8789

{% endhighlight %}

Metrics exposed through JMX are identified by a domain and a list of key-properties, which together form the object name.

The domain always begins with org.apache.flink followed by a generalized metric identifier. In contrast to the usual identifier it is not affected by scope-formats, does not contain any variables and is constant across jobs. An example for such a domain would be org.apache.flink.job.task.numBytesOut.

The key-property list contains the values for all variables, regardless of configured scope formats, that are associated with a given metric. An example for such a list would be host=localhost,job_name=MyJob,task_name=MyTask.

The domain thus identifies a metric class, while the key-property list identifies one (or multiple) instances of that metric.

Ganglia (org.apache.flink.metrics.ganglia.GangliaReporter)

In order to use this reporter you must copy /opt/flink-metrics-ganglia-{{site.version}}.jar into the /lib folder of your Flink distribution.

Parameters:

  • host - the gmond host address configured under udp_recv_channel.bind in gmond.conf
  • port - the gmond port configured under udp_recv_channel.port in gmond.conf
  • tmax - soft limit for how long an old metric should be retained
  • dmax - hard limit for how long an old metric should be retained
  • ttl - time-to-live for transmitted UDP packets
  • addressingMode - UDP addressing mode to use (UNICAST/MULTICAST)

Example configuration:

{% highlight yaml %}

metrics.reporter.gang.class: org.apache.flink.metrics.ganglia.GangliaReporter metrics.reporter.gang.host: localhost metrics.reporter.gang.port: 8649 metrics.reporter.gang.tmax: 60 metrics.reporter.gang.dmax: 0 metrics.reporter.gang.ttl: 1 metrics.reporter.gang.addressingMode: MULTICAST

{% endhighlight %}

Graphite (org.apache.flink.metrics.graphite.GraphiteReporter)

In order to use this reporter you must copy /opt/flink-metrics-graphite-{{site.version}}.jar into the /lib folder of your Flink distribution.

Parameters:

  • host - the Graphite server host
  • port - the Graphite server port
  • protocol - protocol to use (TCP/UDP)

Example configuration:

{% highlight yaml %}

metrics.reporter.grph.class: org.apache.flink.metrics.graphite.GraphiteReporter metrics.reporter.grph.host: localhost metrics.reporter.grph.port: 2003 metrics.reporter.grph.protocol: TCP

{% endhighlight %}

Prometheus (org.apache.flink.metrics.prometheus.PrometheusReporter)

In order to use this reporter you must copy /opt/flink-metrics-prometheus-{{site.version}}.jar into the /lib folder of your Flink distribution.

Parameters:

  • port - (optional) the port the Prometheus exporter listens on, defaults to 9249. In order to be able to run several instances of the reporter on one host (e.g. when one TaskManager is colocated with the JobManager) it is advisable to use a port range like 9250-9260.

Example configuration:

{% highlight yaml %}

metrics.reporter.prom.class: org.apache.flink.metrics.prometheus.PrometheusReporter

{% endhighlight %}

Flink metric types are mapped to Prometheus metric types as follows:

FlinkPrometheusNote
CounterGaugePrometheus counters cannot be decremented.
GaugeGaugeOnly numbers and booleans are supported.
HistogramSummaryQuantiles .5, .75, .95, .98, .99 and .999
MeterGaugeThe gauge exports the meter's rate.

All Flink metrics variables (see List of all Variables) are exported to Prometheus as labels.

StatsD (org.apache.flink.metrics.statsd.StatsDReporter)

In order to use this reporter you must copy /opt/flink-metrics-statsd-{{site.version}}.jar into the /lib folder of your Flink distribution.

Parameters:

  • host - the StatsD server host
  • port - the StatsD server port

Example configuration:

{% highlight yaml %}

metrics.reporter.stsd.class: org.apache.flink.metrics.statsd.StatsDReporter metrics.reporter.stsd.host: localhost metrics.reporter.stsd.port: 8125

{% endhighlight %}

Datadog (org.apache.flink.metrics.datadog.DatadogHttpReporter)

In order to use this reporter you must copy /opt/flink-metrics-datadog-{{site.version}}.jar into the /lib folder of your Flink distribution.

Note any variables in Flink metrics, such as <host>, <job_name>, <tm_id>, <subtask_index>, <task_name>, and <operator_name>, will be sent to Datadog as tags. Tags will look like host:localhost and job_name:myjobname.

Parameters:

  • apikey - the Datadog API key
  • tags - (optional) the global tags that will be applied to metrics when sending to Datadog. Tags should be separated by comma only

Example configuration:

{% highlight yaml %}

metrics.reporter.dghttp.class: org.apache.flink.metrics.datadog.DatadogHttpReporter metrics.reporter.dghttp.apikey: xxx metrics.reporter.dghttp.tags: myflinkapp,prod

{% endhighlight %}

Slf4j (org.apache.flink.metrics.slf4j.Slf4jReporter)

In order to use this reporter you must copy /opt/flink-metrics-slf4j-{{site.version}}.jar into the /lib folder of your Flink distribution.

Example configuration:

{% highlight yaml %}

metrics.reporter.slf4j.class: org.apache.flink.metrics.slf4j.Slf4jReporter metrics.reporter.slf4j.interval: 60 SECONDS

{% endhighlight %}

System metrics

By default Flink gathers several metrics that provide deep insights on the current state. This section is a reference of all these metrics.

The tables below generally feature 5 columns:

  • The “Scope” column describes which scope format is used to generate the system scope. For example, if the cell contains “Operator” then the scope format for “metrics.scope.operator” is used. If the cell contains multiple values, separated by a slash, then the metrics are reported multiple times for different entities, like for both job- and taskmanagers.

  • The (optional)“Infix” column describes which infix is appended to the system scope.

  • The “Metrics” column lists the names of all metrics that are registered for the given scope and infix.

  • The “Description” column provides information as to what a given metric is measuring.

  • The “Type” column describes which metric type is used for the measurement.

Note that all dots in the infix/metric name columns are still subject to the “metrics.delimiter” setting.

Thus, in order to infer the metric identifier:

  1. Take the scope-format based on the “Scope” column
  2. Append the value in the “Infix” column if present, and account for the “metrics.delimiter” setting
  3. Append metric name.

CPU

Memory

Threads

GarbageCollection

ClassLoader

Network

Availability

Checkpointing

IO

Connectors

Kafka Connectors

Kinesis Connectors

Latency tracking

Flink allows to track the latency of records traveling through the system. To enable the latency tracking a latencyTrackingInterval (in milliseconds) has to be set to a positive value in the ExecutionConfig.

At the latencyTrackingInterval, the sources will periodically emit a special record, called a LatencyMarker. The marker contains a timestamp from the time when the record has been emitted at the sources. Latency markers can not overtake regular user records, thus if records are queuing up in front of an operator, it will add to the latency tracked by the marker.

Note that the latency markers are not accounting for the time user records spend in operators as they are bypassing them. In particular the markers are not accounting for the time records spend for example in window buffers. Only if operators are not able to accept new records, thus they are queuing up, the latency measured using the markers will reflect that.

All intermediate operators keep a list of the last n latencies from each source to compute a latency distribution. The sink operators keep a list from each source, and each parallel source instance to allow detecting latency issues caused by individual machines.

Currently, Flink assumes that the clocks of all machines in the cluster are in sync. We recommend setting up an automated clock synchronisation service (like NTP) to avoid false latency results.

REST API integration

Metrics can be queried through the [Monitoring REST API]({{ site.baseurl }}/monitoring/rest_api.html).

Below is a list of available endpoints, with a sample JSON response. All endpoints are of the sample form http://hostname:8081/jobmanager/metrics, below we list only the path part of the URLs.

Values in angle brackets are variables, for example http://hostname:8081/jobs/<jobid>/metrics will have to be requested for example as http://hostname:8081/jobs/7684be6004e4e955c2a558a9bc463f65/metrics.

Request metrics for a specific entity:

  • /jobmanager/metrics
  • /taskmanagers/<taskmanagerid>/metrics
  • /jobs/<jobid>/metrics
  • /jobs/<jobid>/vertices/<vertexid>/subtasks/<subtaskindex>

Request metrics aggregated across all entities of the respective type:

  • /taskmanagers/metrics
  • /jobs/metrics
  • /jobs/<jobid>/vertices/<vertexid>/subtasks/metrics

Request metrics aggregated over a subset of all entities of the respective type:

  • /taskmanagers/metrics?taskmanagers=A,B,C
  • /jobs/metrics?jobs=D,E,F
  • /jobs/<jobid>/vertices/<vertexid>/subtasks/metrics?subtask=1,2,3

Request a list of available metrics:

GET /jobmanager/metrics

{% highlight json %} [ { “id”: “metric1” }, { “id”: “metric2” } ] {% endhighlight %}

Request the values for specific (unaggregated) metrics:

GET taskmanagers/ABCDE/metrics?get=metric1,metric2

{% highlight json %} [ { “id”: “metric1”, “value”: “34” }, { “id”: “metric2”, “value”: “2” } ] {% endhighlight %}

Request aggregated values for specific metrics:

GET /taskmanagers/metrics?get=metric1,metric2

{% highlight json %} [ { “id”: “metric1”, “min”: 1, “max”: 34, “avg”: 15, “sum”: 45 }, { “id”: “metric2”, “min”: 2, “max”: 14, “avg”: 7, “sum”: 16 } ] {% endhighlight %}

Request specific aggregated values for specific metrics:

GET /taskmanagers/metrics?get=metric1,metric2&agg=min,max

{% highlight json %} [ { “id”: “metric1”, “min”: 1, “max”: 34, }, { “id”: “metric2”, “min”: 2, “max”: 14, } ] {% endhighlight %}

Dashboard integration

Metrics that were gathered for each task or operator can also be visualized in the Dashboard. On the main page for a job, select the Metrics tab. After selecting one of the tasks in the top graph you can select metrics to display using the Add Metric drop-down menu.

  • Task metrics are listed as <subtask_index>.<metric_name>.
  • Operator metrics are listed as <subtask_index>.<operator_name>.<metric_name>.

Each metric will be visualized as a separate graph, with the x-axis representing time and the y-axis the measured value. All graphs are automatically updated every 10 seconds, and continue to do so when navigating to another page.

There is no limit as to the number of visualized metrics; however only numeric metrics can be visualized.

{% top %}