Using the Apache Flink Runner
The Apache Flink Runner can be used to execute Beam pipelines using Apache Flink. When using the Flink Runner you will create a jar file containing your job that can be executed on a regular Flink cluster. It’s also possible to execute a Beam pipeline using Flink’s local execution mode without setting up a cluster. This is helpful for development and debugging of your pipeline.
The Flink Runner and Flink are suitable for large scale, continuous jobs, and provide:
- A streaming-first runtime that supports both batch processing and data streaming programs
- A runtime that supports very high throughput and low event latency at the same time
- Fault-tolerance with exactly-once processing guarantees
- Natural back-pressure in streaming programs
- Custom memory management for efficient and robust switching between in-memory and out-of-core data processing algorithms
- Integration with YARN and other components of the Apache Hadoop ecosystem
The Beam Capability Matrix documents the supported capabilities of the Flink Runner.
Flink Runner prerequisites and setup
If you want to use the local execution mode with the Flink runner to don’t have to complete any setup.
To use the Flink Runner for executing on a cluster, you have to setup a Flink cluster by following the Flink setup quickstart.
To find out which version of Flink you need you can run this command to check the version of the Flink dependency that your project is using:
$ mvn dependency:tree -Pflink-runner |grep flink ... [INFO] | +- org.apache.flink:flink-streaming-java_2.10:jar:1.2.1:runtime ...
Here, we would need Flink 1.2.1. Please also note the Scala version in the dependency name. In this case we need to make sure to use a Flink cluster with Scala version 2.10.
For more information, the Flink Documentation can be helpful.
Specify your dependency
When using Java, you must specify your dependency on the Flink Runner in your
<dependency> <groupId>org.apache.beam</groupId> <artifactId>beam-runners-flink_2.10</artifactId> <version>2.2.0</version> <scope>runtime</scope> </dependency>
This section is not applicable to the Beam SDK for Python.
Executing a pipeline on a Flink cluster
For executing a pipeline on a Flink cluster you need to package your program along will all dependencies in a so-called fat jar. How you do this depends on your build system but if you follow along the Beam Quickstart this is the command that you have to run:
$ mvn package -Pflink-runner
The Beam Quickstart Maven project is setup to use the Maven Shade plugin to create a fat jar and the
-Pflink-runner argument makes sure to include the dependency on the Flink Runner.
For actually running the pipeline you would use this command
$ mvn exec:java -Dexec.mainClass=org.apache.beam.examples.WordCount \ -Pflink-runner \ -Dexec.args="--runner=FlinkRunner \ --inputFile=/path/to/pom.xml \ --output=/path/to/counts \ --flinkMaster=<flink master url> \ --filesToStage=target/word-count-beam--bundled-0.1.jar"
If you have a Flink
JobManager running on your local machine you can give
Pipeline options for the Flink Runner
When executing your pipeline with the Flink Runner, you can set these pipeline options.
||The pipeline runner to use. This option allows you to determine the pipeline runner at runtime.||Set to
||Whether streaming mode is enabled or disabled;
||The url of the Flink JobManager on which to execute pipelines. This can either be the address of a cluster JobManager, in the form
||Jar Files to send to all workers and put on the classpath. Here you have to put the fat jar that contains your program along with all dependencies.||empty|
||The degree of parallelism to be used when distributing operations onto workers.||
||The interval between consecutive checkpoints (i.e. snapshots of the current pipeline state used for fault tolerance).||
||Sets the number of times that failed tasks are re-executed. A value of
||Sets the delay between executions. A value of
||Sets the state backend to use in streaming mode. The default is to read this setting from the Flink config.||
Additional information and caveats
Monitoring your job
You can monitor a running Flink job using the Flink JobManager Dashboard. By default, this is available at port
8081 of the JobManager node. If you have a Flink installation on your local machine that would be
If your pipeline uses an unbounded data source or sink, the Flink Runner will automatically switch to streaming mode. You can enforce streaming mode by using the
streaming setting mentioned above.