Use a single programming model for both batch and streaming use cases.
Write and share new SDKs, IO connectors, and transformation libraries.
Execute pipelines on multiple execution environments.
Community-based development and support to help evolve your application and use cases.
Beam reads your data from a diverse set of supported sources, no matter if it’s on-prem or in the cloud.
Beam executes your business logic for Batch and Streaming use cases.
Beam writes the results of your data processing logic to the most popular data destinations in the industry.
A Beam pipeline can execute in the most popular distributed data processing systems such as Spark, Flink or Samza.
You can write Apache Beam pipelines in your programming language of choice: Java, Python and Go. Learn More.