Top Qs
Timeline
Chat
Perspective

Apache Beam

Unified programming model for data processing pipelines From Wikipedia, the free encyclopedia

Apache Beam
Remove ads

Apache Beam is an open source unified programming model to define and execute data processing pipelines, including ETL, batch and stream (continuous) processing.[2] Beam Pipelines are defined using one of the provided SDKs and executed in one of the Beam’s supported runners (distributed processing back-ends) including Apache Flink, Apache Samza, Apache Spark, and Google Cloud Dataflow.[3]

Quick Facts Original author(s), Developer(s) ...
Remove ads

History

Summarize
Perspective

Apache Beam[3] is one implementation of the Dataflow model paper.[4] The Dataflow model is based on previous work on distributed processing abstractions at Google, in particular on FlumeJava[5] and Millwheel.[6][7]

Google released an open SDK implementation of the Dataflow model in 2014 and an environment to execute Dataflows locally (non-distributed) as well as in the Google Cloud Platform service.

Timeline

Apache Beam makes minor releases every 6 weeks.[8]

More information Version, Release date ...


More information Version, Release date ...
Remove ads

See also

References

Loading related searches...

Wikiwand - on

Seamless Wikipedia browsing. On steroids.

Remove ads