Apache Spark VS Airflow

Compare Apache Spark vs Airflow and see what are their differences.

Our great sponsors
  • InfluxDB - Power Real-Time Data Analytics at Scale
  • WorkOS - The modern identity platform for B2B SaaS
  • SaaSHub - Software Alternatives and Reviews
Apache Spark Airflow
101 169
38,320 34,485
1.1% 2.1%
10.0 10.0
6 days ago about 14 hours ago
Scala Python
Apache License 2.0 Apache License 2.0
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

Apache Spark

Posts with mentions or reviews of Apache Spark. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2024-03-11.

Airflow

Posts with mentions or reviews of Airflow. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2023-12-07.

What are some alternatives?

When comparing Apache Spark and Airflow you can also consider the following projects:

Trino - Official repository of Trino, the distributed SQL query engine for big data, formerly known as PrestoSQL (https://trino.io)

Kedro - Kedro is a toolbox for production-ready data science. It uses software engineering best practices to help you create data engineering and data science pipelines that are reproducible, maintainable, and modular.

Pytorch - Tensors and Dynamic neural networks in Python with strong GPU acceleration

dagster - An orchestration platform for the development, production, and observation of data assets.

Scalding - A Scala API for Cascading

n8n - Free and source-available fair-code licensed workflow automation tool. Easily automate tasks across different services.

mrjob - Run MapReduce jobs on Hadoop or Amazon Web Services

luigi - Luigi is a Python module that helps you build complex pipelines of batch jobs. It handles dependency resolution, workflow management, visualization etc. It also comes with Hadoop support built in.

Dask - Parallel computing with task scheduling

Apache Arrow - Apache Arrow is a multi-language toolbox for accelerated data interchange and in-memory processing

Pandas - Flexible and powerful data analysis / manipulation library for Python, providing labeled data structures similar to R data.frame objects, statistical functions, and much more