airflow-testing-ci-workflow VS patterns-devkit

Compare airflow-testing-ci-workflow vs patterns-devkit and see what are their differences.

Our great sponsors
  • WorkOS - The modern identity platform for B2B SaaS
  • InfluxDB - Power Real-Time Data Analytics at Scale
  • SaaSHub - Software Alternatives and Reviews
airflow-testing-ci-workflow patterns-devkit
2 5
81 106
- 0.0%
0.0 2.9
about 3 years ago about 1 year ago
Python Python
- BSD 3-clause "New" or "Revised" License
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

airflow-testing-ci-workflow

Posts with mentions or reviews of airflow-testing-ci-workflow. We have used some of these posts to build our list of alternatives and similar projects.

patterns-devkit

Posts with mentions or reviews of patterns-devkit. We have used some of these posts to build our list of alternatives and similar projects.

What are some alternatives?

When comparing airflow-testing-ci-workflow and patterns-devkit you can also consider the following projects:

airflow-api-tests - This is a collection of Pytest for the 2.0 Stable Rest Apis for Apache Airflow. I have another repo where you could setup airflow locally and play around with these. I am used to RestAssured, but trying out pytest here.

pyspark-example-project - Implementing best practices for PySpark ETL jobs and applications.

gretel-airflow-pipelines - Runbooks for running Gretel on Apache Airflow

Dataplane - Dataplane is a data platform that makes it easy to construct a data mesh with automated data pipelines and workflows.

pipebird - Pipebird is open source infrastructure for securely sharing data with customers.

hamilton - Hamilton helps data scientists and engineers define testable, modular, self-documenting dataflows, that encode lineage and metadata. Runs and scales everywhere python does.

SmartPipeline - A framework for rapid development of robust data pipelines following a simple design pattern

AWS Data Wrangler - pandas on AWS - Easy integration with Athena, Glue, Redshift, Timestream, Neptune, OpenSearch, QuickSight, Chime, CloudWatchLogs, DynamoDB, EMR, SecretManager, PostgreSQL, MySQL, SQLServer and S3 (Parquet, CSV, JSON and EXCEL).

flowrunner - Flowrunner is a lightweight package to organize and represent Data Engineering/Science workflows

prism - Prism is the easiest way to develop, orchestrate, and execute data pipelines in Python.

versatile-data-kit - One framework to develop, deploy and operate data workflows with Python and SQL.

workshop-realtime-data-pipelines - You will inspect and run a sample architecture making use of Apache Pulsarâ„¢ and Pulsar Functions for real-time, event-streaming-based data ingestion, cleaning and processing.