What’s your process for deploying a data pipeline from a notebook, running it, and managing it in production?

This page summarizes the projects mentioned and recommended in the original post on /r/dataengineering

Our great sponsors
  • WorkOS - The modern identity platform for B2B SaaS
  • InfluxDB - Power Real-Time Data Analytics at Scale
  • SaaSHub - Software Alternatives and Reviews
  • datahub

    The Metadata Platform for your Data Stack

  • Something like this? https://datahubproject.io/

  • Mage

    🧙 The modern replacement for Airflow. Mage is an open-source data pipeline tool for transforming and integrating data. https://github.com/mage-ai/mage-ai

  • Reverse ETL: new hot one https://github.com/mage-ai/mage-ai

  • WorkOS

    The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.

    WorkOS logo
  • featureform

    The Virtual Feature Store. Turn your existing data infrastructure into a feature store.

  • Feature store: new hot one: https://www.featureform.com/

  • neon

    Neon: Serverless Postgres. We separated storage and compute to offer autoscaling, branching, and bottomless storage.

  • App DBs: new hot one on the scene https://neon.tech/

NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts