conduit
timely-dataflow
conduit | timely-dataflow | |
---|---|---|
7 | 11 | |
348 | 3,157 | |
3.2% | 0.9% | |
9.5 | 7.0 | |
about 4 hours ago | 11 days ago | |
Go | Rust | |
Apache License 2.0 | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
conduit
-
Pulling CDC data from Postgres
I'd like to mention Conduit + its Postgres connector. The Pg connector comes built-in, so all that is needed is a single Conduit binary to get started. It relies on WAL, but the connector creates the replication slot itself (if needed).
-
How to connect already setup kafka cluster to mongodb?
GitHub - ConduitIO/Conduit: Data Integration for Production Data Stores. Conduit is meant to be a bit more general-purpose than Kafka Connect and is an easy drop-in replacement. We're working hard to make that even easier. We are still in the early stages of this project and are trying to build more and more connectors. You can find out more about our connector roadmap on the Github repo. Our connector philosophy is to be real-time first, and double down on change data capture (CDC) capabilities, all with permissive licensing.
-
What services you guys used for CDC (Change Data capture) for Sql as well as no sql databases ?
If you're looking for a tool with a UI and in which you can also easily extend the functionality with your own, custom data connectors, you might also want take a look at Conduit which is another open-source tool we've developed to make building and running real-time data infrastructure more straightforward and less time consuming.
-
Alternative Kafka Integration Framework to Kafka Connect?
You might want to check out: https://github.com/conduitio/conduit
-
Where is the modern data stack for software engineers?
This is why we are working on a project called Conduit at Meroxa. We hope to change the experience software engineers have with data.
- Conduit: Data Integration for Production Data Stores
- Conduit: Data Integration Tool for Production Data Stores written in Go
timely-dataflow
-
Readyset: A MySQL and Postgres wire-compatible caching layer
They have a bit about their technical foundation here[0].
Given that Readyset was co-founded by Jon Gjengset (but has apparently since departed the company), who authored the paper on Noria[1], I would assume that Readyset is the continuation of that research.
So it shares some roots with Materialize. They have a common conceptual ancestry in Naiad, where Materialize evolved out of timely-dataflow.
[0]: https://docs.readyset.io/concepts/streaming-dataflow
[1]: https://jon.thesquareplanet.com/papers/osdi18-noria.pdf
[2]: https://dl.acm.org/doi/10.1145/2517349.2522738
[3]: https://github.com/TimelyDataflow/timely-dataflow
-
Mandala: experiment data management as a built-in (Python) language feature
And systems like timely dataflow, https://github.com/TimelyDataflow/timely-dataflow
-
Arroyo: A distributed stream processing engine written in Rust
Project looks cool! Glad you open sourced it. It could use some comments in the code base to help contributors ;). I also like the datafusion usage, that is awesome. BTW I work on github.com/bytewax/bytewax, which is based on https://github.com/TimelyDataflow/timely-dataflow another Rust dataflow computation engine.
-
Rust MPI -- Will there ever be a fully oxidized implementation?
Just found this https://github.com/TimelyDataflow/timely-dataflow and my heart skipped a beat.
-
Streaming processing in Python using Timely Dataflow with Bytewax
Bytewax is a Python native binding to the Timely Dataflow library (written in Rust) for building highly scalable streaming (and batch) processing pipelines.
-
Alternative Kafka Integration Framework to Kafka Connect?
I am working on Bytewax, which is a Python stream processing framework built on Timely Dataflow. It is not exactly a Kafka integration framework because it is a more of a general stream processing framework, but might be interesting for you. We are focused on enabling people to more easily debug, containerize, parallelize and customize and less on enabling a declarative integration framework. It is still early days for us! And we are looking for feedback and ideas from the community.
-
[AskJS] JavaScript for data processing
We used to use a library called Pond.js, https://github.com/esnet/pond, but the reliance on Immutable.JS caused some performance pitfalls, so we wrote a system from scratch that deals with data in a batched streaming fashion. A lot of the concepts were borrowed from a Rust library called timely-dataflow, https://github.com/TimelyDataflow/timely-dataflow.
-
Dataflow: An Efficient Data Processing Library for Machine Learning
Though the name "Dataflow" might be an unfortunate name conflict with another Rust project: https://github.com/TimelyDataflow/timely-dataflow
-
Ask HN: Is there a way to subscribe to an SQL query for changes?
> In the simplest case, I'm talking about regular SQL non-materialized views which are essentially inlined.
I see that now -- makes sense!
> Wish we had some better database primitives to assemble rather than building everything on Postgres - its not ideal for a lot of things.
I'm curious to hear more about this! We agree that better primitives are required and that's why Materialize is written in Rust using using TimelyDataflow[1] and DifferentialDataflow[2] (both developed by Materialize co-founder Frank McSherry). The only relationship between Materialize and Postgres is that we are wire-compatible with Postgres and we don't share any code with Postgres nor do we have a dependence on it.
[1] https://github.com/TimelyDataflow/timely-dataflow
-
7 Real-Time Data Streaming Tools You Should Consider On Your Next Project
Under the hood, Materialize uses Timely Dataflow (TDF) as the stream-processing engine. This allows Materialize to take advantage of the distributed data-parallel compute engine. The great thing about using TDF is that it has been in open source development since 2014 and has since been battle-tested in production at large Fortune 1000-scale companies.
What are some alternatives?
turbine-go - Turbine Library for Go
noria - Fast web applications through dynamic, partially-stateful dataflow
dozer - Dozer is a real-time data movement tool that leverages CDC from various sources and moves data into various sinks.
differential-datalog - DDlog is a programming language for incremental computation. It is well suited for writing programs that continuously update their output in response to input changes. A DDlog programmer does not write incremental algorithms; instead they specify the desired input-output mapping in a declarative manner.
sqlpipe - SQLpipe makes it easy to move the result of one query from one database to another.
materialize - The data warehouse for operational workloads.
bytewax - Python Stream Processing
Benthos - Fancy stream processing made operationally mundane
realtime - Broadcast, Presence, and Postgres Changes via WebSockets
deprecated-core - 🔮 Instill Core contains components for supporting Instill VDP and Instill Model
differential-dataflow - An implementation of differential dataflow using timely dataflow on Rust.