timely-dataflow VS diagnostics

Compare timely-dataflow vs diagnostics and see what are their differences.

timely-dataflow

A modular implementation of timely dataflow in Rust (by TimelyDataflow)

diagnostics

Diagnostic tools for timely dataflow computations (by TimelyDataflow)
InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
timely-dataflow diagnostics
11 1
3,157 41
0.9% -
7.0 0.0
9 days ago almost 2 years ago
Rust Rust
MIT License MIT License
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

timely-dataflow

Posts with mentions or reviews of timely-dataflow. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2024-02-21.
  • Readyset: A MySQL and Postgres wire-compatible caching layer
    5 projects | news.ycombinator.com | 21 Feb 2024
    They have a bit about their technical foundation here[0].

    Given that Readyset was co-founded by Jon Gjengset (but has apparently since departed the company), who authored the paper on Noria[1], I would assume that Readyset is the continuation of that research.

    So it shares some roots with Materialize. They have a common conceptual ancestry in Naiad, where Materialize evolved out of timely-dataflow.

    [0]: https://docs.readyset.io/concepts/streaming-dataflow

    [1]: https://jon.thesquareplanet.com/papers/osdi18-noria.pdf

    [2]: https://dl.acm.org/doi/10.1145/2517349.2522738

    [3]: https://github.com/TimelyDataflow/timely-dataflow

  • Mandala: experiment data management as a built-in (Python) language feature
    4 projects | /r/ProgrammingLanguages | 11 Apr 2023
    And systems like timely dataflow, https://github.com/TimelyDataflow/timely-dataflow
  • Arroyo: A distributed stream processing engine written in Rust
    3 projects | /r/rust | 4 Apr 2023
    Project looks cool! Glad you open sourced it. It could use some comments in the code base to help contributors ;). I also like the datafusion usage, that is awesome. BTW I work on github.com/bytewax/bytewax, which is based on https://github.com/TimelyDataflow/timely-dataflow another Rust dataflow computation engine.
  • Rust MPI -- Will there ever be a fully oxidized implementation?
    4 projects | /r/rust | 5 Mar 2023
    Just found this https://github.com/TimelyDataflow/timely-dataflow and my heart skipped a beat.
  • Streaming processing in Python using Timely Dataflow with Bytewax
    1 project | /r/Python | 9 Nov 2022
    Bytewax is a Python native binding to the Timely Dataflow library (written in Rust) for building highly scalable streaming (and batch) processing pipelines.
  • Alternative Kafka Integration Framework to Kafka Connect?
    3 projects | /r/apachekafka | 21 Jun 2022
    I am working on Bytewax, which is a Python stream processing framework built on Timely Dataflow. It is not exactly a Kafka integration framework because it is a more of a general stream processing framework, but might be interesting for you. We are focused on enabling people to more easily debug, containerize, parallelize and customize and less on enabling a declarative integration framework. It is still early days for us! And we are looking for feedback and ideas from the community.
  • [AskJS] JavaScript for data processing
    5 projects | /r/javascript | 27 May 2022
    We used to use a library called Pond.js, https://github.com/esnet/pond, but the reliance on Immutable.JS caused some performance pitfalls, so we wrote a system from scratch that deals with data in a batched streaming fashion. A lot of the concepts were borrowed from a Rust library called timely-dataflow, https://github.com/TimelyDataflow/timely-dataflow.
  • Dataflow: An Efficient Data Processing Library for Machine Learning
    2 projects | /r/rust | 17 Jan 2022
    Though the name "Dataflow" might be an unfortunate name conflict with another Rust project: https://github.com/TimelyDataflow/timely-dataflow
  • Ask HN: Is there a way to subscribe to an SQL query for changes?
    17 projects | news.ycombinator.com | 22 Apr 2021
    > In the simplest case, I'm talking about regular SQL non-materialized views which are essentially inlined.

    I see that now -- makes sense!

    > Wish we had some better database primitives to assemble rather than building everything on Postgres - its not ideal for a lot of things.

    I'm curious to hear more about this! We agree that better primitives are required and that's why Materialize is written in Rust using using TimelyDataflow[1] and DifferentialDataflow[2] (both developed by Materialize co-founder Frank McSherry). The only relationship between Materialize and Postgres is that we are wire-compatible with Postgres and we don't share any code with Postgres nor do we have a dependence on it.

    [1] https://github.com/TimelyDataflow/timely-dataflow

  • 7 Real-Time Data Streaming Tools You Should Consider On Your Next Project
    2 projects | dev.to | 20 Mar 2021
    Under the hood, Materialize uses Timely Dataflow (TDF) as the stream-processing engine. This allows Materialize to take advantage of the distributed data-parallel compute engine. The great thing about using TDF is that it has been in open source development since 2014 and has since been battle-tested in production at large Fortune 1000-scale companies.

diagnostics

Posts with mentions or reviews of diagnostics. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2021-01-22.
  • Why isn't differential dataflow more popular?
    13 projects | news.ycombinator.com | 22 Jan 2021
    I've been using DD in production usage for just over a year now for low latency(sub second from event IRL to pipeline CDC output) processing in a geo-distributed environment(100's of locations globally coordinating) some days at the TB per day level of event ingest.

    DD for me was one of the final attempts to find something, anything, that could handle the requirements I was working with, because Spark, Flink, and others just couldn't reasonably get close to what I was looking for. The closest 2nd place was Apache Flink.

    Over the last year I've read through the DD and TD codebases about 5-7 times fully. Even with that, I'm often in a position where I go back to my own applications to see how I had already solved a type of problem. I liken the project to taking someone use to NASCAR and dropping them into a Formula One vehicle. You've seen it work so much faster, and the tech and capabilities are clearly designed for so much more than you can make it do right now.

    A few learning examples that I consider funny:

    1. I had a graph that was on the order of about 1.2 trillion edges with about 90 million nodes. I was using serde derived structs for the edge and node structs(not simplified numerical types), which means I have to implement(or derive) a bunch of traits myself. I spent way more time than I'd like to admit trying to get .reduce() to work to remove 'surplus' edges that have already been processed from the graph to shrink the working dataset. Finally in frustration and reading through the DD codebase again, I 'rediscovered' .consolidate() which 'just worked' taking the 1.2 trillion edges down into the 300 million edges. For instance, some of the edge values I need to work with have histograms for the distributions, and some of the scoring of those histograms is custom. Not usually an issue, except having to figure out how to implement a bunch of the traits has been a significant hurdle.

    2. I get to constantly dance between DD's runtime and trying to ergonomically connect the application into the tonic gRPC and tokio interfaces. Luckily I've found a nice pattern where I create my inter-thread communication constructs, then start up 2 rust threads, and start tokio based interfaces in one, and DD runtime and workers in the other. On bigger servers(packet.net has some great gen3 instances) I usually pin tokio to 2-8 cores, and leave the rest of the cores to DD.

    3. Almost every new app I start, I run into the gotcha where I want to have a worker that runs only once 'globally' and it's usually the thread that I'd want to use to coordinate data ingestion. Super simple to just have a guard for if worker.index() == 0, but when deep in thought about an upcoming pipeline, it's often forgotten.

    4. For diagnostics, there is: https://github.com/TimelyDataflow/diagnostics which has provided much needed insights when things have gotten complex. Usually it's been 'just enough' to point into the right direction, but only once was the output able to point exactly to the issue I was running into.

    5. I have really high hopes for materialize.io That's really the type of system I'd want to use in 80% of the cases I'm using DD right now. I've been following them for about a year now, and the progress is incredible, but my use cases seem more likely to be supported in the 0.8->1.3 roadmap range.

    6. I've wanted to have a way to express 'use no more than 250GB of ram' and have some way to get a compile time feedback that a fixed dataset won't be able to process the pipeline with that much resources. It'd be far better if the system could adjust its internal runtime approach in order to stay within the limits.

What are some alternatives?

When comparing timely-dataflow and diagnostics you can also consider the following projects:

noria - Fast web applications through dynamic, partially-stateful dataflow

differential-datalog - DDlog is a programming language for incremental computation. It is well suited for writing programs that continuously update their output in response to input changes. A DDlog programmer does not write incremental algorithms; instead they specify the desired input-output mapping in a declarative manner.

rslint - A (WIP) Extremely fast JavaScript and TypeScript linter and Rust crate

materialize - The data warehouse for operational workloads.

sliding-window-aggregators - Reference implementations of sliding window aggregation algorithms

bytewax - Python Stream Processing

blog - Some notes on things I find interesting and important.

realtime - Broadcast, Presence, and Postgres Changes via WebSockets

lambdo - Feature engineering and machine learning: together at last!

differential-dataflow - An implementation of differential dataflow using timely dataflow on Rust.

ballista - Distributed compute platform implemented in Rust, and powered by Apache Arrow.