Ingest, store, & analyze all types of time series data in a fully-managed, purpose-built database. Keep data forever with low-cost storage and superior data compression. Learn more →
Apache Flink Alternatives
Similar projects and alternatives to Apache Flink
Apache Spark - A unified analytics engine for large-scale data processing
Suite of tools for deploying and training deep learning models using the JVM. Highlights include model import for keras, tensorflow, and onnx/pytorch, a modular and tiny c++ library for running math code and a java based math library on top of the core c++ library. Also includes samediff: a pytorch/tensorflow like library for running deep learning using automatic differentiation.
Access the most powerful time series database as a service. Ingest, store, & analyze all types of time series data in a fully-managed, purpose-built database. Keep data forever with low-cost storage and superior data compression.
Official repository of Trino, the distributed SQL query engine for big data, formerly known as PrestoSQL (https://trino.io)
Sparkling Water provides H2O functionality inside Spark cluster
Mirror of Apache Kafka
A Scala API for Apache Beam and Google Cloud Dataflow.
Lightweight real-time big data streaming engine over Akka
Write Clean Java Code. Always.. Sonar helps you commit clean code every time. With over 600 unique rules to find Java bugs, code smells & vulnerabilities, Sonar finds the issues while you focus on the work.
Build highly concurrent, distributed, and resilient message-driven applications on the JVM
Deep Java Library (DJL)
An Engine-Agnostic Deep Learning Framework in Java
The Streaming-first HTTP server/module of Akka
Statistical Machine Intelligence & Learning Engine
ClickHouse® is a free analytics DBMS for big data
CockroachDB - the open source, cloud-native distributed SQL database.
The MongoDB Database
📊 Cube — The Semantic Layer for Building Data Applications
An open-source time-series SQL database optimized for fast ingest and complex queries. Packaged as a PostgreSQL extension.
Apache Druid: a high performance real-time analytics database.
An open source time-series database for fast ingest and SQL queries
Firebase Quickstart Samples for Android
SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives
Apache Flink reviews and mentions
We Are Changing the License for Akka
6 projects | news.ycombinator.com | 7 Sep 2022
DeWitt Clause, or Can You Benchmark %DATABASE% and Get Away With It
21 projects | dev.to | 2 Jun 2022
Apache Drill, Druid, Flink, Hive, Kafka, Spark
Computation reuse via fusion in Amazon Athena
2 projects | news.ycombinator.com | 20 May 2022
It took me some time to get a good grasp of the power of SQL; and it really kicked in when I learned about optimization rules. It's a program that you rewrite, just like an optimizing compiler would.
You state what you want; you have different ways to fetch and match and massage data; and you can search through this space to produce a physical plan. Hopefully you used knowledge to weight parts to be optimized (table statistics, like Java's JIT would detect hot spots).
I find it fascinating to peer through database code to see what is going on. Lately, there's been new advances towards streaming databases, which bring a whole new design space. For example, now you have latency of individual new rows to optimize for, as opposed to batch it whole to optimize the latency of a dataset. Batch scanning will be benefit from better use of your CPU caches.
And maybe you could have a hybrid system which reads history from a log and aggregates in a batched manner, and then switches to another execution plan when it reaches the end of the log.
If you want to have a peek at that here are Flink's set of rules , generic and stream-specific ones. The names can be cryptic, but usually give a good sense of what is going on. For example: PushFilterIntoTableSourceScanRule makes the WHERE clause apply the earliest possible, to save some CPU/network bandwidth further down. PushPartitionIntoTableSourceScanRule tries to make a fan-out/shuffle happen the earliest possible, so that parallelism can be made use of.
Avro SpecificRecord File Sink using apache flink is not compiling due to error incompatible types: FileSink<?> cannot be converted to SinkFunction<?>
3 projects | reddit.com/r/apacheflink | 14 Sep 2021
: https://mvnrepository.com/artifact/org.apache.avro/avro-maven-plugin/1.8.2 : https://github.com/apache/flink/blob/master/flink-connectors/flink-connector-files/src/main/java/org/apache/flink/connector/file/sink/FileSink.java : https://ci.apache.org/projects/flink/flink-docs-master/docs/connectors/datastream/file_sink/ : https://github.com/apache/flink/blob/c81b831d5fe08d328251d91f4f255b1508a9feb4/flink-end-to-end-tests/flink-file-sink-test/src/main/java/FileSinkProgram.java : https://github.com/rajcspsg/streaming-file-sink-demo
A note from our sponsor - InfluxDB
www.influxdata.com | 23 Mar 2023
apache/flink is an open source project licensed under Apache License 2.0 which is an OSI approved license.