jvm-serializers
Apache Avro
Our great sponsors
jvm-serializers | Apache Avro | |
---|---|---|
7 | 22 | |
3,275 | 2,764 | |
- | 1.7% | |
4.4 | 9.7 | |
7 months ago | 5 days ago | |
Java | Java | |
- | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
jvm-serializers
-
Fury: 170x faster than JDK, fast serialization powered by JIT and Zero-copy
Compared with protobuf, fury is 3.2x faster. When comparing with avro, fury is 5.3x faster. Compared with flatbuffers, fury is 4.8x faster. See https://github.com/eishay/jvm-serializers/wiki for detailed benchmark data
-
The state of Java Object Serialization libraries in Q2 2023
First, there's benchmarks here if you haven't seen it: jvm-serializers. Not terribly scientific, but it's something. To make any decision, you really need to benchmark your own object graph and it's important to configure the serializer for your particular usage. Still, it is sort of useful for comparing frameworks. It would be interesting to see how Loial performs there. Ping me if you add it.
-
Up to 100x Faster FastAPI with simdjson and io_uring on Linux 5.19+
It depends. Some binary encodings such as flatbuffer are actually slower than some JSON libraries. There's a wide range of performance even in the JSON libraries themselves. Generally the faster JSON libraries are the ones that work on a predefined schema and so are able to generate code specifically for that JSON.
-
Go standard library: structured, leveled logging
> I'm surprised this is up for debate.
I looked into logging in protobuf when I was seeing if there was a better binary encoding for ring-buffer logging, along the same lines as nanolog:
https://tersesystems.com/blog/2020/11/26/queryable-logging-w...
What I found was that it's typically not the binary encoding vs string encoding that makes a difference. The biggest factors are "is there a predefined schema", "is there a precompiler that will generate code for this schema", and "what is the complexity of the output format". With that in mind, if you are dealing with chaotic semi-structured data, JSON is pretty good, and actually faster than some binary encodings:
https://github.com/eishay/jvm-serializers/wiki/Newer-Results...
-
Scala 3.0 serialization
You could use any of the JVM serialisers which should still work.
Apache Avro
-
Open Table Formats Such as Apache Iceberg Are Inevitable for Analytical Data
Apache AVRO [1] is one but it has been largely replaced by Parquet [2] which is a hybrid row/columnar format
[1] https://avro.apache.org/
-
Generating Avro Schemas from Go types
The most common format for describing schema in this scenario is Apache Avro.
-
How do you update an existing avro schema using apache avro SchemaBuilder?
I am testing a new schema registry which loads and retrieves different kinds of avro schemas. In the process of testing, I need to create a bunch of different types of avro schemas. As it involves a lot of permutations, I decided to create the schema programmatically.I am using the apache avro SchemaBuilder to do so.
- The state of Apache Avro in Rust
- How people generate examples for multiple programming languages?
-
gRPC on the client side
Other serialization alternatives have a schema validation option: e.g., Avro, Kryo and Protocol Buffers. Interestingly enough, gRPC uses Protobuf to offer RPC across distributed components:
-
Understanding Azure Event Hubs Capture
Apache Avro is a data serialization system, for more information visit Apache Avro
-
tl;dr of Data Contracts
Once things like JSON became more popular Apache Avro appeared. You can define Avro files which can then be generated into Python, Java C, Ruby, etc.. classes.
-
In One Minute : Hadoop
Avro, a data serialization system based on JSON schemas.
-
Events: Fat or Thin?
Supporting multiple versions of an event schema is a solved problem. Apache Avro with a published schema hash in a message header is one solution.
https://avro.apache.org/
What are some alternatives?
fury-benchmarks - Serialization Benchmarks for fury with other libraries
Protobuf - Protocol Buffers - Google's data interchange format
zio-json - Fast, secure JSON library with tight ZIO integration.
SBE - Simple Binary Encoding (SBE) - High Performance Message Codec
opentelemetry-specificatio
Apache Thrift - Apache Thrift
janino - Janino is a super-small, super-fast Java™ compiler.
iceberg - Apache Iceberg
grpc-dotnet - gRPC for .NET
Apache Parquet - Apache Parquet
honeycomb-opentelemetry-go - Honeycomb's OpenTelemetry Go SDK distribution
gRPC - The C based gRPC (C++, Python, Ruby, Objective-C, PHP, C#)