Our great sponsors
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
The second one is to encode data in such a way that you can read it and operate on it directly from the buffer. You write data in a layout that is the same, or easily transformed as types in memory. To do that you usually need to encode with a known schema, only Sized types to efficiently compute fields locations as offsets in the buffer, and you usually represent pointers as offset into the encode. You can look at capnproto protocol for instance https://capnproto.org/
I'm about to release my dictionary project in a little while, and it uses zero copy deserialization for its dictionary files. You can map the file directly into memory and with the "right" kind of data structure you don't have to perform additional work or buffering to do dictionary lookups. I blogged about it in more in detail here.