Apache Thrift VS Apache Orc

Compare Apache Thrift vs Apache Orc and see what are their differences.

Our great sponsors
  • Scout APM - Less time debugging, more time building
  • OPS - Build and Run Open Source Unikernels
  • SonarLint - Deliver Cleaner and Safer Code - Right in Your IDE of Choice!
Apache Thrift Apache Orc
6 2
8,865 484
1.0% 4.3%
8.9 9.4
5 days ago 5 days ago
C++ HTML
Apache License 2.0 Apache License 2.0
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

Apache Thrift

Posts with mentions or reviews of Apache Thrift. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2022-01-01.

Apache Orc

Posts with mentions or reviews of Apache Orc. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2021-07-27.
  • AWS EMR Cost Optimization Guide
    1 project | dev.to | 14 Dec 2021
    Data formatting is another place to make gains. When dealing with huge amounts of data, finding the data you need can take up a significant amount of your compute time. Apache Parquet and Apache ORC are columnar data formats optimized for analytics that pre-aggregate metadata about columns. If your EMR queries column intensive data like sum, max, or count, you can see significant speed improvements by reformatting data like CSVs into one of these columnar formats.
  • Apache Hudi - The Streaming Data Lake Platform
    8 projects | dev.to | 27 Jul 2021
    The following stack captures layers of software components that make up Hudi, with each layer depending on and drawing strength from the layer below. Typically, data lake users write data out once using an open file format like Apache Parquet/ORC stored on top of extremely scalable cloud storage or distributed file systems. Hudi provides a self-managing data plane to ingest, transform and manage this data, in a way that unlocks incremental data processing on them.

What are some alternatives?

When comparing Apache Thrift and Apache Orc you can also consider the following projects:

gRPC - The C based gRPC (C++, Python, Ruby, Objective-C, PHP, C#)

ZeroMQ - ZeroMQ core engine in C++, implements ZMTP/3.1

Cap'n Proto - Cap'n Proto serialization/RPC system - core tools and C++ library

Protobuf - Protocol Buffers - Google's data interchange format

Apache Avro - Apache Avro is a data serialization system.

Apache Parquet - Apache Parquet

nanomsg - nanomsg library

Big Queue - A big, fast and persistent queue based on memory mapped file.

rpclib - rpclib is a modern C++ msgpack-RPC server and client library

debezium - Change data capture for a variety of databases. Please log issues at https://issues.redhat.com/browse/DBZ.