ClickBench VS pg_partman

Compare ClickBench vs pg_partman and see what are their differences.

InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
ClickBench pg_partman
71 7
571 1,879
3.0% 2.4%
9.0 7.0
2 days ago 21 days ago
HTML PLpgSQL
GNU General Public License v3.0 or later GNU General Public License v3.0 or later
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

ClickBench

Posts with mentions or reviews of ClickBench. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2024-05-02.
  • Umbra: A Disk-Based System with In-Memory Performance [pdf]
    3 projects | news.ycombinator.com | 2 May 2024
    Benchmarks: https://benchmark.clickhouse.com

    So definitely compared against PostgreSQL, MariaDB it is significantly faster.

    On par with lower-end Snowflake.

  • Loading a trillion rows of weather data into TimescaleDB
    8 projects | news.ycombinator.com | 16 Apr 2024
    TimescaleDB primarily serves operational use cases: Developers building products on top of live data, where you are regularly streaming in fresh data, and you often know what many queries look like a priori, because those are powering your live APIs, dashboards, and product experience.

    That's different from a data warehouse or many traditional "OLAP" use cases, where you might dump a big dataset statically, and then people will occasionally do ad-hoc queries against it. This is the big weather dataset file sitting on your desktop that you occasionally query while on holidays.

    So it's less about "can you store weather data", but what does that use case look like? How are the queries shaped? Are you saving a single dataset for ad-hoc queries across the entire dataset, or continuously streaming in new data, and aging out or de-prioritizing old data?

    In most of the products we serve, customers are often interested in recent data in a very granular format ("shallow and wide"), or longer historical queries along a well defined axis ("deep and narrow").

    For example, this is where the benefits of TimescaleDB's segmented columnar compression emerges. It optimizes for those queries which are very common in your application, e.g., an IoT application that groups by or selected by deviceID, crypto/fintech analysis based on the ticker symbol, product analytics based on tenantID, etc.

    If you look at Clickbench, what most of the queries say are: Scan ALL the data in your database, and GROUP BY one of the 100 columns in the web analytics logs.

    - https://github.com/ClickHouse/ClickBench/blob/main/clickhous...

    There are almost no time-predicates in the benchmark that Clickhouse created, but perhaps that is not surprising given it was designed for ad-hoc weblog analytics at Yandex.

    So yes, Timescale serves many products today that use weather data, but has made different choices than Clickhouse (or things like DuckDB, pg_analytics, etc) to serve those more operational use cases.

  • Variant in Apache Doris 2.1.0: a new data type 8 times faster than JSON for semi-structured data analysis
    2 projects | dev.to | 27 Mar 2024
    We tested with 43 Clickbench SQL queries. Queries on the Variant columns are about 10% slower than those on pre-defined static columns, and 8 times faster than those on JSON columns. (For I/O reasons, most cold runs on JSONB data failed with OOM.)
  • Fair Benchmarking Considered Difficult (2018) [pdf]
    2 projects | news.ycombinator.com | 10 Mar 2024
    I have a project dedicated to this topic: https://github.com/ClickHouse/ClickBench

    It is important to explain the limitations of a benchmark, provide a methodology, and make it reproducible. It also has to be simple enough, otherwise it will not be realistic to include a large number of participants.

    I'm also collecting all database benchmarks I could find: https://github.com/ClickHouse/ClickHouse/issues/22398

  • ClickBench – A Benchmark for Analytical DBMS
    1 project | news.ycombinator.com | 8 Feb 2024
  • FLaNK Stack 05 Feb 2024
    49 projects | dev.to | 5 Feb 2024
  • Why Postgres RDS didn't work for us
    4 projects | news.ycombinator.com | 3 Feb 2024
    Indeed, ClickHouse results were run on an older instance type of the same family and size (c5.4xlarge for ClickHouse and c6a.4xlarge for Timescale), so if anything ClickHouse results are at a slight disadvantage.

    This is an open source benchmark - we'd love contributions from Timescale enthusiasts if we missed something: https://github.com/ClickHouse/ClickBench/

  • Show HN: Stanchion – Column-oriented tables in SQLite
    3 projects | news.ycombinator.com | 31 Jan 2024
    Interesting project! Thank you for open sourcing and sharing. Agree that local and embedded analytics are an increasing trend, I see it too.

    A couple of questions:

    * I’m curious what the difficulties were in the implementation. I suspect it is quite a challenge to implement this support in the current SQLite architecture, and would curious to know which parts were tricky and any design trade-off you were faced with.

    * Aside from ease-of-use (install extension, no need for a separate analytical database system), I wonder if there are additional benefits users can anticipate resulting from a single system architecture vs running an embedded OLAP store like DuckDB or clickhouse-local / chdb side-by-side with SQLite? Do you anticipate performance or resource efficiency gains, for instance?

    * I am also curious, what the main difficulty with bringing in a separate analytical database is, assuming it natively integrates with SQLite. I may be biased, but I doubt anything can approach the performance of native column-oriented systems, so I'm curious what the tipping point might be for using this extension vs using an embedded OLAP store in practice.

    Btw, would love for you or someone in the community to benchmark Stanchion in ClickBench and submit results! (https://github.com/ClickHouse/ClickBench/)

    Disclaimer: I work on ClickHouse.

  • ClickBench: A Benchmark for Analytical Databases
    1 project | news.ycombinator.com | 22 Jan 2024
  • DuckDB performance improvements with the latest release
    8 projects | news.ycombinator.com | 6 Nov 2023

pg_partman

Posts with mentions or reviews of pg_partman. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2022-09-22.
  • Dear data engineers
    1 project | /r/dataengineering | 6 Dec 2023
    Assuming these are the types of insights you're looking for, you'll probably look for a way to aggregate data points across/within geographies. postgis is an open source extension for postgres that can help you with this, but there's also quite a few python tools that can help you explore the data, such as geopandas, folium, geoplot. Depending on volume, you might want to partition the data for query performance, and there's another extension pg_partman that can help with that. Just noticed some other posts have recommended something similar.
  • Pgpartman: Partition Management Extension for Postgres
    1 project | news.ycombinator.com | 20 Sep 2023
  • Which is the best way to automate backing up data monthly of tables in a schema and then deleting them?
    1 project | /r/PostgreSQL | 6 Oct 2022
    pg_partman is in RDS >12.5 (here). Pg_partman makes managing data retention relatively easy.
  • Partitioning in Postgres, 2022 Edition
    1 project | news.ycombinator.com | 5 Oct 2022
  • TimescaleDB 2.7 vs. PostgreSQL 14
    4 projects | news.ycombinator.com | 22 Sep 2022
    Whenever I see these posts from TimescaleDB, I always want to ask them how it compares in performance to alternative extensions that implement the same features, rather than just comparing TimescaleDB to vanilla PostgreSQL.

    For example, they mention their automated data retention and how it's achieved with one SQL command, and how DELETEing records is a very costly operation, and how "even if you were using Postgres declarative partitioning you’d still need to automate the process yourself, wasting precious developer time, adding additional requirements, and implementing bespoke code that needs to be supported moving forward".

    There's zero mention anywhere of pg_partman, which does all of these things for you equally as simply, and is a fully OSS free alternative [0].

    I get that it's a PG extension that competes with their product. I know that TimescaleDB does a few other things that pg_partman does not. But I can't help but find its (seemingly) purposeful omission in these, otherwise very thorough blog posts, misleading.

    [0] https://github.com/pgpartman/pg_partman/blob/master/doc/pg_p...

  • Table partitioning by months of the year?
    1 project | /r/PostgreSQL | 23 Aug 2021
    Take a look into this extension which would take care of a good amount of automation for you.
  • Replicating a dynamically partitioned table possible in Postgres 13
    1 project | /r/PostgreSQL | 8 Jan 2021
    You might want to look into pg_partman which has many useful tools around semi-automatic partitioning. According to their documentation they already have a procedure that will do exactly that: create new partitions based on the rows in the default partition.

What are some alternatives?

When comparing ClickBench and pg_partman you can also consider the following projects:

starrocks - StarRocks, a Linux Foundation project, is a next-generation sub-second MPP OLAP database for full analytics scenarios, including multi-dimensional analytics, real-time analytics, and ad-hoc queries. InfoWorld’s 2023 BOSSIE Award for best open source software.

periods - PERIODs and SYSTEM VERSIONING for PostgreSQL

duckdb - DuckDB is an in-process SQL OLAP Database Management System

pgddl - DDL eXtractor functions for PostgreSQL (ddlx)

ClickHouse - ClickHouse® is a free analytics DBMS for big data

blog - OpenSource,Database,Business,Minds. git clone --depth 1 https://github.com/digoal/blog

hosts - 🔒 Consolidating and extending hosts files from several well-curated sources. Optionally pick extensions for porn, social media, and other categories.

postgres-aws-s3 - aws_s3 postgres extension to import/export data from/to s3 (compatible with aws_s3 extension on AWS RDS)

TablePlus - TablePlus macOS issue tracker

practical-sql - Code and Data for the First Edition of "Practical SQL" by Anthony DeBarros, published by No Starch Press (2018).

clickhouse-bulk - Collects many small inserts to ClickHouse and send in big inserts

metagration - Metagration: PostgreSQL Migrator in PostgreSQL