dcs VS rum

Compare dcs vs rum and see what are their differences.

dcs

Debian Code Search (codesearch.debian.net) is a search engine that searches through all the 130 GB of open source software that is included in Debian. Supports regular expressions! (by Debian)

rum

RUM access method - inverted index with additional information in posting lists (by postgrespro)
InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
dcs rum
3 11
197 694
1.0% 0.9%
4.9 4.0
about 2 months ago 4 months ago
Go C
GNU General Public License v3.0 or later GNU General Public License v3.0 or later
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

dcs

Posts with mentions or reviews of dcs. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2024-04-10.
  • Code Search Is Hard
    13 projects | news.ycombinator.com | 10 Apr 2024
  • Sourcegraph is no longer Open Source
    26 projects | news.ycombinator.com | 4 Jul 2023
    What is a good open-source system for code search if I want to plug 100 or so git repos into it and have it available over the web? GH search is not desirable because it would search too broadly and would not cover repos on Gitlab etc.

    I looked at the Debian code search [1] in the past, but for some reason thought it required a bit too much effort and didn't complete my investigation of it. Though [2] looks pretty approachable.

    Sourcegraph mentioned Zoekt [3], but I am not sure how usable it is. If it was pretty good, why did Sourcegraph OSS exist?

    Finally, from all the discussion how Sourcegraph OSS was very behind in the past few years, I guess there is no serious plan to fork it?

    [1]: https://github.com/Debian/dcs

    [2]: https://github.com/Debian/dcs/blob/main/howto/building.md

    [3]: https://github.com/sourcegraph/zoekt

  • Building a custom code search index in Go for searchcode.com
    3 projects | news.ycombinator.com | 23 Nov 2022

rum

Posts with mentions or reviews of rum. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2024-04-10.
  • Code Search Is Hard
    13 projects | news.ycombinator.com | 10 Apr 2024
    the rum index has worked well for us on roughly 1TB of pdfs. written by postgrespro, same folks who wrote core text search and json indexing. not sure why rum not in core. we have no problems.

       https://github.com/postgrespro/rum
  • Is it worth using Postgres' builtin full-text search or should I go straight to Elastic?
    2 projects | /r/PostgreSQL | 25 Apr 2023
    If you need ranking, and you have the possibility to install PostgreSQL extensions, then you can consider an extension providing RUM indexes: https://github.com/postgrespro/rum. Otherwise, you'll have to use an "external" FTS engine like ElasticSearch.
  • Features I'd Like in PostgreSQL
    14 projects | news.ycombinator.com | 28 Jan 2023
    >Reduce the memory usage of prepared queries

    Yes query plan reuse like every other db, this still blows me away PG replans every time unless you explicitly prepare and that's still per connection.

    Better full-text scoring is one for me that's missing in that list, TF/IDF or BM25 please see: https://github.com/postgrespro/rum

  • Ask HN: Books about full text search
    3 projects | news.ycombinator.com | 24 Nov 2022
    for postgres, i highly recommend the rum index over the core fts. rum is written by postgrespro, who also wrote core fts and json indexing in pg.

        https://github.com/postgrespro/rum
  • Postgres Full Text Search vs. the Rest
    21 projects | news.ycombinator.com | 14 Oct 2022
    My experience with Postgres FTS (did a comparison with Elastic a couple years back), is that filtering works fine and is speedy enough, but ranking crumbles when the resulting set is large.

    If you have a large-ish data set with lots of similar data (4M addresses and location names was the test case), Postgres FTS just doesn't perform.

    There is no index that helps scoring results. You would have to install an extension like RUM index (https://github.com/postgrespro/rum) to improve this, which may or may not be an option (often not if you use managed databases).

    If you want a best of both worlds, one could investigate this extensions (again, often not an option for managed databases): https://github.com/matthewfranglen/postgres-elasticsearch-fd...

    Either way, writing something that indexes your postgres database into elastic/opensearch is a one time investment that usually pays off in the long run.

  • Postgres Full-Text Search: A Search Engine in a Database
    3 projects | news.ycombinator.com | 11 Jul 2022
    Mandatory mention of the RUM extension (https://github.com/postgrespro/rum) if this caught your eye. Lots of tutorials and conference presentations out there showcasing the advantages in terms of ranking, timestamps...
    10 projects | news.ycombinator.com | 27 Jul 2021
    You might be just fine adding an unindexed tsvector column, since you've already filtered down the results.

    The GIN indexes for FTS don't really work in conjunction with other indices, which is why https://github.com/postgrespro/rum exists. Luckily, it sounds like you can use your existing indices to filter and let postgres scan for matches on the tsvector.

  • Postgrespro/rum: RUM access method – inverted index with additional information
    1 project | news.ycombinator.com | 17 Dec 2021
  • Debugging random slow writes in PostgreSQL
    1 project | news.ycombinator.com | 15 May 2021
    We have been bitten by the same behavior. I gave a talk with a friend about this exact topic (diagnosing GIN pending list updates) at PGCon 2019 in Ottawa[1][2].

    What you need to know is that the pending list will be merged with the main b-tree during several operations. Only one of them is so extremely critical for your insert performance - that is during actual insert. Both vacuum and autovacuum (including autovacuum analyze but not direct analyze) will merge the pending list. So frequent autovacuums are the first thing you should tune. Merging on insert happens when you exceed the gin_pending_list_limit. In all cases it is also interesting to know which memory parameter is used to rebuild the index as that inpacts how long it will take: work_mem (when triggered on insert), autovacuum_work_mem (when triggered during autovauum) and maintainance_work_mem (triggered by a call to gin_clean_pending_list()) define how much memory can be used for the rebuild.

    What you can do is:

    - tune the size of the pending list (like you did)

    - make sure vacuum runs frequently

    - if you have a bulk insert heavy workload (ie. nightly imports), drop the index and create it after inserting rows (not always makes sense business wise, depends on your app)

    - disable fastupdate, you pay a higher cost per insert but remove the fluctuctuation when the merge needs to happen

    The first thing was done in the article. However I believe the author still relies on the list being merged on insert. If vacuums were tuned agressively along with the limit (vacuums can be tuned per table). Then the list would be merged out of bound of ongoing inserts.

    I also had the pleasure of speaking with one main authors of GIN indexes (Oleg Bartunov) during the mentioned PGCon. He gave probably the best solution and informed me to "just use RUM indexes". RUM[3] indexes are like GIN indexes, without the pending list and with faster ranking, faster phrase searches and faster timestamp based ordering. It is however out of the main postgresql release so it might be hard to get it running if you don't control the extensions that are loaded to your Postgres instance.

    [1] - wideo https://www.youtube.com/watch?v=Brt41xnMZqo&t=1s

    [2] - slides https://www.pgcon.org/2019/schedule/attachments/541_Let's%20...

    [3] - https://github.com/postgrespro/rum

  • Show HN: Full text search Project Gutenberg (60m paragraphs)
    5 projects | news.ycombinator.com | 24 Jan 2021
    I suggest to have a look at https://github.com/postgrespro/rum if you haven’t yet. It solves the issue of slow ranking in PostgreSQL FTS.

What are some alternatives?

When comparing dcs and rum you can also consider the following projects:

dcs-lua-datamine - A reference guide to the lua table values in DCS for weapons and aircraft

postgres-elasticsearch-fdw - Postgres to Elastic Search Foreign Data Wrapper

searchcode-server - The offical home of searchcode-server where you can run searchcode locally. Note that master is generally unstable in the sense that it is not a release. Check releases for release versions https://github.com/boyter/searchcode-server/releases

recoll - recoll with webui in a docker container

dcs_liberation - DCS World dynamic campaign.

zombodb - Making Postgres and Elasticsearch work together like it's 2023

git-peek - git repo to local editor instantly

pgvector - Open-source vector similarity search for Postgres

peek - 1-click from git repo to local editor

pg_search - pg_search builds ActiveRecord named scopes that take advantage of PostgreSQL’s full text search

cs - command line codespelunker or code search

pg_cjk_parser - Postgres CJK Parser pg_cjk_parser is a fts (full text search) parser derived from the default parser in PostgreSQL 11. When a postgres database uses utf-8 encoding, this parser supports all the features of the default parser while splitting CJK (Chinese, Japanese, Korean) characters into 2-gram tokens. If the database's encoding is not utf-8, the parser behaves just like the default parser.