Magnetico
dolt
Magnetico | dolt | |
---|---|---|
14 | 93 | |
2,665 | 16,993 | |
- | 1.9% | |
1.7 | 10.0 | |
over 2 years ago | 5 days ago | |
Go | Go | |
GNU Affero General Public License v3.0 | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Magnetico
-
Bitmagnet: A self-hosted BitTorrent indexer, DHT crawler, and torrent search
This is really neat. I'll need to check it out. A couple years ago I ran my own instance of Magnetico (https://github.com/boramalper/magnetico), but this project looks a lot more polished.
-
DHT crawler
This is likely your best bet, but I'd recommend you use postgres over sqlite, this is however where development died so some features like search or stats aren't implemented while using pg as the db.
-
Why isn't distributed/decentralized archiving currently used?
You can create decentralized bittorrent indexers though (see https://github.com/boramalper/magnetico as an example)! This means you can search for bittorrent files without having to rely on a centralized service (although building the index does require some time & storage space of course).
-
Ask HN: I've Built a DHT Torrent Sniffer and Search Engine. Should I Release?
https://github.com/boramalper/magnetico
Someone else did this a while back, universe continues to exist.
-
Popcorn Time Is Back
I just get a rickroll gif on your link. Did you mean this project?
"Autonomous (self-hosted) BitTorrent DHT search engine suite."
https://github.com/boramalper/magnetico
-
Self Hosted Open Source Torrent Scraper!
Although I admit, that I would love to see a maintained DHT search solution (like magnetico or dhtcrawler2).
-
Magnetico / Docker question
Yeah, you probably found it. Abandoned, yes, but not necessarily unfinished.
- What is the best local DHT search engine?
- What does a 'good' GitHub page like? (Q for the Professionals)
-
I2P applications wish list?
A magnet search is easy enough to self-host, I can provide you with instructions to self-host a clearnet magnet search engine over I2P in the meantime, but you can probably figure it out from this: https://github.com/boramalper/magnetico which is an easy to self-host DHT search engine.
dolt
-
A MySQL compatible database engine written in pure Go
Hi, this is my project :)
For us this package is most important as the query engine that powers Dolt:
https://github.com/dolthub/dolt
We aren't the original authors but have contributed the vast majority of its code at this point. Here's the origin story if you're interested:
https://www.dolthub.com/blog/2020-05-04-adopting-go-mysql-se...
-
The Great Migration from MongoDB to PostgreSQL
It's a pretty good default stance, yeah.
We have been trying to convince people to use our new database [1] for several years and it's an uphill battle, because Postgres really is the best choice for most people. They really have to need our unique feature (version control) to even consider it over Postgres, and I don't blame them.
[1] https://github.com/dolthub/dolt
-
What I Talk About When I Talk About Query Optimizer (Part 1): IR Design
We implemented a query optimizer with a flexible intermediate representation in pure Go:
https://github.com/dolthub/go-mysql-server
Getting the IR correct so that it's both easy to use and flexible enough to be useful is a really interesting design challenge. Our primary abstraction in the query plan is called a Node, and is way more general than the IR type described in the article from OP. This has probably hurt us: we only recently separated the responsibility to fetch rows into its own part of the runtime, out of the IR -- originally row fetching was coupled to the Node type directly.
This is also the query engine that Dolt uses:
https://github.com/dolthub/dolt
But it has a plug-in architecture, so you can use the engine on any data source that implements a handful of Go interface.
- Dolt – Git for Data
- Dolt: A version-controlled SQL database
-
Show HN: DoltgreSQL – Version-Controlled Database, Like Git and PostgreSQL
Just want to point out that we're announcing development on the project. It's absolutely not ready for mainstream use yet! We have Dolt (https://github.com/dolthub/dolt) which is production-ready and widely in use, but it uses MySQL's syntax and wire protocol. We are building the Dolt equivalent for PostgreSQL, which is DoltgreSQL, but it's only pre-alpha.
-
Pg_branch: Pre-alpha Postgres extension brings Neon-like branching
Interesting that branching is now better supported and almost free. I wonder if merging can be simplified or whether it already is as simple and as fast as it can be?
I guess I am inspired by Dolt’s ability to branch and merge: https://github.com/dolthub/dolt
-
SQLedge: Replicate Postgres to SQLite on the Edge
#. SQLite WAL mode
From https://www.sqlite.org/isolation.html https://news.ycombinator.com/item?id=32247085 :
> [sqlite] WAL mode permits simultaneous readers and writers. It can do this because changes do not overwrite the original database file, but rather go into the separate write-ahead log file. That means that readers can continue to read the old, original, unaltered content from the original database file at the same time that the writer is appending to the write-ahead log
#. superfly/litefs: aFUSE-based file system for replicating SQLite https://github.com/superfly/litefs
#. sqldiff: https://www.sqlite.org/sqldiff.html https://news.ycombinator.com/item?id=31265005
#. dolthub/dolt: https://github.com/dolthub/dolt
> Dolt can be set up as a replica of your existing MySQL or MariaDB database using standard MySQL binlog replication. Every write becomes a Dolt commit. This is a great way to get the version control benefits of Dolt and keep an existing MySQL or MariaDB database.
#. pganalyze/libpg_query: https://github.com/pganalyze/libpg_query :
> C library for accessing the PostgreSQL parser outside of the server environment
#. Ibis + Substrait [ + DuckDB ]
> ibis strives to provide a consistent interface for interacting with a multitude of different analytical execution engines, most of which (but not all) speak some dialect of SQL.
> Today, Ibis accomplishes this with a lot of help from `sqlalchemy` and `sqlglot` to handle differences in dialect, or we interact directly with available Python bindings (for instance with the pandas, datafusion, and polars backends).
> [...] `Substrait` is a new cross-language serialization format for communicating (among other things) query plans. It's still in its early days, but there is already nascent support for Substrait in Apache Arrow, DuckDB, and Velox.
#. benbjohnson/postlite: https://github.com/benbjohnson/postlite
> postlite is a network proxy to allow access to remote SQLite databases over the Postgres wire protocol. This allows GUI tools to be used on remote SQLite databases which can make administration easier.
> The proxy works by translating Postgres frontend wire messages into SQLite transactions and converting results back into Postgres response wire messages. Many Postgres clients also inspect the pg_catalog to determine system information so Postlite mirrors this catalog by using an attached in-memory database with virtual tables. The proxy also performs minor rewriting on these system queries to convert them to usable SQLite syntax.
> Note: This software is in alpha. Please report bugs. Postlite doesn't alter your database unless you issue INSERT, UPDATE, DELETE commands so it's probably safe. If anything, the Postlite process may die but it shouldn't affect your database.
#. > "Hosting SQLite Databases on GitHub Pages" (2021) re: sql.js-httpvfs, DuckDB https://news.ycombinator.com/item?id=28021766
#. awesome-db-tools https://github.com/mgramin/awesome-db-tools
- How do you sync dev databases across multiple devices?
-
Ask HN: Data Management for AI Training
If you are just looking for data versioning there is Dolt:
https://github.com/dolthub/dolt
And that has a user-friendly UI in DoltHub:
https://www.dolthub.com/
You wouldn't store the images themselves in Dolt, those would likely be links to S3 but al the labels and surrounding metadata could be stored in Dolt?
DISCLAIMER: I'm the CEO of DoltHub so this is self-promotion.
What are some alternatives?
torrent-paradise - Decentralized DHT search site for IPFS
liquibase - Main Liquibase Source
dhtcrawler2 - dhtcrawler is a DHT crawler written in erlang. It can join a DHT network and crawl many P2P torrents. The program save all torrent info into database and provide an http interface to search a torrent by a keyword
absurd-sql - sqlite3 in ur indexeddb (hopefully a better backend soon)
torrentinim - A very low memory-footprint, self hosted API-only torrent search engine. Sonarr + Radarr Compatible, native support for Linux, Mac and Windows.
noms - The versioned, forkable, syncable database
cloud-torrent - ☁️ Cloud Torrent: a self-hosted remote torrent client
TimescaleDB - An open-source time-series SQL database optimized for fast ingest and complex queries. Packaged as a PostgreSQL extension.
peerflix-server - Streaming torrent client for Node.js with web ui.
vitess - Vitess is a database clustering system for horizontal scaling of MySQL.
bittorrent-tracker - 🌊 Simple, robust, BitTorrent tracker (client & server) implementation
temporal_tables - Temporal Tables PostgreSQL Extension