Our great sponsors
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
I use Restic [0] for my personal backups and I use Backblaze for the backend, but AWS S3 and anything compatible (of which Backblaze is too) is also an option. I preencrypt all my data and use pass for managing my encryption password and the secrets.
[0] https://restic.net/
Take a look at aifiles: https://github.com/jjuliano/aifiles
Getting vicuna or alpaca for this could be the best decision for those that want to keep their data.
Could you imagine the space saving you can achieve by a system that constructs a real normalized duckdb database with zstd compression and join tables and all from your big dump of tar.xml.gz files? Automagically converting all of your media to AV1 and Opus to save space and remove any private codec reqs?
Regarding Arq backup: if you are worried about using a proprietary (enrypted) and closed-source backup format in case the company were to go under, they have an open source command-line restore tool:
https://github.com/arqbackup/arq_restore
I've been using Arq for years, but I need to look into the "Glacier Deep Archive" format which is about 1/20th the cost of the fastest storage class.
You can remove folders from spotlight indexing. I do it for all the source code folders.
I’ve also built a small tool to check whether some git repo is in sync with upstream, so it could be deleted: https://gitlab.com/leipert-projects/git-recon