ParlAI
lrzip
Our great sponsors
ParlAI | lrzip | |
---|---|---|
18 | 6 | |
10,366 | 582 | |
- | - | |
5.6 | 2.6 | |
6 months ago | 10 months ago | |
Python | C | |
MIT License | GNU General Public License v3.0 only |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
ParlAI
-
Why does flake8 require me to copyright facebook and license under MIT?
Do you have https://github.com/facebookresearch/ParlAI installed? Looks like they're doing something weird with their flake8 config.
-
[D] Inner workings of the chatgpt memory
I would suspect similar to blenderbot2 from meta and parl.ai.
-
[D] We're the Meta AI research team behind CICERO, the first AI agent to achieve human-level performance in the game Diplomacy. We’ll be answering your questions on December 8th starting at 10am PT. Ask us anything!
There's quite a few open-source Reinforcement Learning challenges that you can explore with modest amounts of compute in order to build some experience training RL models, for example the Nethack Learning Environment, Atari, Minigrid, etc. For me personally, I had only worked in NLP / dialogue for years but got into RL by implementing Random Network Distillation models for NetHack. It's a fun area that definitely has its own unique challenges vs other domains. -AM
- How to Get Your Backup to Half of Its Size – ZSTD Support in XtraBackup
- re there any places you can download code for a ai chat bot and run on your own system?
-
Tarot Readings for Robots and Tangents
I am intrigued by the model because it develops long term memory that it can access for future conversations which you can see in more detail on the model cards.
-
Meta AI Introduces BlenderBot 3: A 175B Parameter, Publicly Available Chatbot That Improves Its Skills And Safety Over Time
Continue reading | Check out the paper, project, github link and reference article.
-
BlenderBot 3: A 175B parameter, publicly available chatbot
I have tried to use parl.ai in the past. I actually wanted to play with blenderbot 1.0. I kinda hate this library because it isn't exactly quick and easy to learn. I ended up using the Huggingface version.
You probably meant to link this: https://github.com/facebookresearch/ParlAI/blob/main/project...
- BlenderBot 3: A 175B-parameter, publicly available chatbot that improves its skills & safety over time
lrzip
-
How to Get Your Backup to Half of Its Size – ZSTD Support in XtraBackup
lrzip
Long Range ZIP or LZMA RZIP
https://github.com/ckolivas/lrzip
"A compression utility that excels at compressing large files (usually > 10-50 MB). Larger files and/or more free RAM means that the utility will be able to more effectively compress your files (ie: faster / smaller size), especially if the filesize(s) exceed 100 MB. You can either choose to optimise for speed (fast compression / decompression) or size, but not both."
-
File compression
7zip and XZ are almost always the best in any comparison. (They use the same algorithm.) Occasionally something new comes allong that may be bettyer, but it fades away... Like lrzip. https://lkml.org/lkml/2011/6/4/23 https://github.com/ckolivas/lrzip
-
If we found a way to reverse a hashing function, would that make them ultra-compression algorithms?
For example lrzip has an intense "dupe hunting" mode and takes days for large content, but does compress very well once it's done (and expansion is fast). I use it on long term storage backups and disk images and junk. Completely incompatible with streaming, unlike chunk-based like gzip or deflate or etc, although unpacking can stream such as searching or verifying a tarfile archive. But the original source has to be file-based so seeking for the hunting can work across the entire file-as-a-block.
- Lrzip – Long Range Zip or LZMA RZIP
-
Ask HN: How would you store 10PB of data for your startup today?
Best I know of for that is something like lrzip still, but even then it's probably not state of the art. https://github.com/ckolivas/lrzip
It'll also take a hell of a long time to do the compression and decompression. It'd probably be better to do some kind of chunking and deduplication instead of compression itself simply because I don't think you're ever going to have enough ram to store any kind of dictionary that would effectively handle so much data. You'd also not want to have to re-read and reconstruct that dictionary to get at some random image too.
-
Encrypted Backup Shootout
There's also lrzip for large files: https://github.com/ckolivas/lrzip
What are some alternatives?
algoneer - The Algoneer Python library.
bupstash - Easy and efficient encrypted backups.
flake8-copyright - Adds copyright checks to flake8
rdedup - Data deduplication engine, supporting optional compression and public key encryption.
mypy - Optional static typing for Python
duplicity - mirror of duplicity: https://code.launchpad.net/duplicity
webDiplomacy - Play Diplomacy online
LeoFS - The LeoFS Storage System
nle - The NetHack Learning Environment
BorgBackup - Deduplicating archiver with compression and authenticated encryption.
pyre-check - Performant type-checking for python.
nfreezer - nFreezer is an encrypted-at-rest backup tool.