mpifileutils
pigz
mpifileutils | pigz | |
---|---|---|
4 | 9 | |
160 | 2,549 | |
0.6% | - | |
5.1 | 3.4 | |
21 days ago | about 1 month ago | |
C | C | |
BSD 3-clause "New" or "Revised" License | - |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
mpifileutils
-
Pigz: A parallel implementation of gzip for multi-core machines
If you ever run into the limitations of a single machine, dbz2 is also a fun little app for this sort of thing. You can run it across multiple machines and it'll automatically balance the workload across them.
https://github.com/hpc/mpifileutils/blob/master/man/dbz2.1
- MpiFileUtils: File utilities designed for scalability and performance
-
Go Find Duplicates: blazingly-fast simple-to-use tool to find duplicate files
If you want something that scales horizontally, dcmp from https://github.com/hpc/mpifileutils is an option.
- You can list a directory containing 8M files, but not with ls
pigz
-
Pigz: Parallel gzip for modern multi-processor, multi-core machines
You can grab the version from the solaris userland repo I linked and use it without me completing a homework assignment. Just grab the pigz-2.3.4 source then apply the patches from [1] in the proper order. Maybe some of them aren't needed for non-Solaris.
1. https://github.com/oracle/solaris-userland/tree/master/compo...
I thought I had opened a PR for that a long while ago, but it doesn't show up on github these days. In any case, I did ask Mark Adler to review it. It was never a priority, then the code changed in ways that I don't really want to deal with.
While looking through the PRs, I noticed a PR for Blocked GZip Format (BGZF) [2]. That's very interesting, and perhaps suggests that bgzip is a tool you would be interested in.
2. https://github.com/madler/pigz/pull/19
- ZSTD 1.5.5 is released with a corruption fix found at Google
- pigz: A parallel implementation of gzip for multi-core machines
-
Pigz: A parallel implementation of gzip for multi-core machines
The bit I found most interesting was actually:
https://github.com/madler/pigz/blob/master/try.h
https://github.com/madler/pigz/blob/master/try.c
which implements try/catch for C99.
-
Is there stronger zip compression than 7z a -mx9?
pigz seems to be able to do it. https://github.com/madler/pigz
What are some alternatives?
fclones - Efficient Duplicate File Finder
rapidgzip - Gzip Decompression and Random Access for Modern Multi-Core Machines
rmlint - Extremely fast tool to remove duplicates and other lint from your filesystem
TurboBench - Compression Benchmark
duphard - A simple utility to detect duplicate files and replace them with hard links.
zstd
coreutils - Enhancements to the GNU coreutils (especiall head)
isa-l - Intelligent Storage Acceleration Library
jdupes - A powerful duplicate file finder and an enhanced fork of 'fdupes'.
zstd - Zstandard - Fast real-time compression algorithm
rdfind - find duplicate files utility
PostgreSQL - Mirror of the official PostgreSQL GIT repository. Note that this is just a *mirror* - we don't work with pull requests on github. To contribute, please see https://wiki.postgresql.org/wiki/Submitting_a_Patch