mpifileutils
fdupes
mpifileutils | fdupes | |
---|---|---|
4 | 17 | |
160 | 2,363 | |
0.6% | - | |
5.1 | 2.3 | |
21 days ago | 4 days ago | |
C | C | |
BSD 3-clause "New" or "Revised" License | - |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
mpifileutils
-
Pigz: A parallel implementation of gzip for multi-core machines
If you ever run into the limitations of a single machine, dbz2 is also a fun little app for this sort of thing. You can run it across multiple machines and it'll automatically balance the workload across them.
https://github.com/hpc/mpifileutils/blob/master/man/dbz2.1
- MpiFileUtils: File utilities designed for scalability and performance
-
Go Find Duplicates: blazingly-fast simple-to-use tool to find duplicate files
If you want something that scales horizontally, dcmp from https://github.com/hpc/mpifileutils is an option.
- You can list a directory containing 8M files, but not with ls
fdupes
- Fdupes: Identify or Delete Duplicate Files
- fdupes: Identify or Delete Duplicate Files
-
Removing image duplicates
fdupes is simple and easy to use: https://github.com/adrianlopezroche/fdupes
- Backing Up Data: Tips/Advice for Tons of Unorganized Data and Duplicate Files from Multiple Sources
-
File Deduplication
I recently used [fdupes](https://github.com/adrianlopezroche/fdupes) to figure out duplicate files from my amazon cloud drive / photos migration. Took about 2 days to scour through about 1.5TB worth of day.
-
How would I go about copying around 5TB worth of data, from multiple drives to a singular drive/drives (Shared Pools/Raid)?
I would add the content of your current drives with rysnc to the new big drive. I would then run https://github.com/adrianlopezroche/fdupes To remove duplicate files.
- Ask HN: Tool to find identical file subtrees scattered over disks
-
Which tool do you use to find duplicate files?
jdupes, an optimized fork of the popular fdupes. There's 32-bit and 64-bit Win32 packages of jdupes there on Github.
- Mercredi Tech - 2022-05-11
- Suggestions on how to identify & report on old stale data in file shares?
What are some alternatives?
fclones - Efficient Duplicate File Finder
rdfind - find duplicate files utility
rmlint - Extremely fast tool to remove duplicates and other lint from your filesystem
jdupes - A powerful duplicate file finder and an enhanced fork of 'fdupes'.
pigz - A parallel implementation of gzip for modern multi-processor, multi-core machines.
duphard - A simple utility to detect duplicate files and replace them with hard links.
czkawka - Multi functional app to find duplicates, empty folders, similar images etc.
coreutils - Enhancements to the GNU coreutils (especiall head)
go-find-duplicates - Find duplicate files (photos, videos, music, documents) on your computer, portable hard drives etc.
dupeguru - Find duplicate files