bottleneck
jdupes
bottleneck | jdupes | |
---|---|---|
1 | 44 | |
1,006 | 1,681 | |
1.4% | - | |
3.5 | 0.0 | |
4 days ago | 7 months ago | |
Python | C | |
BSD 2-clause "Simplified" License | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
bottleneck
-
Update on my Python, C++ and Rust Library
Fast Array Manipulation in Python: Since Numpy is the de facto standard for storing multi-dimensional data, any performance gain you see using librapid math kernels will need to be realized on data which probably started its life as a numpy array, and needs to be passed to another tool as a numpy array. Hopefully there will be (or already is?) a way to build a librapid array out of a numpy array without copying the data and vice versa. In fact I might suggest that librapid focus on the fast math operations and simply become an accelerator for numpy arrays. For instance, look at CuPy which provides GPU-implemented operations within a numpy-compatible API, and Bottleneck which simply provides fast C-based implementations of some otherwise slow parts of Numpy. Also note that numpy *can* be multi-threaded depending on the operation and some environment variables. Single-threaded to Single-threaded I think you will be hard-pressed to beat Numpy on general math operations, but that doesn't mean there aren't specific "kernels" that are more specialized that can be greatly improved with a C++ back-end.
jdupes
-
File Servers... how are you handling duplicates
I recommend the use of jdupes, a fork of the well-known fdupes, to find duplicate files.
-
fdupes: Identify or Delete Duplicate Files
200 lines of Nim [1] seems to run about 9X faster than the 8000 lines of C in fdupes on a little test dir I have. If you need C, I think jdupes [2] is faster as @TacticalCoder points out a couple of times here. In my testing, `dups` is usually faster than `jdupes`, though.
[1] https://github.com/c-blake/bu/blob/main/dups.nim
[2] https://github.com/jbruchon/jdupes
-
I'm amazed how I find anything & why I have so many dupes!
There's always the well-respected tool, Czkawka. Or, of the CLI is your thing, jdupes is a good option.
- Anyone know of any good file deduplication tools?
-
Johnny Decimal
My research into this many years ago turned out that jdupes was the right / best solution I could find for my usecase.
https://github.com/jbruchon/jdupes
Though that works fine from a script perspective I'd like some more interactive way of sorting directories etc. Identifying is just the first step, jdupes helps with linking the files (both soft and hard links comes with caveats though!) but that is mostly to save space, not to help in reorganisation.
- Jdupes: A powerful duplicate file finder
-
Does jdupes do a 'dry run' if you just specify directory(s) and no other options
I can work it out by looking at https://github.com/jbruchon/jdupes.
-
replace duplicates with hard links - I think jdupes is the answer, or maybe fclones (I have questions)
I have looked at a few alternatives and think jdupes is the one for me. Then I found out it was not multi-threaded so will give it a go but the developer of jdupes recomended fclones (https://github.com/jbruchon/jdupes/issues/186) if you were dealing with large file systems and wanted multi-threading. But as I am using a HD it may not be necessary.
-
De-Duping a file server
jdupes is a fork of the old standby fdupes, but it has a Win32 release as well as supporting POSIX.
-
Any good duplicate file finder for windows?
jdupes is a tuned fork of the well-known fdupes, and has Win32 releases.
What are some alternatives?
cupy - NumPy & SciPy for GPU
fdupes - FDUPES is a program for identifying or deleting duplicate files residing within specified directories.
NumPy - The fundamental package for scientific computing with Python.
dupeguru - Find duplicate files
pyxirr - Rust-powered collection of financial functions.
rmlint - Extremely fast tool to remove duplicates and other lint from your filesystem
segyio - Fast Python library for SEGY files.
rdfind - find duplicate files utility
trusted-traveler-scheduler - Python script for periodically fetching appointment dates from the Trusted Traveler Program API for Global Entry, Nexus, SENTRI, and FAST, with notifications to the user when new appointments are discovered.
duperemove - Tools for deduping file systems
czkawka - Multi functional app to find duplicates, empty folders, similar images etc.
fclones - Efficient Duplicate File Finder