czkawka
dupeguru
czkawka | dupeguru | |
---|---|---|
364 | 154 | |
20,515 | 5,494 | |
- | - | |
7.4 | 6.0 | |
about 2 months ago | 4 months ago | |
Rust | Python | |
GNU General Public License v3.0 or later | GNU General Public License v3.0 only |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
czkawka
-
Ask HN: How do you deduplicate files?
You want content-addressed storage; this works with rolling content hashes that identify common blocks of memory. `rsync` uses that technique to minimize bytes to be transferred. https://github.com/qarmin/czkawka is a GUI app and CLI tool to find identical files in general and similar images in particular.
The task is much simpler if you only want to find bit-identical entire files, not part of files; in that case, you can just run a tool like `sha1sum` over each file and record the hash digest in a database; identical files—and only identical ones, with high probability—will have the same hash, non-identical ones will have different hashes.
- Czkawka: Multi functional app to find duplicates, empty folders, similar images
-
Duperemove – Tools for deduping file systems
You might be interested in this app: https://github.com/qarmin/czkawka
- Is there software to compress large but similar files?
- Merge three separate partial libraries from external USB drives
-
Tools to deduplicate files
https://github.com/qarmin/czkawka by far the best of anything iv tried
-
fdupes: Identify or Delete Duplicate Files
I've used Czkawka (https://github.com/qarmin/czkawka) because it does Lanczos-based image duplicate detection, which makes it more practical for me.
-
AllDup suddenly taking forever to process/delete selections
Maybe it's a setting you made or the files, not sure. You can try another software czkawka to see if you get better results with it.
-
Is there a file duplicate finder that works with animated jpegxl-gif?
For static images i used https://github.com/qarmin/czkawka and it works well enough. I think. But when i used it on a folder with gifs and their jxl conversions, it shows nothing. SURELY this could not be user error, rrrright?
-
PhotoPrism: Browse Your Life in Pictures
I used to use DupeGuru which has some photo-specific dupe detection where you can fuzzy match image dupes based on content: https://dupeguru.voltaicideas.net/
But I switched over to czkawka, which has a better interface for comparing files, and seems to be a bit faster: https://github.com/qarmin/czkawka
Unfortunately, neither of these are integrated into Photoprism, so you still have to do some file management outside the database before importing.
I also haven't used Photoprism extensively yet (I think it's running on one of my boxes, but I haven't gotten around to setting it up), but I did find that it wasn't really built for file-based libraries. It's a little more heavyweight, but my research shows that Nextcloud Memories might be a better choice for me (it's not the first-party Nextcloud photos app, but another one put together by the community): https://apps.nextcloud.com/apps/memories
dupeguru
-
How to use onedrive for culling photos
Dupeguru
- Does anyone know any freeware duplicate file checkers without an upsell similar to awesome duplicate photo finder?
-
DupeGuru: Open-source, cross-platform GUI tool to find duplicate files
Posted link appears be misspelled
This appears to be actual link:
https://dupeguru.voltaicideas.net/
-
PhotoPrism: Browse Your Life in Pictures
I used to use DupeGuru which has some photo-specific dupe detection where you can fuzzy match image dupes based on content: https://dupeguru.voltaicideas.net/
But I switched over to czkawka, which has a better interface for comparing files, and seems to be a bit faster: https://github.com/qarmin/czkawka
Unfortunately, neither of these are integrated into Photoprism, so you still have to do some file management outside the database before importing.
I also haven't used Photoprism extensively yet (I think it's running on one of my boxes, but I haven't gotten around to setting it up), but I did find that it wasn't really built for file-based libraries. It's a little more heavyweight, but my research shows that Nextcloud Memories might be a better choice for me (it's not the first-party Nextcloud photos app, but another one put together by the community): https://apps.nextcloud.com/apps/memories
-
App recommendation for finding duplicate images
If your photos are exact duplicates you can also use the freeware dupeGuru app.
-
I'm amazed how I find anything & why I have so many dupes!
If you want a some other GUI options, AllDup or DupeGuru might be of interest. AllDup has a rather weird interface, though, imo.
-
Johnny Decimal
I used DupeGuru (https://dupeguru.voltaicideas.net/) in the past but I'm not sure it's the best solution for you. Try it, it's open-source.
- DupeGuru: Open-Source, cross-platform GUI software to find duplicate files
-
Diff tool for medias
dupeGuru
-
App to find duplicates across multiple external drives (Mac/Unix)
dupeGuru is a graphical application available for macOS that can scan multiple drives and identify duplicate files based on various criteria like name, size, and content https://dupeguru.voltaicideas.net/
What are some alternatives?
jdupes - A powerful duplicate file finder and an enhanced fork of 'fdupes'.
fdupes - FDUPES is a program for identifying or deleting duplicate files residing within specified directories.
AntiDupl - A program to search similar and defect pictures on the disk
coronavirus-dashboard - Dashboard for tracking Coronavirus (COVID-19) across the UK
PhotoPrism - AI-Powered Photos App for the Decentralized Web 🌈💎✨
video-simili-duplicate-cleaner
darktable - darktable is an open source photography workflow application and raw developer
snapraid - A backup program for disk arrays. It stores parity information of your data and it recovers from up to six disk failures
datacurator-filetree - a standard filetree for /r/datacurator [ and r/datahoarder ]
rmlint - Extremely fast tool to remove duplicates and other lint from your filesystem