Saturday night craziness: implementing things in shell that are already available in C.
**Tool for detecting duplicate folders**
https://github.com/ilario/finddirdupes
Beware: it is much slower than `rmlint -D`.
Recommended #opensource #file #duplicates detection and deletion: #rmlint
Why? - Extremely fast · #CLI · Candidate file filtering by #name, #size, #modification #time · Configurable criteria for determining the original file · Paranoia mode offered (byte-by-byte comparison) · Flexible output #formats, including #bash deletion script, #json, #CSV · Excellent #documentation and #tutorials
https://github.com/sahib/rmlint
More recommendations: https://tuxwise.net/recommended-software/
After having several troubles with #backuppc, which used to be my tried and true reliable backup server software, I am looking at other ways of doing things.
I have learned that #btrfs file system has features that include not taking up extra disk space when making a copy and that #rmlint can be run on a schedule to de-duplicate.
I think some scripting, cron and rsync on top of btrfs may be a more long term reliable solution to how I'd like to do backup than specialized software.
#rmlint is a filesystem junk cleaner.
rmlint is a simple tool that scans a directory or a set of files for duplicates, empty files/directories, orphaned files, and several other problematic things. After rmlint finishes scanning it creates a #JSON file and shell script which contains the to be removed files. Running the shell script deletes these files after confirmation.
Website 🔗️: https://rmlint.readthedocs.io/
apt 📦️: rmlint