#duplicates #deduplicate #disk #remove #removal #directory #media

bin+lib backdown

A smart CLI for removing thousands of duplicates on your disks

11 releases (5 stable)

1.1.2 Sep 10, 2024
1.1.1 Jul 1, 2023
1.1.0 Dec 5, 2021
1.0.1 Nov 17, 2021
0.0.1 Aug 21, 2020

#137 in Filesystem

MIT license

260KB
1.5K SLoC

backdown

MIT Latest Version Build Chat on Miaou

Backdown helps you safely and ergonomically remove duplicate files.

Its design is based upon my observation of frequent patterns regarding build-up of duplicates with time, especially images and other media files.

Finding duplicates is easy. Cleaning the disk when there are thousands of them is the hard part. What Backdown brings is the easy way to select and remove the duplicates you don't want to keep.

A Backdown session goes through the following phases:

  1. Backdown analyzes the directory of your choice and find sets of duplicates (files whose content is exactly the same). Backdown ignores symlinks and files or directories whose name starts with a dot.
  2. Backdown asks you a few questions depending on the analysis. Nothing is removed at this point: you only stage files for removal. Backdown never lets you stage all items in a set of identical files
  3. After having maybe looked at the list of staged files, you confirm the removals
  4. Backdown does the removals on disk

What it looks like

Analysis and first question:

screen 1

Another kind of question:

screen 2

Yet another one:

screen 3

Yet another one:

screen 4

Review and Confirm:

screen 5

At this point you may also export the report as JSON, and you may decide to replace each removed file with a link to one of the kept ones.

Installation

From the crates.io repository

You must have the Rust env installed: https://rustup.rs

Run

cargo install --locked backdown

From Source

You must have the Rust env installed: https://rustup.rs

Download this repository then run

cargo install --path .

Precompiled binaries

Unless you're a Rust developer, I recommend you just download the precompiled binaries, as this will save a lot of space on your disk.

Binaries are made available at https://dystroy.org/backdown/download/

Usage

Deduplicate any kind of files

backdown /some/directory

Deduplicate images

backdown -i /some/directory

JSON report

After the staging phase, you may decide to export a report as JSON. This doesn't prevent doing also the removals.

The JSON looks like this:

{
  "dup_sets": [
    {
      "file_len": 1212746,
      "files": {
        "trav-copy/2006-05 (mai)/HPIM0530.JPG": "remove",
        "trav-copy/2006-06 (juin)/HPIM0530 (another copy).JPG": "remove",
        "trav-copy/2006-06 (juin)/HPIM0530 (copy).JPG": "remove",
        "trav-copy/2006-06 (juin)/HPIM0530.JPG": "keep"
      }
    },
    {
      "file_len": 1980628,
      "files": {
        "trav-copy/2006-03 (mars)/HPIM0608.JPG": "keep",
        "trav-copy/2006-05 (mai)/HPIM0608.JPG": "remove",
        "trav-copy/2006-06 (juin)/HPIM0608.JPG": "keep"
      }
    },
    {
      "file_len": 1124764,
      "files": {
        "trav-copy/2006-05 (mai)/HPIM0529.JPG": "remove",
        "trav-copy/2006-06 (juin)/HPIM0529.JPG": "keep"
      }
    },
    {
      "file_len": 1706672,
      "files": {
        "trav-copy/2006-05 (mai)/test.jpg": "remove",
        "trav-copy/2006-06 (juin)/HPIM0598.JPG": "keep"
      }
    }
  ],
  "len_to_remove": 8450302
}

Advice

  • If you launch backdown on a big directory, it may find more duplicates you suspect there are. Don't force yourself to answer all questions at first: if you stage the removals of the first dozen questions you'll gain already a lot and you may do the other ones another day
  • Don't launch backdown at the root of your disk because you don't want to try and deal with duplicates in system resources, programs, build artefacts, etc. Launch backdown where you store your images, or your videos or musics
  • Backdown isn't designed for dev directories and doesn't respect .gitignore rules
  • If you launch backdown in a directory with millions files on a slow disk, you'll have to wait a long time while the content is hashed. Try with a smaller directory first if you have an HDD
  • If you're only interested in images, use the -i option

Dependencies

~11–22MB
~349K SLoC