Find duplicate files on Mac OS X

39,250

Solution 1

I'll just copy part of my other answer on a similar subject and also link to a possible duplicate as another reference.

There are many commercial options, some may be better than the listed below, I haven't tried them all. Anyway, I'm listing my choice of apps considering which ones I was able to try. Coincidentally, this was not where I first heard about DupeCheck droplet, but thanks so much Systech for building it up! :)

  • TidyUp is a very well known app in this subject. You can specify where to scan for what kind of duplicates. It offers basic and advanced modes, several different strategies and criterias.

  • MrClean is a free tool that just scans for folders for duplicates and trash them. Very simplistic but efficient if you're sure on what you're doing.

  • Chipmunk scans duplicates and let you choose which ones you want to trash. It offers a node-view of folders and you can select to "delete all files in a folder that have duplicates elsewhere, or vice versa" as well as hand-picking. It may take very long to scan all files, but it does a very decent job after that.

  • DupeCheck "drop a file on it and it will use your Spotlight index to see if you have a potential duplicate somewhere." That's about this nice open source app. Not a great tool for space cleaning at once, but over time it helps you keep your space clean.

  • DuplicateFileSearcher from the website: "is a free powerful software utility that will help you to find and delete duplicate files on your computer. It can also be used to calculate MD5 and SHA hashes. The software runs in Windows, Linux, Solaris and MacOS.". Enough said.

Solution 2

I've written a simple python script FileDupeFinder.py that will find file duplicate files on OSX (and probably Linux/*nix) systems. Right now, you pass it a directory root path and a minimum file size and it will walk the directory tree and output a list of files that it believes to be duplicates. It will ignore files that are hard links of each other. You can tell it to exclude directories (right now it just excludes Backups.backupdb, as this script isn't ready to handle Time Machine wackiness) and tell it not to span filesystems (though these aren't currently paramaterized, they are easy enough to set in the script itself.

This script is only minimally tested, but it has worked reasonably enough for me.

Solution 3

I've converted a bash script into a droplet. Drop a file on it, and it will use your Spotlight index to see if you have a potential duplicate somewhere on your system.
Download the free DupeCheck droplet, unzip it, and find a good spot to put it. Drag and drop any file onto it, and it will use Spotlight to identify possible duplicates.

Solution 4

i recommended dupeGuru by Hardcoded Software, it's Open Source Fairware and free to use

Solution 5

There is Araxis Find Duplicate Files which is pretty convenient.

Share:
39,250

Related videos on Youtube

Kurru
Author by

Kurru

You can find my blog at https://pupeno.com where I publish about coding and other stuff.

Updated on September 17, 2022

Comments

  • Kurru
    Kurru over 1 year

    Is there an application that would search duplicate files in my Mac with an easy way to delete the duplicates?

  • Arjan
    Arjan over 14 years
    And after finding duplicates: remember that using Get Info you might find where the file came from (like from what website, email or which other computer), just in case you wonder how you got the duplicates.
  • Marius
    Marius over 11 years
    Thanks seems to work well. Just took me a few secs to realize you don't pass the root directory in (it accepted it without an error), but you need to pass it as a '-d' param. Otherwise it seems very efficient and tidy.
  • studgeek
    studgeek about 11 years
    It's actually not free to use (anymore at least), its limited to ten deletes/moves until you pay...
  • Armel Larcier
    Armel Larcier almost 9 years
    Well that was true 4 years ago but isn't anymore. It is open source and its master needs developers to keep maintaining it actively. It's a great free soft!... that needs our contribution I guess...
  • jjz
    jjz almost 8 years
    DupeGuru works well and is free.
  • SimplGy
    SimplGy over 7 years
    How would I delete duplicates after finding them? The output file includes both the original and the dupe. Say I modify the script to only include the dupe--how would I go from a line-return separated list of dupe files to a rm on each of them?