Profile

User's avatar
npub1kws0...0n5s
npub1kws0...0n5s
Anybody have a preferred way to deduplicate files in Linux? For context, what I’m working with is ~2TB of backups from external drives with the largely the same set of users. I would not trust that identical names are identical files, and I’m sure that identical files live under multiple names. Lots of pictures, some music, etc., and the desired end state is reclaiming space and making the directory structure more usable without losing information from the current directory structure. For the pictures, maybe that means adding EXIF tags that contain the current path, then consolidating EXIF data from identical files, then a second path through something that can pick up duplicates of e.g. different resolutions? Or is my best path creating a working copy of everything and trying to manually rearrange it one bite at a time?