How to find duplicate files in an AWS S3 bucket?
问题 Is there a way to recursively find duplicate files in an Amazon S3 bucket? In a normal file system, I would simply use: fdupes -r /my/directory 回答1: There is no "find duplicates" command in Amazon S3. However, you do do the following: Retrieve a list of objects in the bucket Look for objects that have the same ETag (checksum) and Size They would (extremely likely) be duplicate objects. 回答2: Here's a git repository: https://github.com/chilts/node-awssum-scripts which has a js script file to