finding same files across hardrives
jtc-vS8X3Ji+8Wg6e3DpGhMbh2oLBQzVVOGK at public.gmane.org
Tue Dec 2 13:50:44 UTC 2008
Antonio T. Sun wrote:
> On Mon, 01 Dec 2008 16:53:21 -0500, Jose wrote:
>>> more, if there is any remote possibilities of the following cases, then
>>> the method is not complete:
>>> - Variation in file names. I used to make backups using distinct names,
>>> e.g., file.ver1, file.ver2, etc. If you ever backup your files this
>>> way, then finding duplications by names won't help much.
>>> - Variation in content. Since backups are made over the times, any
>>> slightly change in file content will break the MD5 check sum method
>>> if the above cases (remotely) apply to you, you still need something
>>> that is more suitable for finding duplications, and it is much faster
>>> than creating MD5 check sums.
>> I tried both scripts (thanks guys), but it keeps breaking when it find
>> paths with blanks, like /sdb1/backup/C folder/etc...
>> Basedir breaks, I haven't been able to find a solution to this problem
> Does my above 2 cases apply to you? Are you just trying to find files
> with same name and same content?
> The Toronto Linux Users Group. Meetings: http://gtalug.org/
> TLUG requests: Linux topics, No HTML, wrap text below 80 columns
> How to UNSUBSCRIBE: http://gtalug.org/wiki/Mailing_lists
Basically, yes, like I said, copied same data over different drives with
different mount points and paths, now I am trying to consolidate the all
thing under one single point and properly back up from there.
The Toronto Linux Users Group. Meetings: http://gtalug.org/
TLUG requests: Linux topics, No HTML, wrap text below 80 columns
How to UNSUBSCRIBE: http://gtalug.org/wiki/Mailing_lists
More information about the Legacy