In re: [1], either 'du' or 'find' can do what you want. I'm pretty sure that I had a native Windows application called "scanner.exe" that did that too - but I'm unable to locate it right now.
Regards, Michael B. Smith Consultant and Exchange MVP http://TheEssentialExchange.com -----Original Message----- From: Kurt Buff [mailto:[email protected]] Sent: Monday, August 02, 2010 8:49 PM To: NT System Admin Issues Subject: Finding a huge file dump from June... All, On our file server we have a single 1.5tb partition - it's on a SAN. Over the course of 4 days recently it went from about 30% free to about 13% free - someone slammed around 200gb onto the file server. I have a general idea of where it might be - there are two top-level directories that are over 200gb each. However, windirstat hasn't been completely helpful, as I can't seem to isolate which files were loaded during those days, and none of the files that I've been looking at were huge - no ISO or VHD files worth mentioning, etc.. I also am pretty confident that there are a *bunch* of duplicate files on those directories. So, I'm looking for a couple of things: 1) A way to get a directory listing that supports a time/date stamp (my choice of atime, mtime or ctime) size and a complete path name for each file/directory on a single line - something like: 2009-01-08 16:12 854,509 K:\Groups\training\On-Site_Special_Training\Customer1.doc I've tried every trick I can think of for the 'dir' command and it won't do what I want, and the 'ls' command from gunuwin32 doesn't seem to want to do this either. Is there a powershell one-liner that can do this for me perhaps? 2) A recommendation for a duplicate file finder - cheap or free would be preferred. Kurt ~ Finally, powerful endpoint security that ISN'T a resource hog! ~ ~ <http://www.sunbeltsoftware.com/Business/VIPRE-Enterprise/> ~ ~ Finally, powerful endpoint security that ISN'T a resource hog! ~ ~ <http://www.sunbeltsoftware.com/Business/VIPRE-Enterprise/> ~
