Re: HD 'Analysis'

2009-05-18 Thread Tobias Walkowiak
On Tue, May 05, 2009 at 01:46:00PM +0200, Hannah Schroeter wrote:
 
dd if=/dev/sd1c of=/dev/null bs=64k
^r
 
 Do yourself a favor and use the raw device.

why?

-- 
t.walkow...@wallstreet-online.de



Re: HD 'Analysis'

2009-05-18 Thread Otto Moerbeek
On Mon, May 18, 2009 at 03:21:25PM +0200, Tobias Walkowiak wrote:

 On Tue, May 05, 2009 at 01:46:00PM +0200, Hannah Schroeter wrote:
  
 dd if=/dev/sd1c of=/dev/null bs=64k
 ^r
  
  Do yourself a favor and use the raw device.
 
 why?

Because you use block devices for mounting, and raw devices for
anything else. In some cases a block devices might work for a
particular purpose, but why take chances?

-Otto



Re: HD 'Analysis'

2009-05-18 Thread Hannah Schroeter
Hi!

On Mon, May 18, 2009 at 03:21:25PM +0200, Tobias Walkowiak wrote:
On Tue, May 05, 2009 at 01:46:00PM +0200, Hannah Schroeter wrote:

dd if=/dev/sd1c of=/dev/null bs=64k
^r

 Do yourself a favor and use the raw device.

why?

If nothing else, it'll be much faster.

Kind regards,

Hannah.



Re: HD 'Analysis'

2009-05-18 Thread David Vasek

On Mon, 18 May 2009, Otto Moerbeek wrote:


On Mon, May 18, 2009 at 03:21:25PM +0200, Tobias Walkowiak wrote:


On Tue, May 05, 2009 at 01:46:00PM +0200, Hannah Schroeter wrote:



  dd if=/dev/sd1c of=/dev/null bs=64k

   ^r

Do yourself a favor and use the raw device.


why?


Because you use block devices for mounting, and raw devices for
anything else. In some cases a block devices might work for a
particular purpose, but why take chances?


While this rule is probably right, a few tools, such as atactl(8) or 
smartctl from ports say something else in their respective man pages. 
Well, these two utilities are not handling the block themselves, but 
anyway. Is this a bug in the man pages?


Regads,
David



Re: HD 'Analysis'

2009-05-14 Thread Landry Breuil
On Thu, May 14, 2009 at 12:42 AM, Marco Peereboom sl...@peereboom.us wrote:
 I once wrote a fancy dd to recover a disk that jordan used for pictures.
 It worked well enough to get the crap off before the disk totally.
 Anyway I dusted it off and added a man page and stuff.  Have a look at
 http://www.peereboom.us/diskrescue/
 if you want to play.

Ah, and btw we have sysutils/testdisk in ports for all those kind of
recoveries.

Landry



Re: HD 'Analysis'

2009-05-13 Thread Marco Peereboom
I once wrote a fancy dd to recover a disk that jordan used for pictures.
It worked well enough to get the crap off before the disk totally.
Anyway I dusted it off and added a man page and stuff.  Have a look at
http://www.peereboom.us/diskrescue/
if you want to play.

I'll add some more language to the man page when I get time.



Re: HD 'Analysis'

2009-05-07 Thread Marco Peereboom
You people crack me up.  I have been trying to ignore this post for a
while but can't anymore.  Garbage like badblock are from the era that
you still could low level format a drive.  Remember those fun days?
When you were all excited about your 10MB hard disk?

Use dd to read it; if it is somewhat broken the drive will reallocate
it.  If it is badly broken the IO will fail and it is time to toss the
disk.  Those are about all the flavors you have available.  Running
vendor diags is basically a fancier dd.

On Thu, May 07, 2009 at 01:10:56AM +0200, ropers wrote:
  On Monday 04 May 2009 17:56:43 L. V. Lammert wrote:
   What is the best way to do a surface analysis on a disk?
 
 
 2009/5/5 Tony Abernethy t...@servacorp.com:
  There is, in the e2fsprogs package, something called badblocks.
  I have used it (on Linux) to rescue bad disks.
  (Windows laptops  -- kinda redundant?)
 
  If you care about your data, follow Steve's advice.
 
  The reality seems to be that this does exercise a disk's ability
  to relocate bad sectors so that a bad disk suddenly goes good.
  This is using a destructive surface test  (badblocks -sw ...)
  Realistically, seems like the most reliable test is that disk is slower
  than it should be.
 
  Me, if I want to rely on a disk drive, I will run badblocks on it.
  The long-winded destructive test
  And I will time it, at least sporadically.
  (New disks are not immune from having problems ;-)
  The exercise maybe loses out to watching grass grow.
 
 I also would recommend badblocks(8), but I would recommend
   badblocks -svn
 instead of badblocks -sw.
 
 badblocks -svn also (s)hows its progress as it goes along, but does a
 (v)erbose (n)on-destructive read/write test (as opposed to either the
 default read-only test or the destructive read/write test). You can
 check an entire device with badblocks, or a partition, or a file. The
 great thing about using badblocks to check a partition is that it's
 filesystem-agnostic. It will dutifully check every bit of its target
 partition regardless of what's actually on it. And if you give
 badblocks -svn an entire storage device to test, it will not even care
 about the actual partition scheme used. Because this read/write test
 can trigger the disk's own built-in bad sector relocation, this means
 you can even have a disk that you can't read the partition table from,
 and running badblocks -svn over it may at least temporarily fix
 things. And I've used badblocks -svn e.g. to check old Macintosh
 floppies. Who cares that OpenBSD doesn't know much about the
 filesystem on those? badblocks does the job anyway.
 
 (Because of this agnosticism, it's actually questionable whether
 badblocks(8) ought to be part of a filesystem-specific package, but
 hey, that's what it comes in. Yea, one *could* also argue whether to
 include it elsewhere by default because it's so useful, but I'm not
 the one making those decisions and I guess the folks who do will do
 what makes the most sense to them, so I don't feel like starting to be
 a back seat driver... ;-)
 
 Oh, and of course it would probably be prudent to do a backup before
 read/write tests, even though badblocks is well-established and (with
 -n) supposed to be non-destructive. Supposed to... ;-) I've never been
 disappointed but YMMV.
 
 regards,
 --ropers



Re: HD 'Analysis'

2009-05-07 Thread ropers
  On Monday 04 May 2009 17:56:43 L. V. Lammert wrote:
   What is the best way to do a surface analysis on a disk?

 2009/5/5 Tony Abernethy t...@servacorp.com:
  There is, in the e2fsprogs package, something called badblocks.

 On Thu, May 07, 2009 at 01:10:56AM +0200, ropers wrote:
 I also would recommend badblocks(8), but I would recommend
   badblocks -svn
 instead of badblocks -sw.

 badblocks -svn also (s)hows its progress as it goes along, but does a
 (v)erbose (n)on-destructive read/write test (as opposed to either the
 default read-only test or the destructive read/write test). You can
 check an entire device with badblocks, or a partition, or a file. The
 great thing about using badblocks to check a partition is that it's
 filesystem-agnostic. It will dutifully check every bit of its target
 partition regardless of what's actually on it. And if you give
 badblocks -svn an entire storage device to test, it will not even care
 about the actual partition scheme used. Because this read/write test
 can trigger the disk's own built-in bad sector relocation, this means
 you can even have a disk that you can't read the partition table from,
 and running badblocks -svn over it may at least temporarily fix
 things. And I've used badblocks -svn e.g. to check old Macintosh
 floppies. Who cares that OpenBSD doesn't know much about the
 filesystem on those? badblocks does the job anyway.

 Oh, and of course it would probably be prudent to do a backup before
 read/write tests, even though badblocks is well-established and (with
 -n) supposed to be non-destructive. Supposed to... ;-) I've never been
 disappointed but YMMV.

2009/5/7 Marco Peereboom sl...@peereboom.us:
 You people crack me up.  I have been trying to ignore this post for a
 while but can't anymore.  Garbage like badblock are from the era that
 you still could low level format a drive.  Remember those fun days?
 When you were all excited about your 10MB hard disk?

 Use dd to read it; if it is somewhat broken the drive will reallocate
 it.  If it is badly broken the IO will fail and it is time to toss the
 disk.  Those are about all the flavors you have available.  Running
 vendor diags is basically a fancier dd.

Why do you consider badblocks garbage?

I remember now that we talked about this before over a year ago, when
I first asked about using badblocks on OpenBSD. Back then I eventually
surmised that using dd to do the same thing as badblocks -svn would be
possible but a lot more cumbersome, cf.:
http://kerneltrap.org/mailarchive/openbsd-misc/2008/4/19/1499524

Am I/was I mistaken, and if so, where?

Thanks and regards,
--ropers



Re: HD 'Analysis'

2009-05-07 Thread Tony Abernethy
Marco Peereboom wrote:
   On Monday 04 May 2009 17:56:43 L. V. Lammert wrote:
What is the best way to do a surface analysis on a disk?
 
  2009/5/5 Tony Abernethy t...@servacorp.com:
   There is, in the e2fsprogs package, something called badblocks.
 
  On Thu, May 07, 2009 at 01:10:56AM +0200, ropers wrote:
  I also would recommend badblocks(8), but I would recommend
badblocks -svn
  instead of badblocks -sw.
 
  badblocks -svn also (s)hows its progress as it goes along, 
 but does a
  (v)erbose (n)on-destructive read/write test (as opposed to 
 either the
  default read-only test or the destructive read/write test). You can
  check an entire device with badblocks, or a partition, or 
 a file. The
  great thing about using badblocks to check a partition is that it's
  filesystem-agnostic. It will dutifully check every bit of 
 its target
  partition regardless of what's actually on it. And if you give
  badblocks -svn an entire storage device to test, it will 
 not even care
  about the actual partition scheme used. Because this 
 read/write test
  can trigger the disk's own built-in bad sector relocation, 
 this means
  you can even have a disk that you can't read the partition 
 table from,
  and running badblocks -svn over it may at least temporarily fix
  things. And I've used badblocks -svn e.g. to check old Macintosh
  floppies. Who cares that OpenBSD doesn't know much about the
  filesystem on those? badblocks does the job anyway.
 
  Oh, and of course it would probably be prudent to do a 
 backup before
  read/write tests, even though badblocks is 
 well-established and (with
  -n) supposed to be non-destructive. Supposed to... ;-) 
 I've never been
  disappointed but YMMV.
 
 2009/5/7 Marco Peereboom sl...@peereboom.us:
  You people crack me up.  I have been trying to ignore this 
 post for a
  while but can't anymore.  Garbage like badblock are from 
 the era that
  you still could low level format a drive.  Remember those fun days?
  When you were all excited about your 10MB hard disk?
 
  Use dd to read it; if it is somewhat broken the drive will 
 reallocate
  it.  If it is badly broken the IO will fail and it is time 
 to toss the
  disk.  Those are about all the flavors you have available.  Running
  vendor diags is basically a fancier dd.
 
 Why do you consider badblocks garbage?
OK, I'll take a nibble. (flames invited where I've got anything wrong)

You use OpenBSD where sloppy doesn't quite do what you need to be done.
This is a world where a false sense of security is not your friend.
This disk is good because it passed badblocks is NOT valid.
I've got too many rescued disks that will probably keep on working.
probably: better then 50%. (but it sounds good)
depending on lots of probables is really instant death.

IF badblocks passed a disk as clean, and there were good reason to 
beleieve that that disk was actually clean, and that it would STAY
clean, then it (badblocks) would be a good program.
Unfortunately, there is not much of anything that badblocks, or the
vendors' programs CAN do that is much of an assurance of reliability.
You might get some idea from the reliability of reconditioned 
drives versus the reliability of actually new drives. And the vendors
have better tools (if such as better tools actually exist).

WITHOUT going into HW or OS handling of bad sectors, simply rename
files or directories something like BAD_STUFF and NEVER delete 'em.
There are exotic ways of increasing risk by keeping the most of the
not-failed-yet neighbors as supposedly good sectors.
You can do much of that by partitioning to avoid places with a lot
of bad stuff. With the prices and capacities of modern disks, all
of this must assume that you have lots of time and need something to
occupy that time. Watching grass grow is probably more exciting.

For a new disk (one that does not need to go into production soon)
you can run a very long winded excercise. Seroing and reading 
probably as effective and certainly faster than 0xAA 0x55 0xFF 0x00

There SHOULD be good data forthcoming from the SMART stuff.
BUT, so far I've haven't heard noises from that corner, just wise-
cracks about vendor diags. Presumably, SHOULD does not imply IS.
IF you have anything resembling money, and do not have lots of 
free time on your hands, the best advice seems to be to replace 
quickly anything that shows any sign of trouble.
(This might be an actual good use of benchmarks ;-)

Reading will reallocate sectors.
The sector after the reallocation will be readable.
The contents of this now readable sector will be the orginal contents 
if the drive managed to successully eventually read those original 
contents, seems like whatever the drive can fake in some cases. 
Seems like with NO indication of problems in some cases at least.
Very hard to be certain at this level (using inferior OSes)

Short answer, is that AFTER a long and complicated process, there
is no reason to believe that the contents of the 

Re: HD 'Analysis'

2009-05-07 Thread Sean Kamath
On May 7, 2009, at 4:50 PM, Tony Abernethy wrote:


 There are exotic ways of increasing risk by keeping the most of the
 not-failed-yet neighbors as supposedly good sectors.

Not with a modern disk.  The drives now essentially lie about where on  
the disk any given block is, you'll never know if block N is anywhere  
(physically) near block N-1 or N+1.

Starting about 15 years ago, the most reasonable check I could find  
was the 'verify' command in solaris' 'format' command (which I've yet  
to find/write a simple alternative to).  Anything else is just a waste  
of time.

What this did was basically write a block of random bits, then read  
and compare.  You need to do both, because some blocks are readable,  
but not writable, and vice versa. If you get a mismatch, the block was  
unreadable, and was (hopefully) remapped, so try again.  The OS  
usually logs read and write errors (soft and/or hard) and you'd have  
some idea of the relative 'health' of the disk.

Frankly, we would verify a disk if we hit a bad block, and if that  
remapped the bad block and produced no other errors over two passes,  
we'd keep using it (disks weren't that cheap then).  If we got another  
error, we'd replace the disk.  We got so many new disks that would  
encounter a bad block (and the OS would log the error) that we started  
verifying the disk when we got them to map out any bad blocks. . .

Sean

[demime 1.01d removed an attachment of type application/pkcs7-signature which 
had a name of smime.p7s]



Re: HD 'Analysis'

2009-05-06 Thread Steve Shockley

On 5/5/2009 12:50 PM, Josi Quinteiro wrote:

First thing I do with a new hard drive is run a long self-test using
smartctl. If it passes it gets added to the system. I have smartd set to
do a daily short self-test and a weekly long self-test on every drive.
Replace any drives that start to show errors.


The self-tests take the drive offline while they run, right?  Do you 
unmount them first, or is the system okay just waiting until the drive 
responds?




Re: HD 'Analysis'

2009-05-06 Thread Steve Shockley

On 5/5/2009 11:49 AM, L. V. Lammert wrote:

Some good options, .. seems like all are DOS, however g!! I guess
that's no big deal if you're rebooting for the analysis, but it does not
seem 'right'!


No, they have a Windows version of Victoria! g  Personally, I use 
these kinds of utilities to see if a drive is worth saving, when I can 
do destructive tests.  For example I recovered a 250gb disk from an 
XServe RAID that i use as a second drive in my work desktop.  SMART 
reports 300 reallocation events, but no matter what I do that doesn't 
increase.  I use it for temporary storage for easy-to-replace data.




Re: HD 'Analysis'

2009-05-06 Thread Martin Schröder
2009/5/6, Steve Shockley steve.shock...@shockley.net:
  The self-tests take the drive offline while they run, right?  Do you

No. man smartctl

Best
   Martin



Re: HD 'Analysis'

2009-05-06 Thread ropers
 On Monday 04 May 2009 17:56:43 L. V. Lammert wrote:
  What is the best way to do a surface analysis on a disk?


2009/5/5 Tony Abernethy t...@servacorp.com:
 There is, in the e2fsprogs package, something called badblocks.
 I have used it (on Linux) to rescue bad disks.
 (Windows laptops  -- kinda redundant?)

 If you care about your data, follow Steve's advice.

 The reality seems to be that this does exercise a disk's ability
 to relocate bad sectors so that a bad disk suddenly goes good.
 This is using a destructive surface test  (badblocks -sw ...)
 Realistically, seems like the most reliable test is that disk is slower
 than it should be.

 Me, if I want to rely on a disk drive, I will run badblocks on it.
 The long-winded destructive test
 And I will time it, at least sporadically.
 (New disks are not immune from having problems ;-)
 The exercise maybe loses out to watching grass grow.

I also would recommend badblocks(8), but I would recommend
  badblocks -svn
instead of badblocks -sw.

badblocks -svn also (s)hows its progress as it goes along, but does a
(v)erbose (n)on-destructive read/write test (as opposed to either the
default read-only test or the destructive read/write test). You can
check an entire device with badblocks, or a partition, or a file. The
great thing about using badblocks to check a partition is that it's
filesystem-agnostic. It will dutifully check every bit of its target
partition regardless of what's actually on it. And if you give
badblocks -svn an entire storage device to test, it will not even care
about the actual partition scheme used. Because this read/write test
can trigger the disk's own built-in bad sector relocation, this means
you can even have a disk that you can't read the partition table from,
and running badblocks -svn over it may at least temporarily fix
things. And I've used badblocks -svn e.g. to check old Macintosh
floppies. Who cares that OpenBSD doesn't know much about the
filesystem on those? badblocks does the job anyway.

(Because of this agnosticism, it's actually questionable whether
badblocks(8) ought to be part of a filesystem-specific package, but
hey, that's what it comes in. Yea, one *could* also argue whether to
include it elsewhere by default because it's so useful, but I'm not
the one making those decisions and I guess the folks who do will do
what makes the most sense to them, so I don't feel like starting to be
a back seat driver... ;-)

Oh, and of course it would probably be prudent to do a backup before
read/write tests, even though badblocks is well-established and (with
-n) supposed to be non-destructive. Supposed to... ;-) I've never been
disappointed but YMMV.

regards,
--ropers



Re: HD 'Analysis'

2009-05-06 Thread Steve Shockley

On 5/6/2009 11:24 AM, Martin Schrvder wrote:

2009/5/6, Steve Shockleysteve.shock...@shockley.net:

  The self-tests take the drive offline while they run, right?  Do you


No. man smartctl


Huh.  That kind of contradicts the name offline self test, but I guess 
they call that captive.




Re: HD 'Analysis'

2009-05-05 Thread Hannah Schroeter
Hi!

On Mon, May 04, 2009 at 06:34:07PM -0400, STeve Andre' wrote:
[...]

I have a pile of disks that I suspect.  Looking at the drawer, I see 8
of them.  As I have time I test them, usually with dd:

   dd if=/dev/sd1c of=/dev/null bs=64k
   ^r

Do yourself a favor and use the raw device.

[...]

Kind regards,

Hannah.



Re: HD 'Analysis'

2009-05-05 Thread L. V. Lammert

At 10:32 PM 5/4/2009 -0400, Steve Shockley wrote:

On 5/4/2009 5:56 PM, L. V. Lammert wrote:

What is the best way to do a surface analysis on a disk? badsect seems
like a holdover from MB-sized disks, and it doesn't do any analysis.


MHDD might do what you want:

http://hddguru.com/content/en/software/2005.10.02-MHDD/

I haven't used it, but Victoria (http://hdd-911.com/) might be useful if 
you can read Russian.


Gibson's Spinrite is okay to check a drive but he tries to imply that what 
he does is way more complicated than it really is.  That, and the author 
is a weenie media whore.


I rarely see a bad drive lock up the system on modern machines without 
timeout messages on the console, etc.  Your controller or cable may be 
suspect if the drive passes all the tests you throw at it.


Some good options, .. seems like all are DOS, however g!! I guess that's 
no big deal if you're rebooting for the analysis, but it does not seem 'right'!


Lee



Re: HD 'Analysis'

2009-05-05 Thread L. V. Lammert

At 05:45 PM 5/4/2009 -0500, Tony Abernethy wrote:


There is, in the e2fsprogs package, something called badblocks.
I have used it (on Linux) to rescue bad disks.
(Windows laptops  -- kinda redundant?)


Interesting, .. it DNB on 4.0, however, .. and I'm unsure as to any issues 
between utilities designed for ext2 and ffs???



If you care about your data, follow Steve's advice.


Right. How many disks should I throw away before trying to gather some 
USEFUL data?



Me, if I want to rely on a disk drive, I will run badblocks on it.


Sounds like the best idea - do you run it from a Linux CD, or ??

Thanks!

Lee



Re: HD 'Analysis'

2009-05-05 Thread L. V. Lammert

At 03:36 PM 5/4/2009 -0700, Jose Quinteiro wrote:

I use this http://smartmontools.sourceforge.net/

Saludos,
Jose.


Thanks! I have used smart tools in the past, .. but how do you use them for 
testing?


Lee



Re: HD 'Analysis'

2009-05-05 Thread José Quinteiro
First thing I do with a new hard drive is run a long self-test using 
smartctl.  If it passes it gets added to the system.  I have smartd set 
to do a daily short self-test and a weekly long self-test on every 
drive.  Replace any drives that start to show errors.


Saludos,
Jose.

L. V. Lammert wrote:

At 03:36 PM 5/4/2009 -0700, Jose Quinteiro wrote:

I use this http://smartmontools.sourceforge.net/

Saludos,
Jose.


Thanks! I have used smart tools in the past, .. but how do you use them 
for testing?


Lee




Re: HD 'Analysis'

2009-05-05 Thread STeve Andre'
On Tuesday 05 May 2009 12:11:49 L. V. Lammert wrote:
 At 05:45 PM 5/4/2009 -0500, Tony Abernethy wrote:
 There is, in the e2fsprogs package, something called badblocks.
 I have used it (on Linux) to rescue bad disks.
 (Windows laptops  -- kinda redundant?)

 Interesting, .. it DNB on 4.0, however, .. and I'm unsure as to any issues
 between utilities designed for ext2 and ffs???

 If you care about your data, follow Steve's advice.

 Right. How many disks should I throw away before trying to gather some
 USEFUL data?

Perhaps I didn't word my thoughts well enough, and appeared snarky
to you?  That wasn't my intent.

Disks today are 1) VASTLY cheaper per meg of storage, 2) Faster, 3)
less power comsumptive and noisy.

But there is also 4) which is they aren't built as well.  The MTBF figures
are a mathmatical fantasy, and dangerously worthless.  I have many
older systems running small disks from 2G to about 20G that are
still fine since 1996.  In fact, looking at my log of disk disasters, I've
had three disks blow up when being used by my users, when they
were using those machines.  In contrast, the 60G+ disk era has given
me at least 12 problems in the last four to five years, and I'm not
counting friends systems that I've helped out on.  Probably more 
like 18 disasters+ if I count those.

Because of this I've adopted a really careful attitude about disks
in general.  I'm not starting to treat them like airplane parts--replace
them before they fail.  This is especially true for laptop disks (I've
had four disks start to go on various OpenBSD thinkpads I've had).

When you have free time you can beat on a disk, and take weeks
pounding on it.  Look at iogen in the ports tree as another testing
method.  It is also the case that multiple make builds of userland
is a good test.  I'm hesitant to depend on the smart tools, because
I've had laptop disks that failed hours after a check said things
were fine, and I still have a 100G disk generates smart errors
but which is absolutely good.

Remember too that getting a disk replacement under warranty
almost always results in a recertified disk, and I'm nervous about
using them.  Given the cost I get new ones.

Hannah's comment that I should have used the raw device was
quite correct; that was a tyop so it should have said

   dd if=/dev/rsd1c of=/dev/null bs=64k


 Me, if I want to rely on a disk drive, I will run badblocks on it.

 Sounds like the best idea - do you run it from a Linux CD, or ??

  Thanks!

  Lee

--STeve Andre'



HD 'Analysis'

2009-05-04 Thread L. V. Lammert
Been trying to build a replacement HD for a system, .. and it seems 
impossible to verify whether a disk is bad or not (having wasted some hours 
rsync'ing data only to have the HD lock up the system when doing the final 
rsync).


What is the best way to do a surface analysis on a disk? badsect seems like 
a holdover from MB-sized disks, and it doesn't do any analysis.


TIA,

Lee



Re: HD 'Analysis'

2009-05-04 Thread STeve Andre'
On Monday 04 May 2009 17:56:43 L. V. Lammert wrote:
 Been trying to build a replacement HD for a system, .. and it seems
 impossible to verify whether a disk is bad or not (having wasted some hours
 rsync'ing data only to have the HD lock up the system when doing the final
 rsync).

 What is the best way to do a surface analysis on a disk? badsect seems like
 a holdover from MB-sized disks, and it doesn't do any analysis.

   TIA,

   Lee

The best way is to get a new disk.  I'm serious.  Disks are cheap enough, and
the value of whats on them is high enough that if you think its going, get a
new one.  Even if this is a hobby system, I'd do that.

There is disk testing software from the OEMs you can use.

But if you think its acting weird don't trust it.

--STeve Andre'



Re: HD 'Analysis'

2009-05-04 Thread L. V. Lammert

At 06:06 PM 5/4/2009 -0400, STeve Andre' wrote:


The best way is to get a new disk.  I'm serious.  Disks are cheap enough, and
the value of whats on them is high enough that if you think its going, get a
new one.  Even if this is a hobby system, I'd do that.


And I'm serious too - how many hard drives to you throw away before you 
realize that might not be the problem?



There is disk testing software from the OEMs you can use.

But if you think its acting weird don't trust it.


That's why I'm looking for a way to gather some hard data.

Lee



Re: HD 'Analysis'

2009-05-04 Thread Jose Quinteiro

I use this http://smartmontools.sourceforge.net/

Saludos,
Jose.

L. V. Lammert wrote:

At 06:06 PM 5/4/2009 -0400, STeve Andre' wrote:

The best way is to get a new disk.  I'm serious.  Disks are cheap 
enough, and
the value of whats on them is high enough that if you think its going, 
get a

new one.  Even if this is a hobby system, I'd do that.


And I'm serious too - how many hard drives to you throw away before you 
realize that might not be the problem?



There is disk testing software from the OEMs you can use.

But if you think its acting weird don't trust it.


That's why I'm looking for a way to gather some hard data.

Lee




Re: HD 'Analysis'

2009-05-04 Thread STeve Andre'
On Monday 04 May 2009 18:29:26 L. V. Lammert wrote:
 At 06:06 PM 5/4/2009 -0400, STeve Andre' wrote:
 The best way is to get a new disk.  I'm serious.  Disks are cheap enough,
  and the value of whats on them is high enough that if you think its
  going, get a new one.  Even if this is a hobby system, I'd do that.

 And I'm serious too - how many hard drives to you throw away before you
 realize that might not be the problem?

 There is disk testing software from the OEMs you can use.
 
 But if you think its acting weird don't trust it.

 That's why I'm looking for a way to gather some hard data.

  Lee

I have a pile of disks that I suspect.  Looking at the drawer, I see 8
of them.  As I have time I test them, usually with dd:

   dd if=/dev/sd1c of=/dev/null bs=64k

and try that a bunch.  Usually I shake loose a few errors after a 
while.  And of course, listen to them.  Hearing many clicking
noises for recalibrates in a row is another sign of impending
death.

For laptop disks I've used IBM/Hitahi's drive fitness test, and
that usually works well, but earlier this year it gave a drive a
clean bill of health, and the disk died about 2 weeks later.

--STeve Andre'



Re: HD 'Analysis'

2009-05-04 Thread Tony Abernethy
STeve Andre' wrote:
 On Monday 04 May 2009 17:56:43 L. V. Lammert wrote:
  Been trying to build a replacement HD for a system, .. and it seems
  impossible to verify whether a disk is bad or not (having 
 wasted some hours
  rsync'ing data only to have the HD lock up the system when 
 doing the final
  rsync).
 
  What is the best way to do a surface analysis on a disk? 
 badsect seems like
  a holdover from MB-sized disks, and it doesn't do any analysis.
 
  TIA,
 
  Lee
 
 The best way is to get a new disk.  I'm serious.  Disks are 
 cheap enough, and
 the value of whats on them is high enough that if you think 
 its going, get a
 new one.  Even if this is a hobby system, I'd do that.
 
 There is disk testing software from the OEMs you can use.
 
 But if you think its acting weird don't trust it.
 
 --STeve Andre'
 
There is, in the e2fsprogs package, something called badblocks.
I have used it (on Linux) to rescue bad disks.
(Windows laptops  -- kinda redundant?)

If you care about your data, follow Steve's advice.

The reality seems to be that this does exercise a disk's ability
to relocate bad sectors so that a bad disk suddenly goes good.
This is using a destructive surface test  (badblocks -sw ...)
Realistically, seems like the most reliable test is that disk is slower
than it should be.

Me, if I want to rely on a disk drive, I will run badblocks on it.
The long-winded destructive test
And I will time it, at least sporadically.
(New disks are not immune from having problems ;-)
The exercise maybe loses out to watching grass grow.



Re: HD 'Analysis'

2009-05-04 Thread Steve Shockley

On 5/4/2009 5:56 PM, L. V. Lammert wrote:

What is the best way to do a surface analysis on a disk? badsect seems
like a holdover from MB-sized disks, and it doesn't do any analysis.


MHDD might do what you want:

http://hddguru.com/content/en/software/2005.10.02-MHDD/

I haven't used it, but Victoria (http://hdd-911.com/) might be useful if 
you can read Russian.


Gibson's Spinrite is okay to check a drive but he tries to imply that 
what he does is way more complicated than it really is.  That, and the 
author is a weenie media whore.


I rarely see a bad drive lock up the system on modern machines without 
timeout messages on the console, etc.  Your controller or cable may be 
suspect if the drive passes all the tests you throw at it.