Re: 2TB USB hard drive for backing up: XFS info

2012-05-06 Thread Andrei POPESCU
On Vi, 04 mai 12, 15:08:57, Chris Knadle wrote:
 
 Note this means running 'xfs_check' is done when the filesystem is not 
 mounted.  _Supposedly_ it can also be run if the filesystem is mounted read-
 only, but in practice I find it's best (and easier) to run the XFS commands 
 from a LiveCD.  The xfs_check and xfs_repair operations are incredibly fast 
 -- 
 even for a 500GB filesystem it's usually only takes about 10 or 15 seconds.  

Yep, just did that now.

 Speed is generally what XFS is good at, *except* when it comes to deletion of 
 a large number of files -- that's where it's slow.

On advise of a list subscriber I have added the 'delaylog' mount option. 
This is supposed to help if you have a new enough kernel.

 Also in practice I find that any kernel crash or hard-power-off corrupts XFS 
 to at least some extent requiring an xfs_check and xfs_repair, so I have to 
 make sure to keep a LiveCD on hand to be able to do this.

I using xfs only on my laptop, so I have hard power-off only if the 
kernel crashes. I just did the xfs_check and xfs_repair now, but they 
didn't find any problems.

Kind regards,
Andrei
-- 
Offtopic discussions among Debian users and developers:
http://lists.alioth.debian.org/mailman/listinfo/d-community-offtopic


signature.asc
Description: Digital signature


Re: 2TB USB hard drive for backing up: XFS info

2012-05-06 Thread Chris Knadle
On Sunday, May 06, 2012 05:19:20, Andrei POPESCU wrote:
 On Vi, 04 mai 12, 15:08:57, Chris Knadle wrote:
...
  Speed is generally what XFS is good at, *except* when it comes to
  deletion of a large number of files -- that's where it's slow.
 
 On advise of a list subscriber I have added the 'delaylog' mount option.
 This is supposed to help if you have a new enough kernel.

It took me a while to find a reference for what this setting does.  Having 
read the paragraph in the link below [it's question #40 when using the 
contents at the top of the page], it seems there's a speed benefit but also a 
risk of additional corruption in the case of an unclean shutdown.

http://www.xfs.org/index.php/XFS_FAQ#Q:_I_want_to_tune_my_XFS_filesystems_for_.3Csomething.3E

As for the kernel -- I'm currently using Linux 3.3.4.  [I've been custom 
compiling my own kernel using make-kpkg from the 'kernel-package' package for 
a long time now.]

  Also in practice I find that any kernel crash or hard-power-off corrupts
  XFS to at least some extent requiring an xfs_check and xfs_repair, so I
  have to make sure to keep a LiveCD on hand to be able to do this.
 
 I using xfs only on my laptop, so I have hard power-off only if the
 kernel crashes. I just did the xfs_check and xfs_repair now, but they
 didn't find any problems.

The last time I had a hard-power-off both of the XFS filesystems on my laptop 
came up clean also, however I did have corruption and had to clean out about a 
hundred files in /lost+found.  [Mostly temp files, but 'debsums -s' also 
reported files missing from packages, so I needed to reinstall several 
packages to fix that.]

  -- Chris

--
Chris Knadle
chris.kna...@coredump.us
GPG Key: 4096R/0x1E759A726A9FDD74


signature.asc
Description: This is a digitally signed message part.


Re: 2TB USB hard drive for backing up: XFS info

2012-05-05 Thread Chris Knadle
On Friday, May 04, 2012 17:31:23, Camaleón wrote:
 On Fri, 04 May 2012 15:08:57 -0400, Chris Knadle wrote:
  On Monday, April 30, 2012 10:53:46, Camaleón wrote:
  On Sun, 29 Apr 2012 20:16:58 +0200, Martin Steigerwald wrote:
   Am Dienstag, 24. April 2012 schrieb Camaleón:
...
  The steps to actually fsck a XFS filesystem:
 - mount the XFS filesystem to replay the log
 - unmount the XFS filesystem that was just mounted
 - run xfs_check on the filesystem's device
 - if necessary run xfs_repair on the filesystem's device
 
 You mean you don't even notice a file system inconsistency until it
 royally crashes or even something worse? Oh.

Yes that's corret -- XFS does /not/ warn you at boot time (nor mount time) if 
the state of the filesystem is inconsistent.  YOU have to know to check that 
(or find out the hard way),  and if you use XFS for / you somehow need to 
know to do it from a LiveCD.  :-/  [This is rarely explained.]

  Note this means running 'xfs_check' is done when the filesystem is not
  mounted.  _Supposedly_ it can also be run if the filesystem is mounted
  read- only, but in practice I find it's best (and easier) to run the XFS
  commands from a LiveCD.  The xfs_check and xfs_repair operations are
  incredibly fast -- even for a 500GB filesystem it's usually only takes
  about 10 or 15 seconds. Speed is generally what XFS is good at, *except*
  when it comes to deletion of a large number of files -- that's where
  it's slow.
 
 So... is that you don't find it suitable for a standard / partition?

Hmm.  Having given that a thought -- yeah I think that would be a good idea 
and I might be happier using ext4 for / and keep using XFS for /home.  On my 
next reinstall of my laptop (whatever year that will be :-P) I might try that 
and see how I like it.

 I mean, if it's better don't analyze an XFS partition when is mounted
 read-only, that can be really a no-no for many installations running
 24/365.

Yes.  And in addition nounting a / XFS partition read-only to run xfs_repair 
on is fairly tricky.  Even when booting up into single-user mode it's still 
necessary to shut down several processes, which last I recall also includes 
the sshd daemon.  :-/

I'm currently using XFS for / on a remote server I have no physical access 
to, and I'm finding this is a problem because I don't have a good way of 
running xfs_check and xfs_repair on it while the system is running.  The 
hosting company supports remote KVM and bootup to a LiveCD of your choice, and 
I think this is basically what I'd have to resort to using if I wanted to run 
xfs_repair on / on that box.

  Also in practice I find that any kernel crash or hard-power-off corrupts
  XFS to at least some extent requiring an xfs_check and xfs_repair, so I
  have to make sure to keep a LiveCD on hand to be able to do this.
 
 I've also heard about terrific stories of data lose after an unexpected
 power failure on volumes running on XFS but as I said before, I have no
 direct experience with this file system so I can't comment.

I've experienced some data loss due to unclean shutdowns, so I can verify that 
that's possible.  I have not yet had any massive data loss, thankfully.

...
  The main reason I've been running XFS is for speed -- even on top of
  LUKS I'm finding XFS is able to do sustained 40MB/s transfers over 1Gb
  ethernet, where ext4 on the same box is not able to sustain that.
  However ext4 is more reliable and easier to deal with, because it's able
  to run an fsck at boot time and without neeting a LiveCD to fix it.  ;-)
 
 Not bad numbers.

BTW these numbers are with 1Gb ethernet /without/ using jumbo frames -- this 
is because the unmanaged switch I'm using doesn't support them.

 In the event I give XFS a whirl it will be over my /data partition,
 that's for sure... and fortunately all of my system have UPS units on
 behind O:-)

All of the systems I'm using XFS on have UPSes on them too -- yet I find 
systems I run go down hard once in a blue moon, so IMHO a UPS won't completely 
save you from needing to run xfs_check and xfs_repair occasionally.

  -- Chris

--
Chris Knadle
chris.kna...@coredump.us
GPG Key: 4096R/0x1E759A726A9FDD74


signature.asc
Description: This is a digitally signed message part.


Re: 2TB USB hard drive for backing up

2012-05-04 Thread Martin Steigerwald
Am Donnerstag, 3. Mai 2012 schrieb Ellwood Blues:
 2012/4/30 Martin Steigerwald mar...@lichtvoll.de:
  Am Montag, 30. April 2012 schrieb Ellwood Blues:
  2012/4/30 Martin Steigerwald mar...@lichtvoll.de:
   Am Montag, 30. April 2012 schrieb Chris Bannister:
   On Sun, Apr 29, 2012 at 08:27:03PM +0200, Martin Steigerwald 
wrote:
Hmmm, I´d avoid those with 4 KB hardware sectors that lie to
the OS they have 512 byte sectors. Although I think even those
should work. But with 512 byte sectors you have a 2 TB limit
when you use MBR partitioning.

3 TB disks with 4 KB sectors both hardware (physical) and
software (logical) should just work, provided the Linux is new
enough.

On Squeeze use -cu as additional options (see manpage).
   
   Sorry for jumping in here, but I can't figure out (from your
   post) which command requires the additional options: -cu. Which
   manpage?
   
   fdisk. Sorry if I didn´t mention it anywhere in my post.
  
  Thanks, I've tried everything but not success. The problem is that
  the disk is already half full and aligned with WD tools. I am just
  waiting for linux to be able to read it and write it as efficiently
  as Windows does it, at the moment I am not able to read it, which
  is very frustrating.
  
  I would like to see some information from the disk, like
  
  - relevant stuff from hdparm -I /dev/yourdisk (feel free to skip
  serial number if you do not want to post it here)
  - fdisk -cul /dev/yourdisk
  - tail -fn0 /var/log/syslog / dmesg when the kernel detects the disk
  
  for starters.
  
  You need to use GPT if the disk reports 512 byte sectors to the OS.
  Thats no problem, when its just a data disk. Try gdisk on the disk.
 
 Sorry for the delay. Her you have what you've asked attached as files
 and as text on the message:
 
 
 /dev/sdf:
 
 ATA device, with non-removable media
   Model Number:   WDC WD30EZRX-00MMMB0
   Serial Number:  WD-WCAWZxxx
   Firmware Revision:  80.00A80
   Transport:  Serial, SATA 1.0a, SATA II Extensions, SATA Rev
 2.5, SATA Rev 2.6, SATA Rev 3.0
 Standards:
   Supported: 8 7 6 5
   Likely used: 8
 Configuration:
   Logical max current
   cylinders   16383   16383
   heads   16  16
   sectors/track   63  63
   --
   CHS current addressable sectors:   16514064
   LBAuser addressable sectors:  268435455
   LBA48  user addressable sectors: 5860533168
   Logical  Sector size:   512 bytes
   Physical Sector size:  4096 bytes
   Logical Sector-0 offset:  0 bytes
   device size with M = 1024*1024: 2861588 MBytes
   device size with M = 1000*1000: 3000592 MBytes (3000 GB)
[…]

So here we have 3 TB, but…

 May  3 18:18:50 relampago3 kernel: [ 3931.976021] usb 1-4: new
 high-speed USB device number 4 using ehci_hcd
 May  3 18:18:50 relampago3 kernel: [ 3932.108900] usb 1-4: New USB
 device found, idVendor=152d, idProduct=2329
 May  3 18:18:50 relampago3 kernel: [ 3932.108907] usb 1-4: New USB
 device strings: Mfr=1, Product=2, SerialNumber=5
 May  3 18:18:50 relampago3 kernel: [ 3932.108911] usb 1-4: Product:
 USB to ATA/ATAPI Bridge
 May  3 18:18:50 relampago3 kernel: [ 3932.108914] usb 1-4:
 Manufacturer: JMicron May  3 18:18:50 relampago3 kernel: [
 3932.108917] usb 1-4:
 SerialNumber: 152D20329000
 May  3 18:18:50 relampago3 kernel: [ 3932.110324] usb-storage 1-4:1.0:
 Quirks match for vid 152d pid 2329: 8020
 May  3 18:18:50 relampago3 kernel: [ 3932.110360] scsi7 : usb-storage
 1-4:1.0 May  3 18:18:50 relampago3 mtp-probe: checking bus 1, device
 4: /sys/devices/pci:00/:00:1a.7/usb1/1-4
 May  3 18:18:50 relampago3 mtp-probe: bus: 1, device: 4 was not an MTP
 device May  3 18:18:57 relampago3 hddtemp[1872]: /dev/sda: WDC
 WD2500JS-75NCB3: 34 C May  3 18:19:06 relampago3 kernel: [
 3948.470324] scsi 7:0:0:0: Direct-Access WDC WD30 EZRX-00MMMB0
  PQ: 0 ANSI: 2 CCS May  3 18:19:06 relampago3 kernel: [
 3948.472796] sd 7:0:0:0: [sdf] 1565565872 512-byte logical blocks:
 (801 GB/746 GiB)

… here we have 801 GB.

Now thats strange.

What kernel is this?

I think its an kernel issue, cause this is before any partitioning.

The device reports 512-byte logical blocks. But it would be nice to know 
the physical size as well. It might have 4096 byte there. If it has thats 
one of the disks I tell the students of my Linux trainings to through out 
of the window ;). It would report 512 byte to the OS and basically lie to 
it to stay compatible with older Windows versions, while using 4 KiB as 
physical hardware sector size. I recommend either 512/512 or 4096/4096, so 
both the same size.

I would try with 3.2 backport kernel in case you use Squeeze.

 fdisk: opción inválida -- c
 GNU Fdisk 1.2.4

This only works with util-linux fdisk not gnu-fdisk.

merkaba:/sys/block/sda fdisk -c -u -l /dev/sda

Disk /dev/sda: 

Re: 2TB USB hard drive for backing up

2012-05-04 Thread Ellwood Blues
2012/5/4 Martin Steigerwald mar...@lichtvoll.de:

 What kernel is this?

Debian Wheezy  3.2.0-1-686-pae



 This only works with util-linux fdisk not gnu-fdisk.

I have util-linux installed but I can't find the fdisk  you said
support -c switch.

 Ah. I would try a newer kernel. ;)

I've tried the newest on live-cds and on Wheezy but I don't see it working.

Cheers.


-- 
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org 
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: 
http://lists.debian.org/cagzenodoy6bkfcxew5nmkaid8jqa8mhfvgju6ph7jsw8d1z...@mail.gmail.com



Re: 2TB USB hard drive for backing up: XFS info

2012-05-04 Thread Chris Knadle
On Monday, April 30, 2012 10:53:46, Camaleón wrote:
 On Sun, 29 Apr 2012 20:16:58 +0200, Martin Steigerwald wrote:
  Am Dienstag, 24. April 2012 schrieb Camaleón:
...
  XFS might also have long file check times.
 
 I still have not tried this so I can't really tell but what I've heard on
 XFS is not comparable to ext3, I mean, in regards with the time it takes
 to perform a filesystem check.

On XFS, fsck is literally a no-op -- it does *not* fsck the filesystem.

The steps to actually fsck a XFS filesystem:
   - mount the XFS filesystem to replay the log
   - unmount the XFS filesystem that was just mounted
   - run xfs_check on the filesystem's device
   - if necessary run xfs_repair on the filesystem's device

Note this means running 'xfs_check' is done when the filesystem is not 
mounted.  _Supposedly_ it can also be run if the filesystem is mounted read-
only, but in practice I find it's best (and easier) to run the XFS commands 
from a LiveCD.  The xfs_check and xfs_repair operations are incredibly fast -- 
even for a 500GB filesystem it's usually only takes about 10 or 15 seconds.  
Speed is generally what XFS is good at, *except* when it comes to deletion of 
a large number of files -- that's where it's slow.

Also in practice I find that any kernel crash or hard-power-off corrupts XFS 
to at least some extent requiring an xfs_check and xfs_repair, so I have to 
make sure to keep a LiveCD on hand to be able to do this.

This gets even more interesting when running XFS on top of LUKS encryption.  
I'm currently doing that, and I have had to do an xfs_repair operation -- it 
involves running cryptsetup manually at the command line within a LiveCD and 
then running xfs_repair on the newly created unencrypted device.  [And of 
course you have to know to look and make sure the LiveCD contains those 
utilities.]  Definitely an interesting experience.

The main reason I've been running XFS is for speed -- even on top of LUKS I'm 
finding XFS is able to do sustained 40MB/s transfers over 1Gb ethernet, where 
ext4 on the same box is not able to sustain that.  However ext4 is more 
reliable and easier to deal with, because it's able to run an fsck at boot 
time and without neeting a LiveCD to fix it.  ;-)

  -- Chris

--
Chris Knadle
chris.kna...@coredump.us
GPG Key: 4096R/0x1E759A726A9FDD74


signature.asc
Description: This is a digitally signed message part.


Re: 2TB USB hard drive for backing up: XFS info

2012-05-04 Thread Camaleón
On Fri, 04 May 2012 15:08:57 -0400, Chris Knadle wrote:

 On Monday, April 30, 2012 10:53:46, Camaleón wrote:
 On Sun, 29 Apr 2012 20:16:58 +0200, Martin Steigerwald wrote:
  Am Dienstag, 24. April 2012 schrieb Camaleón:
 ...
  XFS might also have long file check times.
 
 I still have not tried this so I can't really tell but what I've heard
 on XFS is not comparable to ext3, I mean, in regards with the time it
 takes to perform a filesystem check.
 
 On XFS, fsck is literally a no-op -- it does *not* fsck the filesystem.

Well, at least there are xfsprogs.

 The steps to actually fsck a XFS filesystem:
- mount the XFS filesystem to replay the log 
- unmount the XFS filesystem that was just mounted 
- run xfs_check on the filesystem's device 
- if necessary run xfs_repair on the filesystem's device

You mean you don't even notice a file system inconsistency until it 
royally crashes or even something worse? Oh.

 Note this means running 'xfs_check' is done when the filesystem is not
 mounted.  _Supposedly_ it can also be run if the filesystem is mounted
 read- only, but in practice I find it's best (and easier) to run the XFS
 commands from a LiveCD.  The xfs_check and xfs_repair operations are
 incredibly fast -- even for a 500GB filesystem it's usually only takes
 about 10 or 15 seconds. Speed is generally what XFS is good at, *except*
 when it comes to deletion of a large number of files -- that's where
 it's slow.

So... is that you don't find it suitable for a standard / partition? 

I mean, if it's better don't analyze an XFS partition when is mounted 
read-only, that can be really a no-no for many installations running 
24/365.

 Also in practice I find that any kernel crash or hard-power-off corrupts
 XFS to at least some extent requiring an xfs_check and xfs_repair, so I
 have to make sure to keep a LiveCD on hand to be able to do this.

I've also heard about terrific stories of data lose after an unexpected 
power failure on volumes running on XFS but as I said before, I have no 
direct experience with this file system so I can't comment.

 This gets even more interesting when running XFS on top of LUKS
 encryption. I'm currently doing that, and I have had to do an xfs_repair
 operation -- it involves running cryptsetup manually at the command line
 within a LiveCD and then running xfs_repair on the newly created
 unencrypted device.  [And of course you have to know to look and make
 sure the LiveCD contains those utilities.]  Definitely an interesting
 experience.

File systems need a twist :-)

 The main reason I've been running XFS is for speed -- even on top of
 LUKS I'm finding XFS is able to do sustained 40MB/s transfers over 1Gb
 ethernet, where ext4 on the same box is not able to sustain that. 
 However ext4 is more reliable and easier to deal with, because it's able
 to run an fsck at boot time and without neeting a LiveCD to fix it.  ;-)

Not bad numbers. 

In the event I give XFS a whirl it will be over my /data partition, 
that's for sure... and fortunately all of my system have UPS units on 
behind O:-)

Greetings,

-- 
Camaleón


-- 
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org 
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: http://lists.debian.org/jo1hra$3du$1...@dough.gmane.org



Re: 2TB USB hard drive for backing up

2012-05-03 Thread Ellwood Blues
2012/4/30 Martin Steigerwald mar...@lichtvoll.de:
 Am Montag, 30. April 2012 schrieb Ellwood Blues:
 2012/4/30 Martin Steigerwald mar...@lichtvoll.de:
  Am Montag, 30. April 2012 schrieb Chris Bannister:
  On Sun, Apr 29, 2012 at 08:27:03PM +0200, Martin Steigerwald wrote:
   Hmmm, I´d avoid those with 4 KB hardware sectors that lie to the
   OS they have 512 byte sectors. Although I think even those should
   work. But with 512 byte sectors you have a 2 TB limit when you
   use MBR partitioning.
  
   3 TB disks with 4 KB sectors both hardware (physical) and software
   (logical) should just work, provided the Linux is new enough.
  
   On Squeeze use -cu as additional options (see manpage).
 
  Sorry for jumping in here, but I can't figure out (from your post)
  which command requires the additional options: -cu. Which manpage?
 
  fdisk. Sorry if I didn´t mention it anywhere in my post.

 Thanks, I've tried everything but not success. The problem is that the
 disk is already half full and aligned with WD tools. I am just waiting
 for linux to be able to read it and write it as efficiently as Windows
 does it, at the moment I am not able to read it, which is very
 frustrating.

 I would like to see some information from the disk, like

 - relevant stuff from hdparm -I /dev/yourdisk (feel free to skip serial
 number if you do not want to post it here)
 - fdisk -cul /dev/yourdisk
 - tail -fn0 /var/log/syslog / dmesg when the kernel detects the disk

 for starters.

 You need to use GPT if the disk reports 512 byte sectors to the OS. Thats
 no problem, when its just a data disk. Try gdisk on the disk.

Sorry for the delay. Her you have what you've asked attached as files
and as text on the message:


/dev/sdf:

ATA device, with non-removable media
Model Number:   WDC WD30EZRX-00MMMB0
Serial Number:  WD-WCAWZxxx
Firmware Revision:  80.00A80
Transport:  Serial, SATA 1.0a, SATA II Extensions, SATA Rev
2.5, SATA Rev 2.6, SATA Rev 3.0
Standards:
Supported: 8 7 6 5
Likely used: 8
Configuration:
Logical max current
cylinders   16383   16383
heads   16  16
sectors/track   63  63
--
CHS current addressable sectors:   16514064
LBAuser addressable sectors:  268435455
LBA48  user addressable sectors: 5860533168
Logical  Sector size:   512 bytes
Physical Sector size:  4096 bytes
Logical Sector-0 offset:  0 bytes
device size with M = 1024*1024: 2861588 MBytes
device size with M = 1000*1000: 3000592 MBytes (3000 GB)
cache/buffer size  = unknown
Capabilities:
LBA, IORDY(can be disabled)
Queue depth: 32
Standby timer values: spec'd by Standard, with device specific minimum
R/W multiple sector transfer: Max = 16  Current = 0
DMA: mdma0 mdma1 mdma2 udma0 udma1 udma2 udma3 udma4 udma5 *udma6
 Cycle time: min=120ns recommended=120ns
PIO: pio0 pio1 pio2 pio3 pio4
 Cycle time: no flow control=120ns  IORDY flow control=120ns
Commands/features:
Enabled Supported:
   *SMART feature set
Security Mode feature set
   *Power Management feature set
   *Write cache
   *Look-ahead
   *Host Protected Area feature set
   *WRITE_BUFFER command
   *READ_BUFFER command
   *NOP cmd
   *DOWNLOAD_MICROCODE
Power-Up In Standby feature set
   *SET_FEATURES required to spinup after power up
SET_MAX security extension
   *48-bit Address feature set
   *Device Configuration Overlay feature set
   *Mandatory FLUSH_CACHE
   *FLUSH_CACHE_EXT
   *SMART error logging
   *SMART self-test
   *General Purpose Logging feature set
   *64-bit World wide name
   *{READ,WRITE}_DMA_EXT_GPL commands
   *Segmented DOWNLOAD_MICROCODE
   *Gen1 signaling speed (1.5Gb/s)
   *Gen2 signaling speed (3.0Gb/s)
   *Gen3 signaling speed (6.0Gb/s)
   *Native Command Queueing (NCQ)
   *Host-initiated interface power management
   *Phy event counters
   *NCQ priority information
DMA Setup Auto-Activate optimization
   *Software settings preservation
   *SMART Command Transport (SCT) feature set
   *SCT LBA Segment Access (AC2)
   *SCT Features Control (AC4)
   *SCT Data Tables (AC5)
unknown 206[12] (vendor specific)
unknown 206[13] (vendor specific)
Security:
Master password revision code = 65534
supported
not enabled
not 

Re: 2TB USB hard drive for backing up

2012-05-01 Thread Ellwood Blues
2012/4/30 Martin Steigerwald mar...@lichtvoll.de:
 Am Montag, 30. April 2012 schrieb Ellwood Blues:
 2012/4/30 Martin Steigerwald mar...@lichtvoll.de:
  Am Montag, 30. April 2012 schrieb Chris Bannister:
  On Sun, Apr 29, 2012 at 08:27:03PM +0200, Martin Steigerwald wrote:
   Hmmm, I´d avoid those with 4 KB hardware sectors that lie to the
   OS they have 512 byte sectors. Although I think even those should
   work. But with 512 byte sectors you have a 2 TB limit when you
   use MBR partitioning.
  
   3 TB disks with 4 KB sectors both hardware (physical) and software
   (logical) should just work, provided the Linux is new enough.
  
   On Squeeze use -cu as additional options (see manpage).
 
  Sorry for jumping in here, but I can't figure out (from your post)
  which command requires the additional options: -cu. Which manpage?
 
  fdisk. Sorry if I didn´t mention it anywhere in my post.

 Thanks, I've tried everything but not success. The problem is that the
 disk is already half full and aligned with WD tools. I am just waiting
 for linux to be able to read it and write it as efficiently as Windows
 does it, at the moment I am not able to read it, which is very
 frustrating.

 I would like to see some information from the disk, like

 - relevant stuff from hdparm -I /dev/yourdisk (feel free to skip serial
 number if you do not want to post it here)
 - fdisk -cul /dev/yourdisk
 - tail -fn0 /var/log/syslog / dmesg when the kernel detects the disk

 for starters.

 You need to use GPT if the disk reports 512 byte sectors to the OS. Thats
 no problem, when its just a data disk. Try gdisk on the disk.

Thanks. I can't do it right now. I think tomorrow I'll send what you want.


--
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: 
http://lists.debian.org/cagzenofuddiq0kmagyfjnk_b41+1-envqplthb_zjszzh-r...@mail.gmail.com



Re: 2TB USB hard drive for backing up

2012-04-30 Thread Jon Dowland
On Sun, Apr 29, 2012 at 08:16:58PM +0200, Martin Steigerwald wrote:
 I use LVM quite a lot, but I am a bit annoyed by having to vgchange -ay 
 after inserting and vgchange -an before removal. I like to see a removable 
 flag for LVs that lessens the locking restrictions that are important in 
 certain enterprise setups, so that I can have an LVM be scanned 
 automatically and do not need to run vgchange.

We may one day be blessed by udisks handling this properly on hotplug, perhaps
with a suitably-configured policykit.


-- 
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org 
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: http://lists.debian.org/20120430133250.GG7795@debian



Re: 2TB USB hard drive for backing up

2012-04-30 Thread Camaleón
On Sun, 29 Apr 2012 20:16:58 +0200, Martin Steigerwald wrote:

 Am Dienstag, 24. April 2012 schrieb Camaleón:

(...)
 
 As an aside note, I've been using ReiserFS in all of my linux boxes (in
 both, servers and workstations) for the / partition and I never had
 to face the long waits at booting till the fsck ends. I know that
 ReiserFS (v3) is not actively developed (just security patches) and
 performs better for small files but I'm considering in moving ext3 to
 XFS or another mature file system (JFS) because the annoying and
 absurdly long time fsck takes.
 
 Reiserfs has some more drawbacks:
 
 - long mount times on large volumes
 
 - doesn´t distinguish between different filesystems on fsck like
   - loopback file with reiserfs on reiserfs 
   - vm image file with reiserfs on reiserfs
 
 Especially the latter is a no-go for me, cause an fsck would mix up the
 several reiserfs filesstems.

I haven't had experienced any of that issues in any of my systems so I 
still find ReaiserFS the most suitable filesystem for me.

The only big drawback I see for ReiserFS v3 is that it does not receive 
more enhancements nor features just patches, which can leave your 
installation in a very delicated state if something goes wrong. OTOH, I 
don't know about the current status of ReiserFS v4, I have not heard any 
news since many years :-?

 XFS might also have long file check times.

I still have not tried this so I can't really tell but what I've heard on 
XFS is not comparable to ext3, I mean, in regards with the time it takes 
to perform a filesystem check.

 fsck times depend more on the number of inodes than on the size of the
 filesystem. They also depend on the version of the check tool.
 
 XFS develovers tuned xfs_repair heavily regarding speed and memory
 usage. And I seem to recall some optimizations for Ext4 as well, like
 unitialized block groups (bguninit or something lile that in tunefs -l).
 
 My Ext4 fsck times on large volumes are quite good. A few minutes
 usually. Even on the backup drive which holds at least 1 million inodes
 in one logical volume.

My experience with ext3 and fsck times also varies depending of the 
volume file size: my 500 GiB hard disk can delay up to 5 min. but the 
server's 1.2 TiB volume do take longer...

  Its very convenient to have a large sack to toss the stuff, but it
  has its own set of drawbacks :)
 
 Yes, and I'm against partitioning too much but considering the big
 sizes of the actual hard disks it has become a must, also to minimize a
 filesystem corruption in a partition that can affect the whole data.
 
 I use LVM quite a lot, but I am a bit annoyed by having to vgchange -ay
 after inserting and vgchange -an before removal. I like to see a
 removable flag for LVs that lessens the locking restrictions that are
 important in certain enterprise setups, so that I can have an LVM be
 scanned automatically and do not need to run vgchange.

I wish hard disk partitioning becomes more handly and flexible and it 
gets implemented on the filesystem natively without having to deal with 
external tools, such as LVM. 

Greetings,

-- 
Camaleón


-- 
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org 
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: http://lists.debian.org/jnm91q$kbo$3...@dough.gmane.org



Re: 2TB USB hard drive for backing up

2012-04-30 Thread Camaleón
On Sun, 29 Apr 2012 20:21:52 +0200, Martin Steigerwald wrote:

 Am Dienstag, 24. April 2012 schrieb Camaleón:
  While I wait for the new 2 TB HDD, I am copying data across my home
  LAN to my laptop which has a 120 GB HDD. I was using the WiFi network
  provided by the computer consultants downstairs: it gave me 1.2 MB/s
  transfer rate. When I succeeded in getting the Ethernet working, I
  got 9 MB/s. I've now checked the ethernet NICs at both ends and they
  are both Gigabit devices. So I've ordered a 5m Cat5e ethernet cable
  from Amazon which should give me 120 MB/s transfer rate. That would
  certainly make it convenient for security backups.
 
 Yes, and that's why my primary backup is directly attached to one of
 the  SATA ports of the motherboard: speed between SATA buses is good
 enough when using a powerful system. The average size of my /home
 backup is now at ~10 GiB and I always do full (not incremental nor
 differential) backups becasue 10 GiB is still a manageable size
 (although for servers and workstations I do differential backups).
 
 I use esata disks for my laptop backups. Together with
 Cardbus-/Expresscard esata controllers, even for the ThinkPad T520 cause
 its Intel only version didn´t sport an esata connector.
 
 SATA flies. A bus where practically possible speeds come quite close to
 the theoretic maximum. Unlike USB 2 and currently it seems also USB 3.

That's another option but again, you need to rely on the ExpressCard 
capabilities (v1.2 or v2.0?) and its performance that can slim the whole 
throughout.

 Initial rsync based backup of 300 GB Intel SSD 320 to new 2 TB Hitachi
 harddisk has been 1 hour. For about 200 GiB of data, at least one
 million inodes, including a virtual server on the internet and the
 Debian on the USB-Stick in my ASUS WL-500g Premium DSL router.
 
 Current times are usually less than 10 minutes for differential backups.

Well, yes, differential backups shorten the backup process time (though 
incremental copies are even faster).

Greetings,

-- 
Camaleón


-- 
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org 
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: http://lists.debian.org/jnma07$kbo$4...@dough.gmane.org



Re: 2TB USB hard drive for backing up

2012-04-30 Thread Jon Dowland
On Mon, Apr 30, 2012 at 02:53:46PM +, Camaleón wrote:
 I haven't had experienced any of that issues in any of my systems so I still
 find ReaiserFS the most suitable filesystem for me.

As long as you don't have any VM images using reiserfs v3 on top of a reiserfs
3 filesystem, you're probably OK.

 OTOH, I don't know about the current status of ReiserFS v4, I have not heard
 any news since many years :-?

It's essentially dead, unfortunately.  Once the main author was convicted of
murder, another maintainer took over, but nobody is interested enough to keep
it maintained as part of the main kernel.


-- 
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org 
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: http://lists.debian.org/20120430155033.GN7795@debian



Re: 2TB USB hard drive for backing up

2012-04-30 Thread Martin Steigerwald
Am Montag, 30. April 2012 schrieb Chris Bannister:
 On Sun, Apr 29, 2012 at 08:27:03PM +0200, Martin Steigerwald wrote:
  Hmmm, I´d avoid those with 4 KB hardware sectors that lie to the OS
  they have 512 byte sectors. Although I think even those should work.
  But with 512 byte sectors you have a 2 TB limit when you use MBR
  partitioning.
  
  3 TB disks with 4 KB sectors both hardware (physical) and software
  (logical) should just work, provided the Linux is new enough.
  
  On Squeeze use -cu as additional options (see manpage).
 
 Sorry for jumping in here, but I can't figure out (from your post)
 which command requires the additional options: -cu. Which manpage?

fdisk. Sorry if I didn´t mention it anywhere in my post.

-- 
Martin 'Helios' Steigerwald - http://www.Lichtvoll.de
GPG: 03B0 0D6C 0040 0710 4AFA  B82F 991B EAAC A599 84C7


--
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: http://lists.debian.org/201204301814.21525.mar...@lichtvoll.de



Re: 2TB USB hard drive for backing up

2012-04-30 Thread Ellwood Blues
2012/4/30 Martin Steigerwald mar...@lichtvoll.de:
 Am Montag, 30. April 2012 schrieb Chris Bannister:
 On Sun, Apr 29, 2012 at 08:27:03PM +0200, Martin Steigerwald wrote:
  Hmmm, I´d avoid those with 4 KB hardware sectors that lie to the OS
  they have 512 byte sectors. Although I think even those should work.
  But with 512 byte sectors you have a 2 TB limit when you use MBR
  partitioning.
 
  3 TB disks with 4 KB sectors both hardware (physical) and software
  (logical) should just work, provided the Linux is new enough.
 
  On Squeeze use -cu as additional options (see manpage).

 Sorry for jumping in here, but I can't figure out (from your post)
 which command requires the additional options: -cu. Which manpage?

 fdisk. Sorry if I didn´t mention it anywhere in my post.

Thanks, I've tried everything but not success. The problem is that the
disk is already half full and aligned with WD tools. I am just waiting
for linux to be able to read it and write it as efficiently as Windows
does it, at the moment I am not able to read it, which is very
frustrating.


--
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: 
http://lists.debian.org/CAGZeNoc82T0jS=sg3d2umqi6pjawjokjqk2rbuk+qfazn3p...@mail.gmail.com



Re: 2TB USB hard drive for backing up

2012-04-30 Thread Martin Steigerwald
Am Montag, 30. April 2012 schrieb Ellwood Blues:
 2012/4/30 Martin Steigerwald mar...@lichtvoll.de:
  Am Montag, 30. April 2012 schrieb Chris Bannister:
  On Sun, Apr 29, 2012 at 08:27:03PM +0200, Martin Steigerwald wrote:
   Hmmm, I´d avoid those with 4 KB hardware sectors that lie to the
   OS they have 512 byte sectors. Although I think even those should
   work. But with 512 byte sectors you have a 2 TB limit when you
   use MBR partitioning.
   
   3 TB disks with 4 KB sectors both hardware (physical) and software
   (logical) should just work, provided the Linux is new enough.
   
   On Squeeze use -cu as additional options (see manpage).
  
  Sorry for jumping in here, but I can't figure out (from your post)
  which command requires the additional options: -cu. Which manpage?
  
  fdisk. Sorry if I didn´t mention it anywhere in my post.
 
 Thanks, I've tried everything but not success. The problem is that the
 disk is already half full and aligned with WD tools. I am just waiting
 for linux to be able to read it and write it as efficiently as Windows
 does it, at the moment I am not able to read it, which is very
 frustrating.

I would like to see some information from the disk, like 

- relevant stuff from hdparm -I /dev/yourdisk (feel free to skip serial 
number if you do not want to post it here)
- fdisk -cul /dev/yourdisk
- tail -fn0 /var/log/syslog / dmesg when the kernel detects the disk

for starters.

You need to use GPT if the disk reports 512 byte sectors to the OS. Thats 
no problem, when its just a data disk. Try gdisk on the disk.

-- 
Martin 'Helios' Steigerwald - http://www.Lichtvoll.de
GPG: 03B0 0D6C 0040 0710 4AFA  B82F 991B EAAC A599 84C7


--
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: http://lists.debian.org/201205010044.24643.mar...@lichtvoll.de



Re: 2TB USB hard drive for backing up

2012-04-29 Thread Martin Steigerwald
Am Dienstag, 24. April 2012 schrieb Camaleón:
 On Tue, 24 Apr 2012 13:20:05 +0200, Alberto Fuentes wrote:
  On 21/04/12 19:34, Camaleón wrote:
  I'm against big partitions (500 GiB is the limit I have auto-imposed
  to my systems) so I would make 4 slices and spread the data over
  them. Anyway, I don't think you are going to have any problem to
  manage a single partition of 2 TiB, even more if you plain to store
  plain data (not a bootable system) there.
  
  I have ext4 in one partition of 1.8 Tb or so. It takes about 40 min
  to chkfs... just something to bear in mind...
 
 Good point :-)
 
 As an aside note, I've been using ReiserFS in all of my linux boxes (in
 both, servers and workstations) for the / partition and I never had
 to face the long waits at booting till the fsck ends. I know that
 ReiserFS (v3) is not actively developed (just security patches) and
 performs better for small files but I'm considering in moving ext3 to
 XFS or another mature file system (JFS) because the annoying and
 absurdly long time fsck takes.

Reiserfs has some more drawbacks:

- long mount times on large volumes

- doesn´t distinguish between different filesystems on fsck like
  - loopback file with reiserfs on reiserfs
  - vm image file with reiserfs on reiserfs

Especially the latter is a no-go for me, cause an fsck would mix up the 
several reiserfs filesstems.

XFS might also have long file check times.

fsck times depend more on the number of inodes than on the size of the 
filesystem. They also depend on the version of the check tool.

XFS develovers tuned xfs_repair heavily regarding speed and memory usage. 
And I seem to recall some optimizations for Ext4 as well, like unitialized 
block groups (bguninit or something lile that in tunefs -l).

My Ext4 fsck times on large volumes are quite good. A few minutes usually. 
Even on the backup drive which holds at least 1 million inodes in one 
logical volume.

  Its very convenient to have a large sack to toss the stuff, but it
  has its own set of drawbacks :)
 
 Yes, and I'm against partitioning too much but considering the big
 sizes of the actual hard disks it has become a must, also to minimize
 a filesystem corruption in a partition that can affect the whole data.

I use LVM quite a lot, but I am a bit annoyed by having to vgchange -ay 
after inserting and vgchange -an before removal. I like to see a removable 
flag for LVs that lessens the locking restrictions that are important in 
certain enterprise setups, so that I can have an LVM be scanned 
automatically and do not need to run vgchange.

-- 
Martin 'Helios' Steigerwald - http://www.Lichtvoll.de
GPG: 03B0 0D6C 0040 0710 4AFA  B82F 991B EAAC A599 84C7


--
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: http://lists.debian.org/201204292016.59038.mar...@lichtvoll.de



Re: 2TB USB hard drive for backing up

2012-04-29 Thread Martin Steigerwald
Am Dienstag, 24. April 2012 schrieb Camaleón:
  While I wait for the new 2 TB HDD, I am copying data across my home
  LAN to my laptop which has a 120 GB HDD. I was using the WiFi
  network provided by the computer consultants downstairs: it gave me
  1.2 MB/s transfer rate. When I succeeded in getting the Ethernet
  working, I got 9 MB/s. I've now checked the ethernet NICs at both
  ends and they are both Gigabit devices. So I've ordered a 5m Cat5e
  ethernet cable from Amazon which should give me 120 MB/s transfer
  rate. That would certainly make it convenient for security backups.
 
 Yes, and that's why my primary backup is directly attached to one of
 the  SATA ports of the motherboard: speed between SATA buses is good
 enough when using a powerful system. The average size of my /home
 backup is now at ~10 GiB and I always do full (not incremental nor
 differential) backups becasue 10 GiB is still a manageable size
 (although for servers and workstations I do differential backups).

I use esata disks for my laptop backups. Together with 
Cardbus-/Expresscard esata controllers, even for the ThinkPad T520 cause 
its Intel only version didn´t sport an esata connector.

SATA flies. A bus where practically possible speeds come quite close to the 
theoretic maximum. Unlike USB 2 and currently it seems also USB 3.

Initial rsync based backup of 300 GB Intel SSD 320 to new 2 TB Hitachi 
harddisk has been 1 hour. For about 200 GiB of data, at least one million 
inodes, including a virtual server on the internet and the Debian on the 
USB-Stick in my ASUS WL-500g Premium DSL router.

Current times are usually less than 10 minutes for differential backups.

-- 
Martin 'Helios' Steigerwald - http://www.Lichtvoll.de
GPG: 03B0 0D6C 0040 0710 4AFA  B82F 991B EAAC A599 84C7


--
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: http://lists.debian.org/201204292021.52282.mar...@lichtvoll.de



Re: 2TB USB hard drive for backing up

2012-04-29 Thread Martin Steigerwald
Am Samstag, 21. April 2012 schrieb tv.deb...@googlemail.com:
 On 21/04/2012 18:36, Sian Mountbatten wrote:
  I am going to buy a big USB hard drive to backup films, music and
  photos. In fact, I might backup the system (at least, the list of
  packages installed).
  
  Anybody out there tried a 2TB USB hard drive? What did you use as
  filing system. I am intending to use ext4. Any comments?
 
 Hi, I use seagate's ST2000DL003-9VT1 (it's some kind of eco drive,
 but  still fast) for that purpose, in raid1. No complains. I have had
 disastrous experiences with WD 2TB drives (those green monstrosity).
 I use gpt partition table, with LUKS and RAID without trouble.

Two of them with 1,5 TB still running as RAID 1 on the workstation at 
work.

-- 
Martin 'Helios' Steigerwald - http://www.Lichtvoll.de
GPG: 03B0 0D6C 0040 0710 4AFA  B82F 991B EAAC A599 84C7


-- 
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org 
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: http://lists.debian.org/201204292022.50499.mar...@lichtvoll.de



Re: 2TB USB hard drive for backing up

2012-04-29 Thread Martin Steigerwald
Am Dienstag, 24. April 2012 schrieb Ellwood Blues:
 2012/4/21 Sian Mountbatten poenik...@fastmail.co.uk:
  I am going to buy a big USB hard drive to backup films, music and
  photos. In fact, I might backup the system (at least, the list of
  packages installed).
  
  Anybody out there tried a 2TB USB hard drive? What did you use as
  filing system. I am intending to use ext4. Any comments?

 If you want to be happy and trouble-free, try to avoid «Advanced
 Format» disks. I have 2TB disks and 1.5TB too that work perfectly but
 I've made the mistake to buy a 3TB WD Caviar Green («Advanced Format»)
 which I can only use on Windows as a 3TB disk, Linux only detects
 about 780MB if I remember well. Try to not make the same mistake.

Hmmm, I´d avoid those with 4 KB hardware sectors that lie to the OS they 
have 512 byte sectors. Although I think even those should work. But with 
512 byte sectors you have a 2 TB limit when you use MBR partitioning.

3 TB disks with 4 KB sectors both hardware (physical) and software 
(logical) should just work, provided the Linux is new enough.

On Squeeze use -cu as additional options (see manpage).

-- 
Martin 'Helios' Steigerwald - http://www.Lichtvoll.de
GPG: 03B0 0D6C 0040 0710 4AFA  B82F 991B EAAC A599 84C7


--
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: http://lists.debian.org/201204292027.04152.mar...@lichtvoll.de



Re: 2TB USB hard drive for backing up

2012-04-29 Thread Chris Bannister
On Sun, Apr 29, 2012 at 08:27:03PM +0200, Martin Steigerwald wrote:
 Hmmm, I´d avoid those with 4 KB hardware sectors that lie to the OS they 
 have 512 byte sectors. Although I think even those should work. But with 
 512 byte sectors you have a 2 TB limit when you use MBR partitioning.
 
 3 TB disks with 4 KB sectors both hardware (physical) and software 
 (logical) should just work, provided the Linux is new enough.
 
 On Squeeze use -cu as additional options (see manpage).

Sorry for jumping in here, but I can't figure out (from your post) which
command requires the additional options: -cu. Which manpage?

-- 
Religion is excellent stuff for keeping common people quiet.
   -- Napoleon Bonaparte


-- 
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org 
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: http://lists.debian.org/20120430052304.GD1947@tal



Re: 2TB USB hard drive for backing up

2012-04-25 Thread Sian Mountbatten

On 25/04/12 00:58, debian-user@lists.debian.org wrote:
Drive has now arrived. Creating a single partition and formatting it to ext4
was a doddle. Have now backed up 15GB of films and over 4GB of music.
Just used the commands
   cp -av /opt/music /mnt
   cp -av /opt/iso /mnt
Now I shall copy the compressed tarballs of my daily backup using the
command
   cp -av /opt/cdrw /mnt
Simple really.

__
Type your response ABOVE THIS LINE to reply


Re: 2TB USB hard drive for backing up

*1QA4xxx2a*  | APR 24, 2012  |  11:57PM UTC
Thank you for submitting your request. We have received your request 
and are working on responding to you as soon as possible. If you have 
any additional information to add to this case, please reply to this 
email.


Thanks in advance for your patience and support.
This message was sent to poenik...@fastmail.co.uk in reference to Case 
#79728.


[[d05c73fe3a70dd2b1f417f53be3258203ab0563f-1700709]] 



--
Sian Mountbatten
ex-Algol 68 specialist



[going OT]Re: 2TB USB hard drive for backing up

2012-04-24 Thread tv.deb...@googlemail.com

On 24/04/2012 04:02, Philipp Schneider wrote:

On Sat, 21 Apr 2012 21:50:02 +0200, tv.deb...@googlemail.com wrote:


Hi, I use seagate's ST2000DL003-9VT1 (it's some kind of eco drive, but
still fast) for that purpose, in raid1. No complains. I have had
disastrous experiences with WD 2TB drives (those green monstrosity). I
use gpt partition table, with LUKS and RAID without trouble.


What kind of disastrous experiences did you have with the WD drives?




I was a happy WD user with their caviar black line, until I started 
getting drives which where either faulty out of the box or quickly 
dying on me (they where 500GB and 1TB). Arguably every manufacturer as 
it shares of defective drives, during the same period I was also bitten 
by a Seagate firmware bug (quickly patched) rendering the drives 
suddenly unusable at boot time without any prior warning, just to keep 
things in fair perspective.
After those bad batches I changed my WD drives for Samsung's, they were 
both cheap and efficient, and since they are still running I can now say 
reliable too.
Later I felt the need to try so called green drive for use in external 
enclosures and NAS, due to their lower power consumption and operating 
temperature. I gave WD caviar green EAR* series a shot (1TB and later 
2TB), long story short the performances where very poor, the 
implementation of advanced format would render them unusable 
(incompatible) with most NAS, and tricky to partition in Linux. Even 
when properly running the operating temperature wasn't a lot different 
from my Samsung's drive, the performance far behind, the price higher, 
and the reliability no better. In raid the drives would be regularly 
kicked out of the volume, for no apparent reason. After loosing time in 
forum and tech support I sent the drives back, got Seagate's Eco green 
as a replacement (seagate bought samsung's disk business around that 
time), and they are running fine where the WD weren't.


I am no google, on my comparatively very small yearly drive consumption 
every bad series becomes a statistical monster, it would probably not be 
noticeable on a bigger scale. Still search engines aren't short on WD 
green bad stories.



--
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org 
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org

Archive: http://lists.debian.org/4f96650e.1090...@googlemail.com



Re: 2TB USB hard drive for backing up

2012-04-24 Thread Alberto Fuentes

On 21/04/12 19:34, Camaleón wrote:

I'm against big partitions (500 GiB is the limit I have auto-imposed to
my systems) so I would make 4 slices and spread the data over them.
Anyway, I don't think you are going to have any problem to manage a
single partition of 2 TiB, even more if you plain to store plain data
(not a bootable system) there.



I have ext4 in one partition of 1.8 Tb or so. It takes about 40 min to 
chkfs... just something to bear in mind...


Its very convenient to have a large sack to toss the stuff, but it has 
its own set of drawbacks :)


greets!
aL


--
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org 
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org

Archive: http://lists.debian.org/4f968c65.4070...@qindel.com



Re: 2TB USB hard drive for backing up

2012-04-24 Thread Camaleón
On Tue, 24 Apr 2012 13:20:05 +0200, Alberto Fuentes wrote:

 On 21/04/12 19:34, Camaleón wrote:
 I'm against big partitions (500 GiB is the limit I have auto-imposed to
 my systems) so I would make 4 slices and spread the data over them.
 Anyway, I don't think you are going to have any problem to manage a
 single partition of 2 TiB, even more if you plain to store plain data
 (not a bootable system) there.
 
 
 I have ext4 in one partition of 1.8 Tb or so. It takes about 40 min to
 chkfs... just something to bear in mind...

Good point :-)

As an aside note, I've been using ReiserFS in all of my linux boxes (in 
both, servers and workstations) for the / partition and I never had to 
face the long waits at booting till the fsck ends. I know that ReiserFS 
(v3) is not actively developed (just security patches) and performs 
better for small files but I'm considering in moving ext3 to XFS or 
another mature file system (JFS) because the annoying and absurdly long 
time fsck takes.
 
 Its very convenient to have a large sack to toss the stuff, but it has
 its own set of drawbacks :)

Yes, and I'm against partitioning too much but considering the big sizes 
of the actual hard disks it has become a must, also to minimize a  
filesystem corruption in a partition that can affect the whole data.

Greetings,

-- 
Camaleón


-- 
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org 
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: http://lists.debian.org/jn69je$gop$1...@dough.gmane.org



Re: 2TB USB hard drive for backing up

2012-04-24 Thread Sian Mountbatten

On 24/04/12 14:30, Camaleón wrote:

On Tue, 24 Apr 2012 13:20:05 +0200, Alberto Fuentes wrote:


On 21/04/12 19:34, Camaleón wrote:

snip

Yes, and I'm against partitioning too much but considering the big sizes
of the actual hard disks it has become a must, also to minimize a
filesystem corruption in a partition that can affect the whole data.


I'm also against partitioning too much. I've had very little trouble 
with hard disk drives over the years; I've certainly had no crashes or 
file corruption.


Not so long ago, my new SSD failed and I had to re-install the system. 
Unfortunately, I had not backed up all my data, so I lost some.


While I wait for the new 2 TB HDD, I am copying data across my home LAN 
to my laptop which has a 120 GB HDD. I was using the WiFi network 
provided by the computer consultants downstairs: it gave me 1.2 MB/s 
transfer rate. When I succeeded in getting the Ethernet working, I got 9 
MB/s. I've now checked the ethernet NICs at both ends and they are both 
Gigabit devices. So I've ordered a 5m Cat5e ethernet cable from Amazon 
which should give me 120 MB/s transfer rate. That would certainly make 
it convenient for security backups.


I've also installed ukopp which looks as though it will do incremental 
backups.


I intend using the 2 TB HDD for three purposes:-
 1. Full security backups
 2. Incremental security backups
 3. Archival backups

Last night, I split two film files into 4 GB pieces and used Brasero to 
write the pieces to 4 DVD+RW discs. So that's another, safe, way of 
backing up my data. I am now going to write my 4.3 GB music collection 
to another DVD+RW disc.


Greetings
--
Sian Mountbatten
ex-Algol 68 specialist


--
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: http://lists.debian.org/jn6apu$760$1...@speranza.aioe.org



Re: 2TB USB hard drive for backing up

2012-04-24 Thread Camaleón
On Tue, 24 Apr 2012 14:45:34 +0100, Sian Mountbatten wrote:

 On 24/04/12 14:30, Camaleón wrote:
 On Tue, 24 Apr 2012 13:20:05 +0200, Alberto Fuentes wrote:

 On 21/04/12 19:34, Camaleón wrote:
 snip
 Yes, and I'm against partitioning too much but considering the big
 sizes of the actual hard disks it has become a must, also to minimize a
 filesystem corruption in a partition that can affect the whole data.
 
 I'm also against partitioning too much. I've had very little trouble
 with hard disk drives over the years; I've certainly had no crashes or
 file corruption.

Yes, but it's dangerous. By not partitioning a tebibyte hard disk we're 
exposing gratituously our beloved data.
 
 Not so long ago, my new SSD failed and I had to re-install the system.
 Unfortunately, I had not backed up all my data, so I lost some.

The first thing a computer user has to learn (and I'm talking in general, 
not specifically to you) is backup, backup and backup. Then you can 
start to play and enjoy your system. People tend to treat their computers 
like washing machines and heck, not, is not the same :-)

 While I wait for the new 2 TB HDD, I am copying data across my home LAN
 to my laptop which has a 120 GB HDD. I was using the WiFi network
 provided by the computer consultants downstairs: it gave me 1.2 MB/s
 transfer rate. When I succeeded in getting the Ethernet working, I got 9
 MB/s. I've now checked the ethernet NICs at both ends and they are both
 Gigabit devices. So I've ordered a 5m Cat5e ethernet cable from Amazon
 which should give me 120 MB/s transfer rate. That would certainly make
 it convenient for security backups.

Yes, and that's why my primary backup is directly attached to one of the 
SATA ports of the motherboard: speed between SATA buses is good enough 
when using a powerful system. The average size of my /home backup is now 
at ~10 GiB and I always do full (not incremental nor differential) 
backups becasue 10 GiB is still a manageable size (although for servers 
and workstations I do differential backups).

 I've also installed ukopp which looks as though it will do incremental
 backups.
 
 I intend using the 2 TB HDD for three purposes:-
   1. Full security backups
   2. Incremental security backups
   3. Archival backups
 
 Last night, I split two film files into 4 GB pieces and used Brasero to
 write the pieces to 4 DVD+RW discs. So that's another, safe, way of
 backing up my data. I am now going to write my 4.3 GB music collection
 to another DVD+RW disc.

Yes, but I'm very reluctant when it comes to optical media: it's painly 
slow, hard to deal with (requires from extra software) and data on them 
does not last forever despite what advertising says (moreover, 
rewritable media has a limited number of writes).

Greetings,

-- 
Camaleón


-- 
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org 
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: http://lists.debian.org/jn6fbq$got$3...@dough.gmane.org



Re: 2TB USB hard drive for backing up

2012-04-24 Thread Jon Dowland
On Tue, Apr 24, 2012 at 02:45:34PM +0100, Sian Mountbatten wrote:
 Last night, I split two film files into 4 GB pieces and used Brasero
 to write the pieces to 4 DVD+RW discs. So that's another, safe, way
 of backing up my data. I am now going to write my 4.3 GB music
 collection to another DVD+RW disc.

Make sure to burn two copies of each disc and verify each disc, too. My
experience with brasero recently has been very poor.


-- 
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org 
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: http://lists.debian.org/20120424173416.GA14891@debian



Re: 2TB USB hard drive for backing up

2012-04-24 Thread Ellwood Blues
If you want to be happy and trouble-free, try to avoid «Advanced
Format» disks. I have 2TB disks and 1.5TB too that work perfectly but
I've made the mistake to buy a 3TB WD Caviar Green («Advanced Format»)
which I can only use on Windows as a 3TB disk, Linux only detects
about 780MB if I remember well. Try to not make the same mistake.



2012/4/21 Sian Mountbatten poenik...@fastmail.co.uk:
 I am going to buy a big USB hard drive to backup films, music and photos. In
 fact, I might backup the system (at least, the list of packages installed).

 Anybody out there tried a 2TB USB hard drive? What did you use as filing
 system. I am intending to use ext4. Any comments?
 --
 Sian Mountbatten
 ex-Algol 68 specialist


 --
 To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org with a subject
 of unsubscribe. Trouble? Contact listmas...@lists.debian.org
 Archive: http://lists.debian.org/jmunm9$90r$2...@speranza.aioe.org



--
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: 
http://lists.debian.org/CAGZeNodQTkb0osZKhbttr8n9Q6sNmLAZLSv=fp82nbngqku...@mail.gmail.com



Re: 2TB USB hard drive for backing up

2012-04-24 Thread Sian Mountbatten

On 24/04/12 22:10, Ellwood Blues wrote:

If you want to be happy and trouble-free, try to avoid «Advanced
Format» disks. I have 2TB disks and 1.5TB too that work perfectly but
I've made the mistake to buy a 3TB WD Caviar Green («Advanced Format»)
which I can only use on Windows as a 3TB disk, Linux only detects
about 780MB if I remember well. Try to not make the same mistake.



2012/4/21 Sian Mountbattenpoenik...@fastmail.co.uk:

I am going to buy a big USB hard drive to backup films, music and photos. In
fact, I might backup the system (at least, the list of packages installed).

Anybody out there tried a 2TB USB hard drive? What did you use as filing
system. I am intending to use ext4. Any comments?
--
Sian Mountbatten
ex-Algol 68 specialist


--
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org with a subject
of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: http://lists.debian.org/jmunm9$90r$2...@speranza.aioe.org

The 2TB HDD is not mentioned as an Advanced Format drive. I gather that 
it will probably have a VFAT file system on it which I should be able to 
remove by repartitioning the drive. Thank you for your comment. I don't 
use Windows in any shape or form, so Windows access is not for me.


Regards
--
Sian Mountbatten
ex-Algol 68 specialist


--
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org 
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org

Archive: http://lists.debian.org/jn7dia$6bp$1...@speranza.aioe.org



Re: 2TB USB hard drive for backing up

2012-04-23 Thread Philipp Schneider
On Sat, 21 Apr 2012 21:50:02 +0200, tv.deb...@googlemail.com wrote:

 Hi, I use seagate's ST2000DL003-9VT1 (it's some kind of eco drive, but
 still fast) for that purpose, in raid1. No complains. I have had
 disastrous experiences with WD 2TB drives (those green monstrosity). I
 use gpt partition table, with LUKS and RAID without trouble.

What kind of disastrous experiences did you have with the WD drives?


-- 
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org 
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: http://lists.debian.org/jn51kc$3mr$1...@dont-email.me



Re: 2TB USB hard drive for backing up

2012-04-22 Thread Jon Dowland
On Sat, Apr 21, 2012 at 05:36:28PM +0100, Sian Mountbatten wrote:
 I am going to buy a big USB hard drive to backup films, music and
 photos. In fact, I might backup the system (at least, the list of
 packages installed).
 
 Anybody out there tried a 2TB USB hard drive? What did you use as
 filing system. I am intending to use ext4. Any comments?

You are conflating backup and filing here.  If you are using it for
backup, set up a large partition (larger than the source partition)
and use something like rdiff-backup, bup or duplicity to back up
your source machine(s) onto the disk. Don't file onto the disk, file
on your computer and use the disk for backup.

Or, if you are to use the disc for filing, be aware that it can fail
just as an internal disc can.

If you really want to use the drive for both backup and filing
(perhaps you have more multimedia than you can fit on your computer's
internal storage), I'd recommend using seperate LVM logical volumes,
one purely for backup and the other for files, and still be aware
that you haven't solved the backup issue for the filing partition.
(maybe you can burn those to disc and divide up your file storage into
directories corresponding to the discs the backups reside on; a tool
such as 'datapacker' could help you with this)


-- 
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org 
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: http://lists.debian.org/20120422200158.GC11237@debian



2TB USB hard drive for backing up

2012-04-21 Thread Sian Mountbatten
I am going to buy a big USB hard drive to backup films, music and 
photos. In fact, I might backup the system (at least, the list of 
packages installed).


Anybody out there tried a 2TB USB hard drive? What did you use as filing 
system. I am intending to use ext4. Any comments?

--
Sian Mountbatten
ex-Algol 68 specialist


--
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org 
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org

Archive: http://lists.debian.org/jmunm9$90r$2...@speranza.aioe.org



Re: 2TB USB hard drive for backing up

2012-04-21 Thread Camaleón
On Sat, 21 Apr 2012 17:36:28 +0100, Sian Mountbatten wrote:

 I am going to buy a big USB hard drive to backup films, music and
 photos. In fact, I might backup the system (at least, the list of
 packages installed).

If you're using the disk to make a backup for one machine, USB (better 
3.0) can be convenient and easy but despite the interface, have you 
considered in buying a good external USB enclosure -with external power- 
and a separate hard disk? I find it to be a better solution than those 
ready-made-and-closed external tera-disks.

 Anybody out there tried a 2TB USB hard drive? What did you use as filing
 system. I am intending to use ext4. Any comments?

For home, I use an internal SATA disk (500 GiB) as primary backup (ext3) 
and then a NAS system as a secondary (backup for the backup) level of 
security, also formatted with ext3 which is the best option available for 
the NAS (only ext2, ext3 or FAT32 is allowed ;-) ).

I'm against big partitions (500 GiB is the limit I have auto-imposed to 
my systems) so I would make 4 slices and spread the data over them. 
Anyway, I don't think you are going to have any problem to manage a 
single partition of 2 TiB, even more if you plain to store plain data 
(not a bootable system) there.

Greetings,

-- 
Camaleón


-- 
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org 
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org
Archive: http://lists.debian.org/jmur37$g9$2...@dough.gmane.org



Re: 2TB USB hard drive for backing up

2012-04-21 Thread tv.deb...@googlemail.com

On 21/04/2012 18:36, Sian Mountbatten wrote:

I am going to buy a big USB hard drive to backup films, music and
photos. In fact, I might backup the system (at least, the list of
packages installed).

Anybody out there tried a 2TB USB hard drive? What did you use as filing
system. I am intending to use ext4. Any comments?


Hi, I use seagate's ST2000DL003-9VT1 (it's some kind of eco drive, but 
still fast) for that purpose, in raid1. No complains. I have had 
disastrous experiences with WD 2TB drives (those green monstrosity).

I use gpt partition table, with LUKS and RAID without trouble.



--
To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org 
with a subject of unsubscribe. Trouble? Contact listmas...@lists.debian.org

Archive: http://lists.debian.org/4f930e50.60...@googlemail.com