RAID mapper device size wrong after replacing drives

2007-12-06 Thread Ian P

Hi,

I have a problem with my RAID array under Linux after upgrading to larger
drives. I have a machine with Windows and Linux dual-boot which had a pair
of 160GB drives in a RAID-1 mirror with 3 partitions: partiton 1 = Windows
boot partition (FAT32), partiton 2 = Linux /boot (ext3), partiton 3 =
Windows system (NTFS). The Linux /root is on a separate physical drive. The
dual boot is via Grub installed on the /boot partiton, and this was all
working fine.

But I just upgraded the drives in the RAID pair, replacing them with 500GB
drives. I did this by replacing one of the 160s with a new 500 and letting
the RAID copy the drive, splitting the drives out of the RAID array and
increasing the size of the last partition of the 500 (which I did under
Windows since its the Windows partiton) then replacing the last 160 with the
other 500 and having the RAID controller create a new array with the two
500s, copying the drive that I'd copied from the 160. This worked great for
Windows, and that now boots and sees a 500GB RAID drive with all the data
intact.

However, Linux has a problem and will not now boot all the way. It reports
that the RAID /dev/mapper volume failed - the partition is beyond the
boundaries of the disk. Running fdisk shows that it is seeing the larger
partiton, but still sees the size of the RAID /dev/mapper drive as 160GB.
Here is the fdisk output for one of the physical drives and for the RAID
mapper drive:

Disk /dev/sda: 500.1 GB, 500107862016 bytes
255 heads, 63 sectors/track, 60801 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

   Device Boot  Start End  Blocks   Id  System
/dev/sda1   1 625 5018624b  W95 FAT32
Partition 1 does not end on cylinder boundary.
/dev/sda2 626 637   96390   83  Linux
/dev/sda3   * 638   60802   4832645127  HPFS/NTFS


Disk /dev/mapper/isw_bcifcijdi_Raid-0: 163.9 GB, 163925983232 bytes
255 heads, 63 sectors/track, 19929 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

Device Boot  Start End  Blocks  
Id  System
/dev/mapper/isw_bcifcijdi_Raid-0p1   1 625 5018624   
b  W95 FAT32
Partition 1 does not end on cylinder boundary.
/dev/mapper/isw_bcifcijdi_Raid-0p2 626 637   96390  
83  Linux
/dev/mapper/isw_bcifcijdi_Raid-0p3   * 638   60802   483264512   
7  HPFS/NTFS


They differ only in the drive capacity and number of cylinders.

I started to try to run a Linux reinstall, but it reports that the partiion
table on the mapper drive is invalid, giving an option to re-initialize it
but saying that doing so will lose all the data on the drive.

So questions:

1. Where is the drive size information for the RAID mapper drive kept, and
is there some way to patch it?

2. Is there some way to re-initialize the RAID mapper drive without
destroying the data on the drive?

Thanks,
Ian
-- 
View this message in context: 
http://www.nabble.com/RAID-mapper-device-size-wrong-after-replacing-drives-tf4958354.html#a14200241
Sent from the linux-raid mailing list archive at Nabble.com.

-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: RAID mapper device size wrong after replacing drives

2007-12-06 Thread Neil Brown

I think you would have more luck posting this to
[EMAIL PROTECTED] - I think that is where support for device mapper
happens.

NeilBrown


On Thursday December 6, [EMAIL PROTECTED] wrote:
 
 Hi,
 
 I have a problem with my RAID array under Linux after upgrading to larger
 drives. I have a machine with Windows and Linux dual-boot which had a pair
 of 160GB drives in a RAID-1 mirror with 3 partitions: partiton 1 = Windows
 boot partition (FAT32), partiton 2 = Linux /boot (ext3), partiton 3 =
 Windows system (NTFS). The Linux /root is on a separate physical drive. The
 dual boot is via Grub installed on the /boot partiton, and this was all
 working fine.
 
 But I just upgraded the drives in the RAID pair, replacing them with 500GB
 drives. I did this by replacing one of the 160s with a new 500 and letting
 the RAID copy the drive, splitting the drives out of the RAID array and
 increasing the size of the last partition of the 500 (which I did under
 Windows since its the Windows partiton) then replacing the last 160 with the
 other 500 and having the RAID controller create a new array with the two
 500s, copying the drive that I'd copied from the 160. This worked great for
 Windows, and that now boots and sees a 500GB RAID drive with all the data
 intact.
 
 However, Linux has a problem and will not now boot all the way. It reports
 that the RAID /dev/mapper volume failed - the partition is beyond the
 boundaries of the disk. Running fdisk shows that it is seeing the larger
 partiton, but still sees the size of the RAID /dev/mapper drive as 160GB.
 Here is the fdisk output for one of the physical drives and for the RAID
 mapper drive:
 
 Disk /dev/sda: 500.1 GB, 500107862016 bytes
 255 heads, 63 sectors/track, 60801 cylinders
 Units = cylinders of 16065 * 512 = 8225280 bytes
 
Device Boot  Start End  Blocks   Id  System
 /dev/sda1   1 625 5018624b  W95 FAT32
 Partition 1 does not end on cylinder boundary.
 /dev/sda2 626 637   96390   83  Linux
 /dev/sda3   * 638   60802   4832645127  HPFS/NTFS
 
 
 Disk /dev/mapper/isw_bcifcijdi_Raid-0: 163.9 GB, 163925983232 bytes
 255 heads, 63 sectors/track, 19929 cylinders
 Units = cylinders of 16065 * 512 = 8225280 bytes
 
 Device Boot  Start End  Blocks  
 Id  System
 /dev/mapper/isw_bcifcijdi_Raid-0p1   1 625 5018624   
 b  W95 FAT32
 Partition 1 does not end on cylinder boundary.
 /dev/mapper/isw_bcifcijdi_Raid-0p2 626 637   96390  
 83  Linux
 /dev/mapper/isw_bcifcijdi_Raid-0p3   * 638   60802   483264512   
 7  HPFS/NTFS
 
 
 They differ only in the drive capacity and number of cylinders.
 
 I started to try to run a Linux reinstall, but it reports that the partiion
 table on the mapper drive is invalid, giving an option to re-initialize it
 but saying that doing so will lose all the data on the drive.
 
 So questions:
 
 1. Where is the drive size information for the RAID mapper drive kept, and
 is there some way to patch it?
 
 2. Is there some way to re-initialize the RAID mapper drive without
 destroying the data on the drive?
 
 Thanks,
 Ian
 -- 
 View this message in context: 
 http://www.nabble.com/RAID-mapper-device-size-wrong-after-replacing-drives-tf4958354.html#a14200241
 Sent from the linux-raid mailing list archive at Nabble.com.
 
 -
 To unsubscribe from this list: send the line unsubscribe linux-raid in
 the body of a message to [EMAIL PROTECTED]
 More majordomo info at  http://vger.kernel.org/majordomo-info.html
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html