On Thu, Nov 17, 2011 at 4:21 PM, Neil Bothwick <[email protected]> wrote:
> On Thu, 17 Nov 2011 16:13:09 -0800, Mark Knecht wrote:
>
>> > Do you have separately partitioned drives with those partitions
>> > arranged into single-partition arrays, or do you have one RAID device
>> > that is then partitioned?
>> >
>> > If the latter, you should certainly work with the md device.
>> >
>> > I prefer to avoid all this confusion by creating a large, single
>> > partition array that I use an an LVM physical volume.
>>
>> Separately partitioned drives arranged into RAID arrays.
>
> So you have three partitions arranged into a single RAID5 partition,
> say /dev/md1?
>
> In that case, the size of /dev/md1 should already be correct and you only
> need to resize the filesystem and you should ignore my witterings about
> fdisk that filed to take into account your use of RAID.
> resize2fs /dev/md1 should be all you need, you shouldn't even need to
> unmount the filesystem.
>
I have 3 partitions which were previously RAID-1. I've already failed
one drive so at this moment it's a 2-drive RAID-1. I'm attempting to
get those two remaining 2 partitions converted to RAID-5 the command
suggested on the RAID list for doing that isn't working for me.
Once the 250GB RAID-1 is converted to RAID-5 i have to add a new drive
back in to become a 3-drive RAID-5. The drive I add will be the drive
I just failed.
c2stable ~ # mdadm --grow /dev/md6 --level=5
mdadm: /dev/md6: could not set level to raid5
c2stable ~ #
c2stable ~ # mdadm -D /dev/md6
/dev/md6:
Version : 1.1
Creation Time : Thu Apr 15 10:45:35 2010
Raid Level : raid1
Array Size : 247416933 (235.96 GiB 253.35 GB)
Used Dev Size : 247416933 (235.96 GiB 253.35 GB)
Raid Devices : 2
Total Devices : 2
Persistence : Superblock is persistent
Update Time : Thu Nov 17 13:27:20 2011
State : clean
Active Devices : 2
Working Devices : 2
Failed Devices : 0
Spare Devices : 0
Name : c2stable:6 (local to host c2stable)
UUID : 249c7331:a8203540:c8f3b020:fb30a66b
Events : 1039
Number Major Minor RaidDevice State
0 8 6 0 active sync /dev/sda6
1 8 22 1 active sync /dev/sdb6
c2stable ~ #