Michael Stumpf wrote:

I get this from the latest stable resize2fs:

[EMAIL PROTECTED] parted]# resize2fs /dev/my_vol_grp/my_log_vol
resize2fs 1.38 (30-Jun-2005)
Resizing the filesystem on /dev/my_vol_grp/my_log_vol to 488390656 (4k) blocks.
Killed

Parted (again, latest stable) tells me the following:
Using /dev/mapper/my_vol_grp-my_log_vol
(parted) resize 1 0 100% No Implementation: This ext2 file system has a rather strange layout! Parted can't resize this (yet).
(parted)
Similar results from ext2resize/ext2online. This is an ordinary ext3 fs, living inside an LVM2 that has already been increased to accomodate (used all free extents).. I've resized it down and up before, though it is possible I am resizing it larger than it has been before (1.8TB). Not sure what's up. Any advice welcome; my research in this has me getting a bit nervous about lvm2 bugs causing loss of data. While I want a single resilient (via raid 5) volume, I may be willing to ditch a whole layer of software (lvm2) to get some security.


Surprised noone has hit this before. It turns out that somehow my swap space disappeared in a system migration. This became more obvious as I explicitly tried to extend the fs to a lower limit (438390656), where resize2fs worked for a while, then informed me that it couldn't allocate some memory.

Add 512mb of swap and problem solved.. never used more than ~100mb of swap (256mb main memory).

Hope this helps someone.



-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to