Problem resolved by removing some large files before do gfs2_grow, as it 
requires some free space to process "grow".

refer to:
https://bugzilla.redhat.com/show_bug.cgi?id=490649

On 7/04/2011, at 1:38 PM, Marco Huang wrote:

> Hi,
> 
> We are running Cenos5.5 (gfs2-utils.x86_64 v0.1.62-20.el5) cluster. We want 
> to add another 5T disk space on the filesystem - expand from 19T to 25T, 
> however it doesn't grow over 20T. There was no error when I did test run. So 
> just wondering if there is a limitation of gfs2_grow which only can grow gfs2 
> filesystem upto 20T ? Is there anyone had same experience?
> 
> Test run
> # gfs2_grow -T /mnt/fsbackup
> (Test mode--File system will not be changed)
> FS: Mount Point: /mnt/fsbackup
> FS: Device:      /dev/mapper/fsbackup-fsbackup01
> FS: Size:        4882811901 (0x12309cbfd)
> FS: RG size:     524244 (0x7ffd4)
> DEV: Size:       6103514112 (0x16bcc3c00)
> The file system grew by 4768368MB.
> gfs2_grow complete.
> 
> Actual run
> # gfs2_grow /mnt/fsbackup
> FS: Mount Point: /mnt/fsbackup
> FS: Device:      /dev/mapper/fsbackup-fsbackup01
> FS: Size:        4882811901 (0x12309cbfd)
> FS: RG size:     524244 (0x7ffd4)
> DEV: Size:       6103514112 (0x16bcc3c00)
> The file system grew by 4768368MB.
> Error writing new rindex entries;aborted.
> gfs2_grow complete.
> 
> 
> Before gfs2_grow
> # df -h /mnt/fsbackup/
> Filesystem            Size  Used Avail Use% Mounted on
> /dev/mapper/fsbackup-fsbackup01
>                        19T   19T  651M 100% /mnt/fsbackup
> 
> # gfs2_tool df
> /mnt/fsbackup:
>   SB lock proto = "lock_dlm"
>   SB lock table = "FSC:fsbackup01"
>   SB ondisk format = 1801
>   SB multihost format = 1900
>   Block size = 4096
>   Journals = 8
>   Resource Groups = 10112
>   Mounted lock proto = "lock_dlm"
>   Mounted lock table = "FSC:fsbackup01"
>   Mounted host data = "jid=0:id=65539:first=1"
>   Journal number = 0
>   Lock module flags = 0
>   Local flocks = FALSE
>   Local caching = FALSE
> 
>   Type           Total Blocks   Used Blocks    Free Blocks    use%           
>   ------------------------------------------------------------------------
>   data           5300270360     4882310027     417960333      92%
>   inodes         447901122      29940789       417960333      7%
> 
> After gfs2_grow
> #  df -h /mnt/fsbackup/
> Filesystem            Size  Used Avail Use% Mounted on
> /dev/mapper/fsbackup-fsbackup01
>                        20T   19T  1.6T  93% /mnt/fsbackup
> 
> 
> # gfs2_tool df
> /mnt/fsbackup:
>   SB lock proto = "lock_dlm"
>   SB lock table = "FSC:fsbackup01"
>   SB ondisk format = 1801
>   SB multihost format = 1900
>   Block size = 4096
>   Journals = 8
>   Resource Groups = 9314
>   Mounted lock proto = "lock_dlm"
>   Mounted lock table = "FSC:fsbackup01"
>   Mounted host data = "jid=0:id=65539:first=1"
>   Journal number = 0
>   Lock module flags = 0
>   Local flocks = FALSE
>   Local caching = FALSE
> 
>   Type           Total Blocks   Used Blocks    Free Blocks    use%           
>   ------------------------------------------------------------------------
>   data           4882476584     4882310009     166575         100%
>   inodes         30107364       29940789       166575         99%
> 
> 
> 
> cheers
> --
> Marco
> 
> --
> Linux-cluster mailing list
> [email protected]
> https://www.redhat.com/mailman/listinfo/linux-cluster

--
Linux-cluster mailing list
[email protected]
https://www.redhat.com/mailman/listinfo/linux-cluster

Reply via email to