On 02/07/2014 05:40 AM, Roman Mamedov wrote: > On Thu, 06 Feb 2014 20:54:19 +0100 > Goffredo Baroncelli <kreij...@libero.it> wrote: > [...]
As Roman pointed out, df show the "raw" space available. However when a RAID level is used, the space available to the user is less. This patch try to address this estimation correcting the value on the basis of the RAID level. This is my third revision of this patch. In this last issue, I addressed the bugs related to an uncorrected evaluation of the free space in case of RAID1 [1] and DUP. I have to point out that the free space estimation is quite approximative, because it assumes: a) all the new files are allocated in data "chunk" b) the free space will not consumed by metadata c) the already allocated chunk are not evaluated for the free space estimation Both these assumptions are unrelated to my patch. I performed some tests with a filesystem composed by 7 51GB disks. Here my "df" results: Profile: single Filesystem Size Used Avail Use% Mounted on /dev/vdb 351G 512K 348G 1% /mnt/btrfs1 Profile: raid1 Filesystem Size Used Avail Use% Mounted on /dev/vdb 351G 1.3M 175G 1% /mnt/btrfs1 Profile: raid10 Filesystem Size Used Avail Use% Mounted on /dev/vdb 351G 2.3M 177G 1% /mnt/btrfs1 Profile: raid5 Filesystem Size Used Avail Use% Mounted on /dev/vdb 351G 2.0M 298G 1% /mnt/btrfs1 Profile: raid6 Filesystem Size Used Avail Use% Mounted on /dev/vdb 351G 1.8M 248G 1% /mnt/btrfs1 Profile: DUP (only one 50GB disk was used) Filesystem Size Used Avail Use% Mounted on /dev/vdc 51G 576K 26G 1% /mnt/btrfs1 Below my patch. BR G.Baroncelli [1] the bug is before my patch; try to see what happens when you create a RAID1 filesystem with three disks. Changes history: V1 First issue V2 Correct a (old) bug when in RAID10 the disks aren't a multiple of 4 V3 Correct the free space estimation in RAID1 (when the number of disks are odd) and DUP diff --git a/fs/btrfs/super.c b/fs/btrfs/super.c index d71a11d..4064a5f 100644 --- a/fs/btrfs/super.c +++ b/fs/btrfs/super.c @@ -1481,10 +1481,16 @@ static int btrfs_calc_avail_data_space(struct btrfs_root *root, u64 *free_bytes) num_stripes = nr_devices; } else if (type & BTRFS_BLOCK_GROUP_RAID1) { min_stripes = 2; - num_stripes = 2; + num_stripes = nr_devices; } else if (type & BTRFS_BLOCK_GROUP_RAID10) { min_stripes = 4; - num_stripes = 4; + num_stripes = nr_devices; + } else if (type & BTRFS_BLOCK_GROUP_RAID5) { + min_stripes = 3; + num_stripes = nr_devices; + } else if (type & BTRFS_BLOCK_GROUP_RAID6) { + min_stripes = 4; + num_stripes = nr_devices; } if (type & BTRFS_BLOCK_GROUP_DUP) @@ -1560,9 +1566,44 @@ static int btrfs_calc_avail_data_space(struct btrfs_root *root, u64 *free_bytes) if (devices_info[i].max_avail >= min_stripe_size) { int j; - u64 alloc_size; + u64 alloc_size, delta; + int k, div; + + /* + * Depending by the RAID profile, we use some + * disk space as redundancy: + * RAID1, RAID10, DUP -> half of space used as redundancy + * RAID5 -> 1 stripe used as redundancy + * RAID6 -> 2 stripes used as redundancy + * RAID0,LINEAR -> no redundancy + */ + if (type & BTRFS_BLOCK_GROUP_RAID1) { + k = num_stripes; + div = 2; + } else if (type & BTRFS_BLOCK_GROUP_DUP) { + k = num_stripes; + div = 2; + } else if (type & BTRFS_BLOCK_GROUP_RAID10) { + k = num_stripes; + div = 2; + } else if (type & BTRFS_BLOCK_GROUP_RAID5) { + k = num_stripes-1; + div = 1; + } else if (type & BTRFS_BLOCK_GROUP_RAID6) { + k = num_stripes-2; + div = 1; + } else { /* RAID0/LINEAR */ + k = num_stripes; + div = 1; + } + + delta = devices_info[i].max_avail*k; + if (div==2) + delta >>= 1; + else if (div>2) + do_div(delta, div); + avail_space += delta; - avail_space += devices_info[i].max_avail * num_stripes; alloc_size = devices_info[i].max_avail; for (j = i + 1 - num_stripes; j <= i; j++) devices_info[j].max_avail -= alloc_size; -- gpg @keyserver.linux.it: Goffredo Baroncelli (kreijackATinwind.it> Key fingerprint BBF5 1610 0B64 DAC6 5F7D 17B2 0EDA 9B37 8B82 E0B5 -- To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html