.. I have the same puzzlement?

0bo0 wrote:
> I created a btrfs RAID-10 array across 4-drives,
>
>  mkfs.btrfs -L TEST -m raid10 -d raid10 /dev/sda /dev/sdb /dev/sdc /dev/sdd
>  btrfs-show
>       Label: TEST  uuid: 2ac85206-2d88-47d7-a1e7-a93d80b199f8
>               Total devices 4 FS bytes used 28.00KB
>               devid    1 size 931.51GB used 2.03GB path /dev/sda
>               devid    2 size 931.51GB used 2.01GB path /dev/sdb
>               devid    4 size 931.51GB used 2.01GB path /dev/sdd
>               devid    3 size 931.51GB used 2.01GB path /dev/sdc
>
> @ mount,
>
>  mount /dev/sda /mnt
>  df -H | grep /dev/sda
>       /dev/sda               4.1T    29k   4.1T   1% /mnt
>
> for RAID-10 across 4-drives, shouldn't the reported/available size be
> 1/2x4TB ~ 2TB?
>
> e.g., using mdadm to build a RAID-10 array across the same drives,
>
>  mdadm -v --create /dev/md0 --level=raid10 --raid-devices=4 /dev/sd[abcd]1
>  pvcreate     /dev/md0
> pvs
>   PV         VG   Fmt  Attr PSize   PFree
>   /dev/md0        lvm2 --   1.82T 1.82T
>
> is the difference in available array space real, an artifact, or a
> misunderstanding on my part?
>
> thanks.
> --
> To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
> the body of a message to majord...@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>   

--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to