.. I have the same puzzlement? 0bo0 wrote: > I created a btrfs RAID-10 array across 4-drives, > > mkfs.btrfs -L TEST -m raid10 -d raid10 /dev/sda /dev/sdb /dev/sdc /dev/sdd > btrfs-show > Label: TEST uuid: 2ac85206-2d88-47d7-a1e7-a93d80b199f8 > Total devices 4 FS bytes used 28.00KB > devid 1 size 931.51GB used 2.03GB path /dev/sda > devid 2 size 931.51GB used 2.01GB path /dev/sdb > devid 4 size 931.51GB used 2.01GB path /dev/sdd > devid 3 size 931.51GB used 2.01GB path /dev/sdc > > @ mount, > > mount /dev/sda /mnt > df -H | grep /dev/sda > /dev/sda 4.1T 29k 4.1T 1% /mnt > > for RAID-10 across 4-drives, shouldn't the reported/available size be > 1/2x4TB ~ 2TB? > > e.g., using mdadm to build a RAID-10 array across the same drives, > > mdadm -v --create /dev/md0 --level=raid10 --raid-devices=4 /dev/sd[abcd]1 > pvcreate /dev/md0 > pvs > PV VG Fmt Attr PSize PFree > /dev/md0 lvm2 -- 1.82T 1.82T > > is the difference in available array space real, an artifact, or a > misunderstanding on my part? > > thanks. > -- > To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in > the body of a message to majord...@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html >
-- To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html