I could be wrong, but I don't think that the original poster was asking
why the SIZE field of zpool list was wrong, but rather why the AVAIL
space in zfs list was lower than he expected.
I would find it easier to answer the question if I knew his drive count
and drive size.
Chris
On 08/24/2015 02:12 PM, Alexander I Kulyavtsev wrote:
Same question here.
6TB/65TB is 11% . In our case about the same fraction was "missing."
My speculation was, It may happen if at some point between zpool and linux the
value reported in TB is interpreted as in TiB, and then converted to TB. Or
unneeded conversion MB to MiB done twice, etc.
Here is my numbers:
We have 12* 4TB drives per pool, it is 48 TB (decimal).
zpool created as raidz2 10+2.
zpool reports 43.5T.
Pool size shall be 48T=4T*12, or 40T=4T*10 (depending what zpool shows, before
raiding or after raiding).
From the Oracle ZFS documentation, "zpool list" returns the total space without
overheads, thus 48 TB shall be reported by zpool instead of 43.5TB.
In my case, it looked like conversion error/interpretation issue between TB and
TiB:
48*1000*1000*1000*1000/1024/1024/1024/1024 = 43.65574568510055541992
At disk level:
~/sas2ircu 0 display
Device is a Hard disk
Enclosure # : 2
Slot # : 12
SAS Address : 5003048-0-015a-a918
State : Ready (RDY)
Size (in MB)/(in sectors) : 3815447/7814037167
Manufacturer : ATA
Model Number : HGST HUS724040AL
Firmware Revision : AA70
Serial No : PN2334PBJPW14T
GUID : 5000cca23de6204b
Protocol : SATA
Drive Type : SATA_HDD
One disk size is about 4 TB (decimal):
3815447*1024*1024 = 4000786153472
7814037167*512 = 4000787029504
vdev presents whole disk to zpool. There is some overhead, some space left on
sdq9 .
[root@lfs1 scripts]# head -4 /etc/zfs/vdev_id.conf
alias s0 /dev/disk/by-path/pci-0000:03:00.0-sas-0x50030480015aa90c-lun-0
alias s1 /dev/disk/by-path/pci-0000:03:00.0-sas-0x50030480015aa90d-lun-0
alias s2 /dev/disk/by-path/pci-0000:03:00.0-sas-0x50030480015aa90e-lun-0
alias s3 /dev/disk/by-path/pci-0000:03:00.0-sas-0x50030480015aa90f-lun-0
...
alias s12 /dev/disk/by-path/pci-0000:03:00.0-sas-0x50030480015aa918-lun-0
...
[root@lfs1 scripts]# ls -l /dev/disk/by-path/
...
lrwxrwxrwx 1 root root 9 Jul 23 16:27
pci-0000:03:00.0-sas-0x50030480015aa918-lun-0 -> ../../sdq
lrwxrwxrwx 1 root root 10 Jul 23 16:27
pci-0000:03:00.0-sas-0x50030480015aa918-lun-0-part1 -> ../../sdq1
lrwxrwxrwx 1 root root 10 Jul 23 16:27
pci-0000:03:00.0-sas-0x50030480015aa918-lun-0-part9 -> ../../sdq9
Pool report:
[root@lfs1 scripts]# zpool list
NAME SIZE ALLOC FREE EXPANDSZ FRAG CAP DEDUP HEALTH ALTROOT
zpla-0000 43.5T 10.9T 32.6T - 16% 24% 1.00x ONLINE -
zpla-0001 43.5T 11.0T 32.5T - 17% 25% 1.00x ONLINE -
zpla-0002 43.5T 10.8T 32.7T - 17% 24% 1.00x ONLINE -
[root@lfs1 scripts]#
[root@lfs1 ~]# zpool list -v zpla-0001
NAME SIZE ALLOC FREE EXPANDSZ FRAG CAP DEDUP HEALTH ALTROOT
zpla-0001 43.5T 11.0T 32.5T - 17% 25% 1.00x ONLINE -
raidz2 43.5T 11.0T 32.5T - 17% 25%
s12 - - - - - -
s13 - - - - - -
s14 - - - - - -
s15 - - - - - -
s16 - - - - - -
s17 - - - - - -
s18 - - - - - -
s19 - - - - - -
s20 - - - - - -
s21 - - - - - -
s22 - - - - - -
s23 - - - - - -
[root@lfs1 ~]#
[root@lfs1 ~]# zpool get all zpla-0001
NAME PROPERTY VALUE SOURCE
zpla-0001 size 43.5T -
zpla-0001 capacity 25% -
zpla-0001 altroot - default
zpla-0001 health ONLINE -
zpla-0001 guid 5472902975201420000 default
zpla-0001 version - default
zpla-0001 bootfs - default
zpla-0001 delegation on default
zpla-0001 autoreplace off default
zpla-0001 cachefile - default
zpla-0001 failmode wait default
zpla-0001 listsnapshots off default
zpla-0001 autoexpand off default
zpla-0001 dedupditto 0 default
zpla-0001 dedupratio 1.00x -
zpla-0001 free 32.5T -
zpla-0001 allocated 11.0T -
zpla-0001 readonly off -
zpla-0001 ashift 12 local
zpla-0001 comment - default
zpla-0001 expandsize - -
zpla-0001 freeing 0 default
zpla-0001 fragmentation 17% -
zpla-0001 leaked 0 default
zpla-0001 feature@async_destroy enabled local
zpla-0001 feature@empty_bpobj active local
zpla-0001 feature@lz4_compress active local
zpla-0001 feature@spacemap_histogram active local
zpla-0001 feature@enabled_txg active local
zpla-0001 feature@hole_birth active local
zpla-0001 feature@extensible_dataset enabled local
zpla-0001 feature@embedded_data active local
zpla-0001 feature@bookmarks enabled local
Alex.
On Aug 19, 2015, at 8:18 AM, Götz Waschk <goetz.was...@gmail.com> wrote:
Dear Lustre experts,
I have configured two different Lustre instances, both using Lustre
2.5.3, one with ldiskfs on RAID-6 hardware RAID and one using ZFS and
RAID-Z2, using the same type of hardware. I was wondering, why I 24 TB
less space available, when I should have the same amount of parity
used:
# lfs df
UUID 1K-blocks Used Available Use% Mounted on
fs19-MDT0000_UUID 50322916 472696 46494784 1%
/testlustre/fs19[MDT:0]
fs19-OST0000_UUID 51923288320 12672 51923273600 0%
/testlustre/fs19[OST:0]
fs19-OST0001_UUID 51923288320 12672 51923273600 0%
/testlustre/fs19[OST:1]
fs19-OST0002_UUID 51923288320 12672 51923273600 0%
/testlustre/fs19[OST:2]
fs19-OST0003_UUID 51923288320 12672 51923273600 0%
/testlustre/fs19[OST:3]
filesystem summary: 207693153280 50688 207693094400 0% /testlustre/fs19
UUID 1K-blocks Used Available Use% Mounted on
fs18-MDT0000_UUID 47177700 482152 43550028 1%
/lustre/fs18[MDT:0]
fs18-OST0000_UUID 58387106064 6014088200 49452733560 11%
/lustre/fs18[OST:0]
fs18-OST0001_UUID 58387106064 5919753028 49547068928 11%
/lustre/fs18[OST:1]
fs18-OST0002_UUID 58387106064 5944542316 49522279640 11%
/lustre/fs18[OST:2]
fs18-OST0003_UUID 58387106064 5906712004 49560109952 11%
/lustre/fs18[OST:3]
filesystem summary: 233548424256 23785095548 198082192080 11% /lustre/fs18
fs18 is using ldiskfs, while fs19 is ZFS:
# zpool list
NAME SIZE ALLOC FREE CAP DEDUP HEALTH ALTROOT
lustre-ost1 65T 18,1M 65,0T 0% 1.00x ONLINE -
# zfs list
NAME USED AVAIL REFER MOUNTPOINT
lustre-ost1 13,6M 48,7T 311K /lustre-ost1
lustre-ost1/ost1 12,4M 48,7T 12,4M /lustre-ost1/ost1
Any idea on why my 6TB per OST went?
Regards, Götz Waschk
_______________________________________________
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
_______________________________________________
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
.
_______________________________________________
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org