Hello,
we are running Lustre 2.12.5 on ZFS 0.8.3, where the MDT is backed by a ZFS
pool consisting of 8 mirror vdevs. Recently we had two drives fail in the same
mirror vdev, from which we were able to mostly recover (details on the
zfs-discuss list [1]). We have brought the Lustre filesystem back up and
mounted it on a single client read-only. Directly after mount, checking the
filesystem space with "df" shows the correct/expected size of the fs. But
running it again shows the total/used/available space inflated by a factor of
32 (almost exactly):
# mount /lustreTEST; sleep 1; df /lustreTEST
Filesystem 1K-blocks Used
Available Use% Mounted on
10.128.104.102@o2ib:10.128.104.101@o2ib:/lustrefs 1131965562880 527126408192
604839130112 47% /lustreTEST
# df /lustreTEST
Filesystem 1K-blocks Used
Available Use% Mounted on
10.128.104.102@o2ib:10.128.104.101@o2ib:/lustrefs 36222898339840 16868045127680
19354852425728 47% /lustreTEST
For comparison, "lfs df" shows the expected size (1.1PB):
# lfs df
UUID 1K-blocks Used Available Use% Mounted on
lustrefs-MDT0000_UUID 2236427008 1183398912 1053026048 53%
/lustreTEST[MDT:0]
lustrefs-OST0000_UUID 94330528768 42854200320 51476326400 46%
/lustreTEST[OST:0]
lustrefs-OST0001_UUID 94329859072 46121248768 48208608256 49%
/lustreTEST[OST:1]
lustrefs-OST0002_UUID 94330471424 44617906176 49712563200 48%
/lustreTEST[OST:2]
lustrefs-OST0003_UUID 94330831872 42602903552 51727926272 46%
/lustreTEST[OST:3]
lustrefs-OST0004_UUID 94330233856 42467627008 51862604800 46%
/lustreTEST[OST:4]
lustrefs-OST0005_UUID 94330768384 47377309696 46953456640 51%
/lustreTEST[OST:5]
lustrefs-OST0006_UUID 94330462208 42833333248 51497126912 46%
/lustreTEST[OST:6]
lustrefs-OST0007_UUID 94330291200 42838326272 51491962880 46%
/lustreTEST[OST:7]
lustrefs-OST0008_UUID 94330214400 45438008320 48892204032 49%
/lustreTEST[OST:8]
lustrefs-OST0009_UUID 94330670080 44649268224 49681399808 48%
/lustreTEST[OST:9]
lustrefs-OST000a_UUID 94330542080 43913293824 50417246208 47%
/lustreTEST[OST:10]
lustrefs-OST000b_UUID 94330704896 41412985856 52917716992 44%
/lustreTEST[OST:11]
filesystem_summary: 1131965578240 527126411264 604839142400 47% /lustreTEST
We are wondering why the df output is so inflated, and how worried we should be
about it. For now we've refrained from putting the fs back into production.
Anyway, we are running LFSCK now. Is there a way to estimate how long it will
take? I can see progress in the "checked" and "current_position" counters
reported by "lctl get_param osd-zfs.lustrefs-MDT*.oi_scrub /
mdd.lustrefs-MDT*.lfsck_layout / mdd.lustrefs-MDT*.lfsck_namespace" but how
could we estimate what the 100% values of those counters are? Is it related to
how full the MDT is? We seem to have 264M inodes in use.
Many thanks,
Frank
[1]
https://zfsonlinux.topicbox.com/groups/zfs-discuss/T28e59deb8ff2c26d/two-failed-drives-in-mirror-vdev
_______________________________________________
lustre-discuss mailing list
[email protected]
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org