You had it here:
[root@server ~]# mmlsrecoverygroup BB1RGL -L declustered recovery group arrays vdisks pdisks format version ----------------- ----------- ------ ------ -------------- BB1RGL 3 18 119 4.2.0.1 declustered needs replace scrub background activity array service vdisks pdisks spares threshold free space duration task progress priority ----------- ------- ------ ------ ------ --------- ---------- -------- ------------------------- LOG no 1 3 0,0 1 558 GiB 14 days scrub 51% low DA1 no 11 58 2,31 2 12 GiB 14 days scrub 78% low DA2 no 6 58 2,31 2 4096 MiB 14 days scrub 10% low 12 GiB in DA1, and 4096 MiB i DA2, but effectively you'll get less when you add a raidCode to the vdisk. Best way to use it id to just don't specify a size to the vdisk, and max possible size will be used. -jf søn. 9. jul. 2017 kl. 14.26 skrev atmane khiredine <[email protected]>: > thank you very much for replying. I can not find the free space > > Here is the output of mmlsrecoverygroup > > [root@server1 ~]#mmlsrecoverygroup > > declustered > arrays with > recovery group vdisks vdisks servers > ------------------ ----------- ------ ------- > BB1RGL 3 18 server1,server2 > BB1RGR 3 18 server2,server1 > -------------------------------------------------------------- > [root@server ~]# mmlsrecoverygroup BB1RGL -L > > declustered > recovery group arrays vdisks pdisks format version > ----------------- ----------- ------ ------ -------------- > BB1RGL 3 18 119 4.2.0.1 > > declustered needs replace > scrub background activity > array service vdisks pdisks spares threshold free space > duration task progress priority > ----------- ------- ------ ------ ------ --------- ---------- > -------- ------------------------- > LOG no 1 3 0,0 1 558 GiB 14 > days scrub 51% low > DA1 no 11 58 2,31 2 12 GiB 14 > days scrub 78% low > DA2 no 6 58 2,31 2 4096 MiB 14 > days scrub 10% low > > declustered > checksum > vdisk RAID code array vdisk size block > size granularity state remarks > ------------------ ------------------ ----------- ---------- > ---------- ----------- ----- ------- > gss0_logtip 3WayReplication LOG 128 MiB 1 > MiB 512 ok logTip > gss0_loghome 4WayReplication DA1 40 GiB 1 > MiB 512 ok log > BB1RGL_GPFS4_META1 4WayReplication DA1 451 GiB 1 > MiB 32 KiB ok > BB1RGL_GPFS4_DATA1 8+2p DA1 5133 GiB 1 > MiB 32 KiB ok > BB1RGL_GPFS1_META1 4WayReplication DA1 451 GiB 1 > MiB 32 KiB ok > BB1RGL_GPFS1_DATA1 8+2p DA1 12 TiB 1 > MiB 32 KiB ok > BB1RGL_GPFS3_META1 4WayReplication DA1 451 GiB 1 > MiB 32 KiB ok > BB1RGL_GPFS3_DATA1 8+2p DA1 12 TiB 1 > MiB 32 KiB ok > BB1RGL_GPFS2_META1 4WayReplication DA1 451 GiB 1 > MiB 32 KiB ok > BB1RGL_GPFS2_DATA1 8+2p DA1 13 TiB 2 > MiB 32 KiB ok > BB1RGL_GPFS2_META2 4WayReplication DA2 451 GiB 1 > MiB 32 KiB ok > BB1RGL_GPFS2_DATA2 8+2p DA2 13 TiB 2 > MiB 32 KiB ok > BB1RGL_GPFS1_META2 4WayReplication DA2 451 GiB 1 > MiB 32 KiB ok > BB1RGL_GPFS1_DATA2 8+2p DA2 12 TiB 1 > MiB 32 KiB ok > BB1RGL_GPFS5_META1 4WayReplication DA1 750 GiB 1 > MiB 32 KiB ok > BB1RGL_GPFS5_DATA1 8+2p DA1 70 TiB 16 > MiB 32 KiB ok > BB1RGL_GPFS5_META2 4WayReplication DA2 750 GiB 1 > MiB 32 KiB ok > BB1RGL_GPFS5_DATA2 8+2p DA2 90 TiB 16 > MiB 32 KiB ok > > config data declustered array VCD spares actual rebuild > spare space remarks > ------------------ ------------------ ------------- > --------------------------------- ---------------- > rebuild space DA1 31 34 pdisk > rebuild space DA2 31 35 pdisk > > > config data max disk group fault tolerance actual disk group > fault tolerance remarks > ------------------ --------------------------------- > --------------------------------- ---------------- > rg descriptor 1 enclosure + 1 drawer 1 enclosure + 1 > drawer limiting fault tolerance > system index 2 enclosure 1 enclosure + 1 > drawer limited by rg descriptor > > vdisk max disk group fault tolerance actual disk group > fault tolerance remarks > ------------------ --------------------------------- > --------------------------------- ---------------- > gss0_logtip 2 enclosure 1 enclosure + 1 > drawer limited by rg descriptor > gss0_loghome 1 enclosure + 1 drawer 1 enclosure + 1 > drawer > BB1RGL_GPFS4_META1 1 enclosure + 1 drawer 1 enclosure + 1 > drawer > BB1RGL_GPFS4_DATA1 2 drawer 2 drawer > BB1RGL_GPFS1_META1 1 enclosure + 1 drawer 1 enclosure + 1 > drawer > BB1RGL_GPFS1_DATA1 2 drawer 2 drawer > BB1RGL_GPFS3_META1 1 enclosure + 1 drawer 1 enclosure + 1 > drawer > BB1RGL_GPFS3_DATA1 2 drawer 2 drawer > BB1RGL_GPFS2_META1 1 enclosure + 1 drawer 1 enclosure + 1 > drawer > BB1RGL_GPFS2_DATA1 2 drawer 2 drawer > BB1RGL_GPFS2_META2 3 enclosure 1 enclosure + 1 > drawer limited by rg descriptor > BB1RGL_GPFS2_DATA2 2 drawer 2 drawer > BB1RGL_GPFS1_META2 3 enclosure 1 enclosure + 1 > drawer limited by rg descriptor > BB1RGL_GPFS1_DATA2 2 drawer 2 drawer > BB1RGL_GPFS5_META1 1 enclosure + 1 drawer 1 enclosure + 1 > drawer > BB1RGL_GPFS5_DATA1 2 drawer 2 drawer > BB1RGL_GPFS5_META2 3 enclosure 1 enclosure + 1 > drawer limited by rg descriptor > BB1RGL_GPFS5_DATA2 2 drawer 2 drawer > > active recovery group server servers > ----------------------------------------------- ------- > server1 server1,server2 > > > Atmane Khiredine > HPC System Administrator | Office National de la Météorologie > Tél : +213 21 50 73 93 # 303 | Fax : +213 21 50 79 40 | E-mail : > [email protected] > ________________________________ > De : Laurence Horrocks-Barlow [[email protected]] > Envoyé : dimanche 9 juillet 2017 09:58 > À : gpfsug main discussion list; atmane khiredine; > [email protected] > Objet : Re: [gpfsug-discuss] get free space in GSS > > You can check the recovery groups to see if there is any remaining space. > > I don't have access to my test system to confirm the syntax however if > memory serves. > > Run mmlsrecoverygroup to get a list of all the recovery groups then: > > mmlsrecoverygroup <YOURRECOVERYGROUP> -L > > This will list all your declustered arrays and their free space. > > Their might be another method, however this way has always worked well for > me. > > -- Lauz > > > > On 9 July 2017 09:00:07 BST, Atmane <[email protected]> wrote: > > Dear all, > > My name is Khiredine Atmane and I am a HPC system administrator at the > National Office of Meteorology Algeria . We have a GSS24 running > gss2.5.10.3<http://2.5.10.3>-3b and gpfs-4.2.0.3<http://4.2.0.3>. > > GSS configuration: 4 enclosures, 6 SSDs, 1 empty slots, 239 disks total, 0 > NVRAM partitions > > disks = 3Tb > SSD = 200 Gb > df -h > Filesystem Size Used Avail Use% Mounted on > > /dev/gpfs1 49T 18T 31T 38% /gpfs1 > /dev/gpfs2 53T 13T 40T 25% /gpfs2 > /dev/gpfs3 25T 4.9T 20T 21% /gpfs3 > /dev/gpfs4 11T 133M 11T 1% /gpfs4 > /dev/gpfs5 323T 34T 290T 11% /gpfs5 > > Total Is 461 To > > I think we have more space > Could anyone make recommendation to troubleshoot find how many free space > in GSS ? > How to find the available space ? > Thank you! > > Atmane > > > > -- > Sent from my Android device with K-9 Mail. Please excuse my brevity. > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at spectrumscale.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss >
_______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss
