Addendum to last message: We haven't upgraded recently as far as I know (I just inherited this a couple of months ago.) but am planning an outage soon to upgrade from 4.2.0-4 to 4.2.3-5.
My growing collection of output files generally contain something like This inode list was generated in the Parallel Inode Traverse on Thu Nov 2 08:34:22 2017 INODE_NUMBER DUMMY_INFO SNAPSHOT_ID ISGLOBAL_SNAPSHOT INDEPENDENT_FSETID MEMO(INODE_FLAGS FILE_TYPE [ERROR]) 53506 0:0 0 1 0 illreplicated REGULAR_FILE RESERVED Error: 28 No space left on device With that inode varying slightly. jbh On Thu, Nov 2, 2017 at 8:55 AM, Scott Fadden <[email protected]> wrote: > Sorry just reread as I hit send and saw this was mmrestripe, in my case it > was mmdeledisk. > > Did you try running the command on just one pool. Or using -B instead? > > What is the file it is complaining about in > "/var/mmfs/tmp/gsfs0.pit.interestingInodes.12888779711" > ? > > Looks like it could be related to the maxfeaturelevel of the cluster. Have > you recently upgraded? Is everything up to the same level? > > Scott Fadden > Spectrum Scale - Technical Marketing > Phone: (503) 880-5833 > [email protected] > http://www.ibm.com/systems/storage/spectrum/scale > > > > ----- Original message ----- > From: Scott Fadden/Portland/IBM > To: [email protected] > Cc: [email protected] > Subject: Re: [gpfsug-discuss] mmrestripefs "No space left on device" > Date: Thu, Nov 2, 2017 8:44 AM > > I opened a defect on this the other day, in my case it was an incorrect > error message. What it meant to say was,"The pool is not empty." Are you > trying to remove the last disk in a pool? If so did you empty the pool with > a MIGRATE policy first? > > > Scott Fadden > Spectrum Scale - Technical Marketing > Phone: (503) 880-5833 > [email protected] > http://www.ibm.com/systems/storage/spectrum/scale > > > > ----- Original message ----- > From: John Hanks <[email protected]> > Sent by: [email protected] > To: gpfsug main discussion list <[email protected]> > Cc: > Subject: Re: [gpfsug-discuss] mmrestripefs "No space left on device" > Date: Thu, Nov 2, 2017 8:34 AM > > We have no snapshots ( they were the first to go when we initially hit the > full metadata NSDs). > > I've increased quotas so that no filesets have hit a space quota. > > Verified that there are no inode quotas anywhere. > > mmdf shows the least amount of free space on any nsd to be 9% free. > > Still getting this error: > > [root@scg-gs0 ~]# mmrestripefs gsfs0 -r -N scg-gs0,scg-gs1,scg-gs2,scg-gs3 > Scanning file system metadata, phase 1 ... > Scan completed successfully. > Scanning file system metadata, phase 2 ... > Scanning file system metadata for sas0 storage pool > Scanning file system metadata for sata0 storage pool > Scan completed successfully. > Scanning file system metadata, phase 3 ... > Scan completed successfully. > Scanning file system metadata, phase 4 ... > Scan completed successfully. > Scanning user file metadata ... > Error processing user file metadata. > No space left on device > Check file '/var/mmfs/tmp/gsfs0.pit.interestingInodes.12888779711' on > scg-gs0 for inodes with broken disk addresses or failures. > mmrestripefs: Command failed. Examine previous error messages to determine > cause. > > I should note too that this fails almost immediately, far to quickly to > fill up any location it could be trying to write to. > > jbh > > On Thu, Nov 2, 2017 at 7:57 AM, David Johnson <[email protected]> > wrote: > > One thing that may be relevant is if you have snapshots, depending on your > release level, > inodes in the snapshot may considered immutable, and will not be > migrated. Once the snapshots > have been deleted, the inodes are freed up and you won’t see the (somewhat > misleading) message > about no space. > > — ddj > Dave Johnson > Brown University > > > On Nov 2, 2017, at 10:43 AM, John Hanks <[email protected]> wrote: > Thanks all for the suggestions. > > Having our metadata NSDs fill up was what prompted this exercise, but > space was previously feed up on those by switching them from metadata+data > to metadataOnly and using a policy to migrate files out of that pool. So > these now have about 30% free space (more if you include fragmented space). > The restripe attempt is just to make a final move of any remaining data off > those devices. All the NSDs now have free space on them. > > df -i shows inode usage at about 84%, so plenty of free inodes for the > filesystem as a whole. > > We did have old .quota files laying around but removing them didn't have > any impact. > > mmlsfileset fs -L -i is taking a while to complete, I'll let it simmer > while getting to work. > > mmrepquota does show about a half-dozen filesets that have hit their quota > for space (we don't set quotas on inodes). Once I'm settled in this morning > I'll try giving them a little extra space and see what happens. > > jbh > > > On Thu, Nov 2, 2017 at 4:19 AM, Oesterlin, Robert < > [email protected]> wrote: > > One thing that I’ve run into before is that on older file systems you had > the “*.quota” files in the file system root. If you upgraded the file > system to a newer version (so these files aren’t used) - There was a bug at > one time where these didn’t get properly migrated during a restripe. > Solution was to just remove them > > > > > > Bob Oesterlin > > Sr Principal Storage Engineer, Nuance > > > > *From: *<[email protected]> on behalf of John > Hanks <[email protected]> > *Reply-To: *gpfsug main discussion list <[email protected]> > *Date: *Wednesday, November 1, 2017 at 5:55 PM > *To: *gpfsug <[email protected]> > *Subject: *[EXTERNAL] [gpfsug-discuss] mmrestripefs "No space left on > device" > > > > Hi all, > > > > I'm trying to do a restripe after setting some nsds to metadataOnly and I > keep running into this error: > > > > Scanning user file metadata ... > > 0.01 % complete on Wed Nov 1 15:36:01 2017 ( 40960 inodes with > total 531689 MB data processed) > > Error processing user file metadata. > > Check file '/var/mmfs/tmp/gsfs0.pit.interestingInodes.12888779708' on > scg-gs0 for inodes with broken disk addresses or failures. > > mmrestripefs: Command failed. Examine previous error messages to determine > cause. > > > > The file it points to says: > > > > This inode list was generated in the Parallel Inode Traverse on Wed Nov 1 > 15:36:06 2017 > > INODE_NUMBER DUMMY_INFO SNAPSHOT_ID ISGLOBAL_SNAPSHOT INDEPENDENT_FSETID > MEMO(INODE_FLAGS FILE_TYPE [ERROR]) > > 53504 0:0 0 1 0 > illreplicated REGULAR_FILE RESERVED Error: 28 No space left on device > > > > > > /var on the node I am running this on has > 128 GB free, all the NSDs have > plenty of free space, the filesystem being restriped has plenty of free > space and if I watch the node while running this no filesystem on it even > starts to get full. Could someone tell me where mmrestripefs is attempting > to write and/or how to point it at a different location? > > > > Thanks, > > > > jbh > > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at spectrumscale.org > <https://urldefense.proofpoint.com/v2/url?u=http-3A__spectrumscale.org&d=DwMFaQ&c=jf_iaSHvJObTbx-siA1ZOg&r=WDtkF9zLTGGYqFnVnJ3rywZM6KHROA4FpMYi6cUkkKY&m=hKtOnoUDijNQoFnSlxQfek9m6h2qKbqjcCswbjHg2-E&s=n5P1NWESV2GUb3EXICXGj62_QDAPfSAWVPz_i59CNKk&e=> > http://gpfsug.org/mailman/listinfo/gpfsug-discuss > > > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at spectrumscale.org > <https://urldefense.proofpoint.com/v2/url?u=http-3A__spectrumscale.org&d=DwMFaQ&c=jf_iaSHvJObTbx-siA1ZOg&r=WDtkF9zLTGGYqFnVnJ3rywZM6KHROA4FpMYi6cUkkKY&m=hKtOnoUDijNQoFnSlxQfek9m6h2qKbqjcCswbjHg2-E&s=n5P1NWESV2GUb3EXICXGj62_QDAPfSAWVPz_i59CNKk&e=> > http://gpfsug.org/mailman/listinfo/gpfsug-discuss > > > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at spectrumscale.org > https://urldefense.proofpoint.com/v2/url?u=http-3A__gpfsug. > org_mailman_listinfo_gpfsug-2Ddiscuss&d=DwICAg&c=jf_iaSHvJObTbx-siA1ZOg&r= > WDtkF9zLTGGYqFnVnJ3rywZM6KHROA4FpMYi6cUkkKY&m= > hKtOnoUDijNQoFnSlxQfek9m6h2qKbqjcCswbjHg2-E&s= > j7eYU1VnwYXrTnflbJki13EfnMjqAro0RdCiLkVrgzE&e= > > > > > > > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at spectrumscale.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss > >
_______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss
