Jeremy Mann wrote:
Nathaniel Rutman wrote:

You have to deactivate the OSCs that reference that OST.  From
https://mail.clusterfs.com/wikis/lustre/MountConf:

As of beta7, an OST can be permanently removed from a filesystem. Note
that any files that have stripes on the removed OST will henceforth
return EIO.

mgs> lctl conf_param testfs-OST0001.osc.active=0

Thanks Nathaniel I found it shortly after posting the message. However,
maybe I didn't do it right but I still get error messages about this node.
The steps I took were:

1. umounted /lustre from the frontend
2. umounted /lustre-storage on node17
3. on frontend, ran lctl conf_param bcffs-OST0002.osc.active=0
4. on frontend, remounted bcffs with:

mount -o exclude=bcffs-OST0002 -t lustre [EMAIL PROTECTED]:/bcffs /lustre

All you need is step 3 on the MGS. You don't have to remount clients or use -o exclude. But that won't hurt anything.


dmesg shows:

Lustre: 31233:0:(quota_master.c:1105:mds_quota_recovery()) Not all osts
are active, abort quota recovery
Apparently quotas don't work with deactivated OSTs. I don't know much about this area, but I suspect that to get rid of these messages, you'll need to have everything active.
Lustre: MDS bcffs-MDT0000: bcffs-OST000c_UUID now active, resetting orphans
This looks like you restarted the MDT.
LustreError: 32542:0:(file.c:1012:ll_glimpse_size()) obd_enqueue returned
rc -5, returning -EIO
Lustre: client 000001017ac3fc00 umount complete
And here the client stopped
Lustre: 1257:0:(obd_mount.c:1675:lustre_check_exclusion()) Excluding
bcffs-OST0002-osc (on exclusion list)
Lustre: 1257:0:(recover.c:231:ptlrpc_set_import_active()) setting import
bcffs-OST0002_UUID INACTIVE by administrator request
That's the -o exlude
Lustre: osc.: set active=0 to 0
LustreError: 1257:0:(lov_obd.c:139:lov_connect_obd()) not connecting OSC
bcffs-OST0002_UUID; administratively disabled
And that's the osc.active=0. You only need one or the other, but both won't break anything.
Lustre: Client bcffs-client has started
And here the client restarted
Lustre: 2484:0:(quota_master.c:1105:mds_quota_recovery()) Not all osts are
active, abort quota recovery
Lustre: MDS bcffs-MDT0000: bcffs-OST000d_UUID now active, resetting orphans
Lustre: MGS: haven't heard from client
454dd520-82b9-e3e6-8fcb-800a75807121 (at [EMAIL PROTECTED]) in 228
seconds. I think it's dead, and I am evicting it.
The MGS eviction of an MGC is a non-destructive event, and I should turn off this scary message. The MGC will re-acquire an MGS lock later. This happened here because (I theorize) that you had a combined MGS/MDT that you restarted, while other Lustre devices were still mounted on the same node. The restart of the MGS means that all live MGC's must get kicked out and reconnect.
LustreError: 4834:0:(file.c:1012:ll_glimpse_size()) obd_enqueue returned
rc -5, returning -EIO
This is the only potential worrying error message, but depending on who generated it, it might be fine.
Are these normal error messages? I'm asking because I'm about to copy all
of the NCBI databases to the lustre filesystem. I don't want to start it,
then have Lustre crash and have to rebuild everything all over again minus
this node.

You can use the "writeconf" procedure described on the wiki to remove all traces of the removed OST. This does not require reformatting anything and will likely fix the quota message.

_______________________________________________
Lustre-discuss mailing list
[email protected]
https://mail.clusterfs.com/mailman/listinfo/lustre-discuss

Reply via email to