Re: [lustre-discuss] Extending Lustre file system

2024-01-08 Thread Andreas Dilger via lustre-discuss
I would recommend *against* mounting all 175 OSTs at the same time. There are (or at least were*) some issues with the MGS registration RPCs timing out when too many config changes happen at once. Your "mount and wait 2 sec" is more robust and doesn't take very much time (a few minutes) vs.

Re: [lustre-discuss] Extending Lustre file system

2024-01-08 Thread Thomas Roth via lustre-discuss
Yes, sorry, I meant the actual procedure of mounting the OSTs for the first time. Last year I did that with 175 OSTs - replacements for EOL hardware. All OSTs had been formatted with a specific index, so probably creating a suitable /etc/fstab everywhere and sending a 'mount -a -t lustre' to

Re: [lustre-discuss] [EXTERNAL] [BULK] MDS hardware - NVME?

2024-01-08 Thread Jeff Johnson
Today nvme/mdraid/ldiskfs will beat nvme/zfs on MDS IOPs but you can close the gap somewhat with tuning, zfs ashift/recordsize and special allocation class vdevs. While the IOPs performance favors nvme/mdraid/ldiskfs there are tradeoffs. The snapshot/backup abilities of ZFS and the security it

Re: [lustre-discuss] Extending Lustre file system

2024-01-08 Thread Andreas Dilger via lustre-discuss
The need to rebalance depends on how full the existing OSTs are. My recommendation if you know that the data will continue to grow is to add new OSTs when the existing ones are at 60-70% full, and add them in larger groups rather than one at a time. Cheers, Andreas > On Jan 8, 2024, at

Re: [lustre-discuss] Extending Lustre file system

2024-01-08 Thread Thomas Roth via lustre-discuss
Just mount the OSTs, one by one and perhaps not if your system is heavily loaded. Follow what happens in the MDS log and the OSS log. And try to rebalance the OSTs fill levels afterwards - very empty OSTs will attract all new files, which might be hot and direct your users's fire to your new

Re: [lustre-discuss] [EXTERNAL] [BULK] MDS hardware - NVME?

2024-01-08 Thread Thomas Roth via lustre-discuss
Hi Cameron, did you run a performance comparison between ZFS and mdadm-raid on the MDTs? I'm currently doing some tests, and the results favor software raid, in particular when it comes to IOPS. Regards Thomas On 1/5/24 19:55, Cameron Harr via lustre-discuss wrote: This doesn't answer your

Re: [lustre-discuss] [EXTERNAL] [BULK] MDS hardware - NVME?

2024-01-08 Thread Vicker, Darby J. (JSC-EG111)[Jacobs Technology, Inc.] via lustre-discuss
Our setup has a single JBOD connected to 2 servers but the JBOD has dual controllers. Each server connects to both controllers for redundancy so there are 4 connections to each server. So we have a paired HA setup where one peer node can take over the OSTs/MDTs of its peer node. Some

[lustre-discuss] Extending Lustre file system

2024-01-08 Thread Backer via lustre-discuss
Hi, Good morning and happy new year! I have a quick question on extending a lustre file system. The extension is performed online. I am looking for any best practices or anything to watchout while doing the file system extension. The file system extension is done adding new OSS and many OSTs