Hi, We have a GPFS cluster where the NSD servers mount the storage over fibre channel and export the file system over InfiniBand for clients.
We will be getting some used equipment that uses OmniPath. As per the "IBM Storage Scale Frequently Asked Questions and Answers" it states [1]: > RDMA is not supported on a node when both Mellanox HCAs and Cornelis Networks > Omni-Path HFIs are enabled for RDMA. Does this mean that we wouldn't be able to consolidate both IB & OPA HCA's in NSD servers and would have to have 2 types of NSD servers? 1) InfiniBand exporting and 2) OmniPath exporting? If so, is it then a matter of using the Multi-Rail over TCP "subnets =" setting in mmchonfig to distinguish which nsd server the clients should connect to? [2]. Or am I completely miss understanding all this? Many thanks in advance. Sean [1] https://www.ibm.com/docs/en/STXKQY/pdf/gpfsclustersfaq.pdf [2] https://www.ibm.com/docs/en/storage-scale/5.1.6?topic=configuring-multi-rail-over-tcp-mrot --- Sean McGrath [email protected] Senior Systems Administrator Research IT, IT Services, Trinity College Dublin https://www.tcd.ie/itservices/ https://www.tchpc.tcd.ie/ _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org
