Jonathan,
my suspicion is that the GPFS daemon on fqdn-new is not reachable via port 1191.
You can double check that by sending a lightweight CCR RPC to this daemon from another quorum node by attempting:
mmccr echo -n fqdn-new;echo $?
If this echo returns with a non-zero exit code the network settings must be verified. And even the other direction must
work: Node fqdn-new must reach another quorum node, like (attempting on fqdn-new):
mmccr echo -n <QUORUM_NODE>;echo $?
Mit freundlichen Grüßen / Kind regards
| Ralf Eberhard Spectrum Scale Developer IBM Systems - Dept. M929 Mobile: +49 162 4159476 E-Mail: [email protected] |
IBM Deutschland Research & Development GmbH
Vorsitzender des Aufsichtsrats: Gregor Pillen
Geschäftsführung: Dirk Wittkopp
Sitz der Gesellschaft: Böblingen / Registergericht: Amtsgericht Stuttgart, HRB 243294
IBM Data Privacy Statement
Vorsitzender des Aufsichtsrats: Gregor Pillen
Geschäftsführung: Dirk Wittkopp
Sitz der Gesellschaft: Böblingen / Registergericht: Amtsgericht Stuttgart, HRB 243294
IBM Data Privacy Statement
[email protected] wrote: -----
To: [email protected]
From: [email protected]
Sent by: [email protected]
Date: 12/09/2021 05:04PM
Subject: [EXTERNAL] gpfsug-discuss Digest, Vol 119, Issue 7
From: [email protected]
Sent by: [email protected]
Date: 12/09/2021 05:04PM
Subject: [EXTERNAL] gpfsug-discuss Digest, Vol 119, Issue 7
Send gpfsug-discuss mailing list submissions to
[email protected]
To subscribe or unsubscribe via the World Wide Web, visit
http://gpfsug.org/mailman/listinfo/gpfsug-discuss
or, via email, send a message with subject or body 'help' to
[email protected]
You can reach the person managing the list at
[email protected]
When replying, please edit your Subject line so it is more specific
than "Re: Contents of gpfsug-discuss digest..."
Today's Topics:
1. Re: alternate path between ESS Servers for Datamigration
(Olaf Weiser)
2. Adding a quorum node (Jonathan Buzzard)
3. Re: alternate path between ESS Servers for Datamigration
(Douglas O'flaherty)
----------------------------------------------------------------------
Message: 1
Date: Thu, 9 Dec 2021 12:04:28 +0000
From: "Olaf Weiser" <[email protected]>
To: [email protected]
Cc: [email protected]
Subject: Re: [gpfsug-discuss] alternate path between ESS Servers for
Datamigration
Message-ID:
<of942131bb.73c8972f-on002587a6.0041d89c-002587a6.00425...@ibm.com>
Content-Type: text/plain; charset="us-ascii"
An HTML attachment was scrubbed...
URL: <http://gpfsug.org/pipermail/gpfsug-discuss/attachments/20211209/33805b31/attachment-0001.html >
------------------------------
Message: 2
Date: Thu, 9 Dec 2021 12:36:08 +0000
From: Jonathan Buzzard <[email protected]>
To: "'[email protected]'"
<[email protected]>
Subject: [gpfsug-discuss] Adding a quorum node
Message-ID: <[email protected]>
Content-Type: text/plain; charset=UTF-8; format=flowed
I am looking to replace the quorum node in our cluster. The RAID card in
the server we are currently using is a casualty of the RHEL8 SAS card
purge :-(
I have a "new" dual core server that is fully supported by RHEL8. After
some toing and throwing with IBM they agreed a Pentium G6400 is 70PVU a
core and two cores :-) That said it is currently running RHEL7 because
that's what the DSS-G nodes are running. The upgrade to RHEL8 is planned
for next year.
Anyway I have added it into the GPFS cluster all well and good and GPFS
is mounted just fine. However when I ran the command to make it a quorum
node I got the following error (sanitized to remove actual DNS names and
IP addresses
initialize (113, '<fqdn-new>', ('<IP address>', 1191)) failed (err 79)
server initialization failed (err 79)
mmchnode: Unexpected error from chnodes -n
1=<fqdn1>:1191,2<fqdn2>:1191,3=<fqdn3>:1191,113=<fqdn-new>:1191 -f 1 -P
1191 . Return code: 149
mmchnode: Unable to change the CCR quorum node configuration.
mmchnode: Command failed. Examine previous error messages to determine
cause.
fqdn-new is the new node and fqdn1/2/3 are the existing quorum nodes. I
want to remove fqdn3 in due course.
Anyone any idea what is going on? I thought you could change the quorum
nodes on the fly?
JAB.
--
Jonathan A. Buzzard Tel: +44141-5483420
HPC System Administrator, ARCHIE-WeSt.
University of Strathclyde, John Anderson Building, Glasgow. G4 0NG
------------------------------
Message: 3
Date: Thu, 9 Dec 2021 16:04:28 +0000
From: "Douglas O'flaherty" <[email protected]>
To: "[email protected]"
<[email protected]>
Subject: Re: [gpfsug-discuss] alternate path between ESS Servers for
Datamigration
Message-ID:
<[email protected]>
Content-Type: text/plain; charset="utf-8"
Walter:
Though not directly about your design, our work with NVIDIA on GPUdirect Storage and SuperPOD has shown how sensitive RDMA (IB & RoCE) to both MOFED and Firmware version compatibility can be.
I would suggest anyone debugging RDMA issues should look at those closely.
Doug
by carrier pigeon
On Dec 9, 2021, 5:04:36 AM, [email protected] wrote:
From: [email protected]
To: [email protected]
Cc:
Date: Dec 9, 2021, 5:04:36 AM
Subject: [EXTERNAL] gpfsug-discuss Digest, Vol 119, Issue 5
Send gpfsug-discuss mailing list submissions to [email protected] subscribe or unsubscribe via the World Wide Web, visit http://gpfsug.org/mailman/listinfo/gpfsug-discuss or, via email, send a message with subject or body 'help' to [email protected] can reach the person managing the list at [email protected] replying, please edit your Subject line so it is more specificthan "Re: Contents of gpfsug-discuss digest..."
Send gpfsug-discuss mailing list submissions to [email protected] subscribe or unsubscribe via the World Wide Web, visit http://gpfsug.org/mailman/listinfo/gpfsug-discuss or, via email, send a message with subject or body 'help' to [email protected] can reach the person managing the list at [email protected] replying, please edit your Subject line so it is more specificthan "Re: Contents of gpfsug-discuss digest..."Today's Topics: 1. alternate path between ESS Servers for Datamigration (Walter Sklenka)
Dear spectrum scale users! May I ask you a design question? We have an IB environment which is very mixed at the moment ( connecX3 ? connect-X6 with FDR , even FDR10 and with arrive of ESS5000SC7 now also HDR100 and HDR switches. We still have some big troubles in this fabric when using RDMA , a case at Mellanox and IBM is open . The environment has 3 old Building blocks 2xESSGL6 and 1x GL4 , from where we want to migrate the data to ess5000 , ( mmdelvdisk +qos) Due to the current problems with RDMA we though eventually we could try a workaround : If you are interested there is Maybe you can find the attachment ? We build 2 separate fabrics , the ess-IO servers attached to both blue and green and all other cluster members and all remote clusters only to fabric blue The daemon interfaces (IPoIP) are on fabric blue It is the aim to setup rdma only on the ess-ioServers in the fabric green , in the blue we must use IPoIB (tcp) Do you think datamigration would work be
tween ess01,ess02,? to ess07,ess08 via RDMA ? Or is it principally not possible to make a rdma network only for a subset of a cluster (though this subset would be reachable via other fabric) ? Thank you very much for any input ! Best regards walter Mit freundlichen Gr??en
Walter Sklenka
Technical Consultant EDV-Design Informationstechnologie GmbH
Giefinggasse 6/1/2, A-1210 Wien
Tel: +43 1 29 22 165-31
Fax: +43 1 29 22 165-90
E-Mail: [email protected]
Internet: www.edv-design.at
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://gpfsug.org/pipermail/gpfsug-discuss/attachments/20211209/f2fa4bf1/attachment.html >
------------------------------
_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss
End of gpfsug-discuss Digest, Vol 119, Issue 7
**********************************************
[email protected]
To subscribe or unsubscribe via the World Wide Web, visit
http://gpfsug.org/mailman/listinfo/gpfsug-discuss
or, via email, send a message with subject or body 'help' to
[email protected]
You can reach the person managing the list at
[email protected]
When replying, please edit your Subject line so it is more specific
than "Re: Contents of gpfsug-discuss digest..."
Today's Topics:
1. Re: alternate path between ESS Servers for Datamigration
(Olaf Weiser)
2. Adding a quorum node (Jonathan Buzzard)
3. Re: alternate path between ESS Servers for Datamigration
(Douglas O'flaherty)
----------------------------------------------------------------------
Message: 1
Date: Thu, 9 Dec 2021 12:04:28 +0000
From: "Olaf Weiser" <[email protected]>
To: [email protected]
Cc: [email protected]
Subject: Re: [gpfsug-discuss] alternate path between ESS Servers for
Datamigration
Message-ID:
<of942131bb.73c8972f-on002587a6.0041d89c-002587a6.00425...@ibm.com>
Content-Type: text/plain; charset="us-ascii"
An HTML attachment was scrubbed...
URL: <http://gpfsug.org/pipermail/gpfsug-discuss/attachments/20211209/33805b31/attachment-0001.html >
------------------------------
Message: 2
Date: Thu, 9 Dec 2021 12:36:08 +0000
From: Jonathan Buzzard <[email protected]>
To: "'[email protected]'"
<[email protected]>
Subject: [gpfsug-discuss] Adding a quorum node
Message-ID: <[email protected]>
Content-Type: text/plain; charset=UTF-8; format=flowed
I am looking to replace the quorum node in our cluster. The RAID card in
the server we are currently using is a casualty of the RHEL8 SAS card
purge :-(
I have a "new" dual core server that is fully supported by RHEL8. After
some toing and throwing with IBM they agreed a Pentium G6400 is 70PVU a
core and two cores :-) That said it is currently running RHEL7 because
that's what the DSS-G nodes are running. The upgrade to RHEL8 is planned
for next year.
Anyway I have added it into the GPFS cluster all well and good and GPFS
is mounted just fine. However when I ran the command to make it a quorum
node I got the following error (sanitized to remove actual DNS names and
IP addresses
initialize (113, '<fqdn-new>', ('<IP address>', 1191)) failed (err 79)
server initialization failed (err 79)
mmchnode: Unexpected error from chnodes -n
1=<fqdn1>:1191,2<fqdn2>:1191,3=<fqdn3>:1191,113=<fqdn-new>:1191 -f 1 -P
1191 . Return code: 149
mmchnode: Unable to change the CCR quorum node configuration.
mmchnode: Command failed. Examine previous error messages to determine
cause.
fqdn-new is the new node and fqdn1/2/3 are the existing quorum nodes. I
want to remove fqdn3 in due course.
Anyone any idea what is going on? I thought you could change the quorum
nodes on the fly?
JAB.
--
Jonathan A. Buzzard Tel: +44141-5483420
HPC System Administrator, ARCHIE-WeSt.
University of Strathclyde, John Anderson Building, Glasgow. G4 0NG
------------------------------
Message: 3
Date: Thu, 9 Dec 2021 16:04:28 +0000
From: "Douglas O'flaherty" <[email protected]>
To: "[email protected]"
<[email protected]>
Subject: Re: [gpfsug-discuss] alternate path between ESS Servers for
Datamigration
Message-ID:
<[email protected]>
Content-Type: text/plain; charset="utf-8"
Walter:
Though not directly about your design, our work with NVIDIA on GPUdirect Storage and SuperPOD has shown how sensitive RDMA (IB & RoCE) to both MOFED and Firmware version compatibility can be.
I would suggest anyone debugging RDMA issues should look at those closely.
Doug
by carrier pigeon
On Dec 9, 2021, 5:04:36 AM, [email protected] wrote:
From: [email protected]
To: [email protected]
Cc:
Date: Dec 9, 2021, 5:04:36 AM
Subject: [EXTERNAL] gpfsug-discuss Digest, Vol 119, Issue 5
Send gpfsug-discuss mailing list submissions to [email protected] subscribe or unsubscribe via the World Wide Web, visit http://gpfsug.org/mailman/listinfo/gpfsug-discuss or, via email, send a message with subject or body 'help' to [email protected] can reach the person managing the list at [email protected] replying, please edit your Subject line so it is more specificthan "Re: Contents of gpfsug-discuss digest..."
Send gpfsug-discuss mailing list submissions to [email protected] subscribe or unsubscribe via the World Wide Web, visit http://gpfsug.org/mailman/listinfo/gpfsug-discuss or, via email, send a message with subject or body 'help' to [email protected] can reach the person managing the list at [email protected] replying, please edit your Subject line so it is more specificthan "Re: Contents of gpfsug-discuss digest..."Today's Topics: 1. alternate path between ESS Servers for Datamigration (Walter Sklenka)
Dear spectrum scale users! May I ask you a design question? We have an IB environment which is very mixed at the moment ( connecX3 ? connect-X6 with FDR , even FDR10 and with arrive of ESS5000SC7 now also HDR100 and HDR switches. We still have some big troubles in this fabric when using RDMA , a case at Mellanox and IBM is open . The environment has 3 old Building blocks 2xESSGL6 and 1x GL4 , from where we want to migrate the data to ess5000 , ( mmdelvdisk +qos) Due to the current problems with RDMA we though eventually we could try a workaround : If you are interested there is Maybe you can find the attachment ? We build 2 separate fabrics , the ess-IO servers attached to both blue and green and all other cluster members and all remote clusters only to fabric blue The daemon interfaces (IPoIP) are on fabric blue It is the aim to setup rdma only on the ess-ioServers in the fabric green , in the blue we must use IPoIB (tcp) Do you think datamigration would work be
tween ess01,ess02,? to ess07,ess08 via RDMA ? Or is it principally not possible to make a rdma network only for a subset of a cluster (though this subset would be reachable via other fabric) ? Thank you very much for any input ! Best regards walter Mit freundlichen Gr??en
Walter Sklenka
Technical Consultant EDV-Design Informationstechnologie GmbH
Giefinggasse 6/1/2, A-1210 Wien
Tel: +43 1 29 22 165-31
Fax: +43 1 29 22 165-90
E-Mail: [email protected]
Internet: www.edv-design.at
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://gpfsug.org/pipermail/gpfsug-discuss/attachments/20211209/f2fa4bf1/attachment.html >
------------------------------
_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss
End of gpfsug-discuss Digest, Vol 119, Issue 7
**********************************************
_______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss
