Why not run the mgr and data networks for the new cluster in different
vlans.

That way you can leave the dhcp on both systems running and they will only
respond to the requests for their respective clusters.

Nick

On Sat., 8 Dec. 2018, 8:22 am Rich Sudlow <r...@nd.edu wrote:

> We do something slightly different by not using a DHCP pool but only allow
> DHCP to answer with MACs that are know.
>
>
> On 12/7/18 1:30 PM, David Johnson wrote:
> > Yes, only one can have a dynamic range.  In my case neither of them do,
> > since I manually paste the MAC addresses into the mac table.
> >
> > My issue with deleting first is when I deleted 16 mac addresses and then
> got
> > sidetracked and went home, those nodes later lost their lease and ended
> up getting
> > evicted from GPFS.  Not an issue if the nodes were otherwise idle, but
> they had
> > been marked to drain, jobs were still running on them.
> >
> >   — ddj
> >
> >> On Dec 7, 2018, at 1:25 PM, Kevin Keane <kke...@sandiego.edu
> >> <mailto:kke...@sandiego.edu>> wrote:
> >>
> >> Thank you, Dave. That is an interesting alternative approach; I might
> actually
> >> consider that.
> >>
> >> So you are saying that the old and new DHCP servers can run in
> parallel? I
> >> assume that they just can't both have dynamic ranges?
> >>
> >> I'm not sure I understand what the problem is with deleting a node
> first, and
> >> then adding it on the new system. Even if the lease expires, wouldn't
> it just
> >> reacquire the new one once you create it?
> >>
> >> _______________________________________________________________________
> >> Kevin Keane | Systems Architect | University of San Diego ITS |
> >> kke...@sandiego.edu <mailto:kke...@sandiego.edu>
> >> Maher Hall, 192 |5998 Alcalá Park | San Diego, CA 92110-2492 |
> 619.260.6859
> >>
> >> *REMEMBER! **_No one from IT at USD will ever ask to confirm or supply
> your
> >> password_*.
> >> These messages are an attempt to steal your username and password.
> Please do
> >> not reply to, click the links within, or open the attachments of these
> >> messages. Delete them!
> >>
> >>
> >>
> >>
> >> On Thu, Dec 6, 2018 at 4:43 PM <david_john...@brown.edu
> >> <mailto:david_john...@brown.edu>> wrote:
> >>
> >>     We’ve kept parallel clusters on the same network for nearly a year
> now
> >>     while transitioning to RH7 from CentOS 6.
> >>     Initially copied the hosts and nodelist and MAC tables into the new
> xcat
> >>     database.  Carefully controlled use of makedhcp so that nodes
> moving to
> >>     the new cluster were first added to new dhcp server and then
> deleted from
> >>     the old. (  Didn’t want a repeat of what happened when I left some
> deleted
> >>     from the old but not added to the new cluster and they lost their
> lease.
> >>     The postscript hardeths also helped.  ). Make new images and use
> nodeset
> >>     to point to them. Reboot and test.
> >>
> >>     Drawback is having to make parallel changes on both management
> servers all
> >>     the time, but we needed both clusters to access gpfs so it was a
> necessary
> >>     evil.
> >>
> >>       -- ddj
> >>     Dave Johnson
> >>
> >>     On Dec 6, 2018, at 4:23 PM, Kevin Keane <kke...@sandiego.edu
> >>     <mailto:kke...@sandiego.edu>> wrote:
> >>
> >>>     I'm in the middle of upgrading our existing HPC (from RHEL 6 to
> RHEL 7).
> >>>     I'm doing most of my testing on a separate "sandbox" test bed, but
> now
> >>>     I'm close to going live. I'm trying to figure out how to do this
> with
> >>>     minimal disruption.
> >>>
> >>>     My question: how can I install the new management node and keep it
> >>>     *almost* completely operational, without interfering with the
> existing
> >>>     cluster? Is it enough to disable DHCP, or do I need to do anything
> else?
> >>>
> >>>     How do I prevent DHCP from accidentally getting enabled before I'm
> ready?
> >>>     Is makedhcp responsible for that?
> >>>
> >>>     Step-by-step, here is what I plan to do:
> >>>
> >>>     - Set up the new management node, but keep it inactive.
> >>>     - Test
> >>>     - Bring down all compute nodes.
> >>>     - Via IPMI, reset all the compute nodes' BMC controllers to DHCP
> >>>     - Other migration steps (home directories, modifications on the
> storage
> >>>     node, etc.)
> >>>     - De-activate the old management node (but keep it running)
> >>>     - Activate the new management node.
> >>>     - Discover and boot compute nodes
> >>>
> >>>     Is there anything glaringly obvious that I overlooked?
> >>>
> >>>     Thanks!
> >>>
> >>>
>  _______________________________________________________________________
> >>>     Kevin Keane | Systems Architect | University of San Diego ITS |
> >>>     kke...@sandiego.edu <mailto:kke...@sandiego.edu>
> >>>     Maher Hall, 192 |5998 Alcalá Park | San Diego, CA 92110-2492 |
> 619.260.6859
> >>>
> >>>     *REMEMBER! **_No one from IT at USD will ever ask to confirm or
> supply
> >>>     your password_*.
> >>>     These messages are an attempt to steal your username and password.
> Please
> >>>     do not reply to, click the links within, or open the attachments
> of these
> >>>     messages. Delete them!
> >>>
> >>>
> >>>     _______________________________________________
> >>>     xCAT-user mailing list
> >>>     xCAT-user@lists.sourceforge.net <mailto:
> xCAT-user@lists.sourceforge.net>
> >>>     https://lists.sourceforge.net/lists/listinfo/xcat-user
> >>     _______________________________________________
> >>     xCAT-user mailing list
> >>     xCAT-user@lists.sourceforge.net <mailto:
> xCAT-user@lists.sourceforge.net>
> >>     https://lists.sourceforge.net/lists/listinfo/xcat-user
> >>
> >> _______________________________________________
> >> xCAT-user mailing list
> >> xCAT-user@lists.sourceforge.net <mailto:xCAT-user@lists.sourceforge.net
> >
> >> https://lists.sourceforge.net/lists/listinfo/xcat-user
> >
> >
> >
> > _______________________________________________
> > xCAT-user mailing list
> > xCAT-user@lists.sourceforge.net
> > https://lists.sourceforge.net/lists/listinfo/xcat-user
> >
>
>
> --
> Rich Sudlow
> University of Notre Dame
> Center for Research Computing - Union Station
> 506 W. South St
> South Bend, In 46601
>
> (574) 631-7258 (office)
> (574) 807-1046 (cell)
>
>
> _______________________________________________
> xCAT-user mailing list
> xCAT-user@lists.sourceforge.net
> https://lists.sourceforge.net/lists/listinfo/xcat-user
>
_______________________________________________
xCAT-user mailing list
xCAT-user@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/xcat-user

Reply via email to