which also means starting in 17.02

in "RPMs INSTALLED" section of https://slurm.schedmd.com/quic
kstart_admin.html should be revised as follows

slurm-sjobexit
slurm-sjstat

should be changed to

slurm-sjobexit (only prior to 17.02)
slurm-sjstat (only prior to 17.02)

On Mon, Feb 27, 2017 at 9:56 AM, <[email protected]> wrote:

> Thanks for the patch. Committed here:
> https://github.com/SchedMD/slurm/commit/95cf960afcdb77cae644
> b7d0709ede123896626d
>
> ----- Message from Daniel Letai <[email protected]> ---------
>     Date: Mon, 27 Feb 2017 05:07:08 -0800
>     From: Daniel Letai <[email protected]>
> Reply-To: slurm-dev <[email protected]>
> Subject: [slurm-dev] Re: Slurm version 17.02.0 is now available
>       To: slurm-dev <[email protected]>
>
> $ git diff
> diff --git a/slurm.spec b/slurm.spec
> index 941b360..6bb3014 100644
> --- a/slurm.spec
> +++ b/slurm.spec
> @@ -346,6 +346,7 @@ Includes the Slurm proctrack/lua and job_submit/lua
> plugin
>  Summary: Perl tool to print Slurm job state information
>  Group: Development/System
>  Requires: slurm
> +Obsoletes: slurm-sjobexit slurm-sjstat slurm-seff
>  %description contribs
>  seff is a mail program used directly by the Slurm daemons. On completion
> of a
>  job, wait for it's accounting information to be available and include that
>
> On 02/27/2017 01:35 PM, dani wrote:
>
> Seems like no obsoletes was set on slurm-contribs, so yum complains of
> conflicts with slurm-sjobs and friends.
>
>
> On 24/02//2017 01:41, Danny Auble wrote:
>
>
> After 9 months of development we are pleased to announce the availability
> of Slurm version 17.02.0.
>
> A brief description of what is contained in this release and other notes
> about it is contained below.  For a fuller description please consult the
> RELEASE_NOTES file available in the source.
>
> Thanks to all involved!
>
> Slurm downloads are available from https://schedmd.com/downloads.php.
>
> RELEASE NOTES FOR SLURM VERSION 17.02
> 23 February 2017
>
> IMPORTANT NOTES:
> THE MAXJOBID IS NOW 67,108,863. ANY PRE-EXISTING JOBS WILL CONTINUE TO RUN
> BUT
> NEW JOB IDS WILL BE WITHIN THE NEW MAXJOBID RANGE. Adjust your configured
> MaxJobID value as needed to eliminate any confusion.
>
> If using the slurmdbd (Slurm DataBase Daemon) you must update this first.
> The 17.02 slurmdbd will work with Slurm daemons of version 15.08 and above.
> You will not need to update all clusters at the same time, but it is very
> important to update slurmdbd first and having it running before updating
> any other clusters making use of it.  No real harm will come from updating
> your systems before the slurmdbd, but they will not talk to each other
> until you do.  Also at least the first time running the slurmdbd you need
> to
> make sure your my.cnf file has innodb_buffer_pool_size equal to at least
> 64M.
> You can accomplish this by adding the line
>
> innodb_buffer_pool_size=64M
>
> under the [mysqld] reference in the my.cnf file and restarting the mysqld.
> The
> buffer pool size must be smaller than the size of the MySQL tmpdir. This is
> needed when converting large tables over to the new database schema.
>
> Slurm can be upgraded from version 15.08 or 16.05 to version 17.02 without
> loss
> of jobs or other state information. Upgrading directly from an earlier
> version
> of Slurm will result in loss of state information.
>
> If using SPANK plugins that use the Slurm APIs, they should be recompiled
> when
> upgrading Slurm to a new major release.
>
> NOTE: systemd services files are installed automatically, but not enabled.
>       You will need to manually enable them on the appropriate systems:
>       - Controller: systemctl enable slurmctld
>       - Database: systemctl enable slurmdbd
>       - Compute Nodes: systemctl enable slurmd
>
> NOTE: If you are not using Munge, but are using the "service" scripts to
>       start Slurm daemons, then you will need to remove this check from the
>       etc/slurm*service scripts.
>
> NOTE: If you are upgrading with any jobs from 14.03 or earlier
>       (i.e. quick upgrade from 14.03 -> 15.08 -> 17.02) you will need
>       to wait until after those jobs are gone before you upgrade to 17.02.
>
> HIGHLIGHTS
> ==========
>  -- Added infrastructure for managing workload across a federation of
> clusters.
>     (partial functionality in version 17.02, fully operational in May 2017)
>  -- In order to support federated jobs, the MaxJobID configuration
> parameter
>     default value has been reduced from 2,147,418,112 to 67,043,328 and its
>     maximum value is now 67,108,863. Upon upgrading, any pre-existing jobs
> that
>     have a job ID above the new range will continue to run and new jobs
> will get
>     job IDs in the new range.
>  -- Added "MailDomain" configuration parameter to qualify email addresses.
>  -- Automatically clean up task/cgroup cpuset and devices cgroups after
> steps
>     are completed.
>  -- Added burst buffer support for job arrays. Added new
> SchedulerParameters
>     configuration parameter of bb_array_stage_cnt=# to indicate how many
> pending
>     tasks of a job array should be made available for burst buffer resource
>     allocation.
>  -- Added new sacctmgr commands: "shutdown" (shutdown the server), "list
> stats"
>     (get server statistics) "clear stats" (clear server statistics).
>  -- The database index for jobs is now 64 bits.  If you happen to be close
> to
>     4 billion jobs in your database you will want to update your slurmctld
> at
>     the same time as your slurmdbd to prevent roll over of this variable as
>     it is 32 bit previous versions of Slurm.
>  -- All memory values (in MB) are now 64 bit. Previously, nodes with more
> than
>     of memory would not schedule or enforce memory limits correctly.
>  -- Removed AIX, BlueGene/L and BlueGene/P support.
>  -- Removed sched/wiki and sched/wiki2 plugins and associated code.
>  -- Added PrologFlags=Serial to disable concurrent execution of
> prolog/epilog
>     scripts.
>
>
>
>
>
>
>
> ----- End message from Daniel Letai <[email protected]> ----
>

Reply via email to