This is a slightly more serious issue than you think. If you can send me the exact sql being sent (Do you know where the wacky '��8' came from?) perhaps it will be clearer where the problem lies. With this situation your DBD will continue to back up information and new info will not be added to the database so it will appear out of sync.

On 03/14/13 09:42, Chris Read wrote:
Re: [slurm-dev] Re: Duplicate jobs in the grid_job_table
We've just had a whole load of nodes fail, and a lot of jobs to rescheduled which recreated the problem of sacct not agreeing with squeue.

Found a load of these in the slurmdbd.log:

[2013-03-14T09:33:39-05:00] error: mysql_query failed: 1064 You have an error in your SQL syntax; check the manual that corresponds to your MySQL server version for the right syntax to use near '��8 ', 'drw', 'short', '**', '49') on duplicate key update job_db_inx=LAST_INSER' at line 1 insert into "grid_job_table" (id_job, id_assoc, id_qos, id_wckey, id_user, id_group, nodelist, id_resv, timelimit, time_eligible, time_submit, time_start, job_name, track_steps, state, priority, cpus_req, cpus_alloc, nodes_alloc, gres_req, gres_alloc, account, partition, wckey, node_inx) values (39641248, 139, 1, 0, 7045, 7045, 'n23', 0, 10, 1363271250, 1363271250, 1363271251, '<JOB_NAME>', 0, 3, 2549, 1, 1, 1, '��i�8 ', '�'��8 ', '<acct>', 'short', '**', '49') on duplicate key update job_db_inx=LAST_INSERT_ID(job_db_inx), id_wckey=0, id_user=7045, id_group=7045, nodelist='n23', id_resv=0, timelimit=10, time_submit=1363271250, time_start=1363271251, job_name='<JOB_NAME>', track_steps=0, id_qos=1, state=greatest(state, 3), priority=2549, cpus_req=1, cpus_alloc=1, nodes_alloc=1, gres_req='��i�8 ', gres_alloc='�'��8 ', account='<acct>', partition='short', wckey='**', node_inx='49'

So I'm guessing that's how the confusing data got in there in the first place.



On Wed, Mar 13, 2013 at 8:43 PM, Danny Auble <[email protected] <http://lists.schedmd.com/cgi-bin/dada/mail.cgi/r/slurmdev/263544490268/>> wrote:

    I would make sure these jobs weren't requeued.  Knowing what the
    times were of the entries in the database would be interesting as
    well.  Any information about the jobs in the slurmctld log would
    probably shed information on the matter.  Outside of being
    requeued I wouldn't ever expect duplicates.

    Danny

    On 03/12/13 10:57, Chris Read wrote:
    Update:

    I've just cleaned things up by deleting the duplicates where
    state = 0 (PENDING). The correct state for the job is actually 7
    (NODE_FAIL), not CANCELLED as I stated above.

    No need to restart slurmdbd either...

    Chris


    On Tue, Mar 12, 2013 at 4:28 PM, Chris Read <[email protected]
    <http://lists.schedmd.com/cgi-bin/dada/mail.cgi/r/slurmdev/062153156925/>>
    wrote:

        Forgot to add another question:

        What's the correct way to clean this up? Just delete the
        record showing PENDING and restart slurmdbd?


        On Tue, Mar 12, 2013 at 4:26 PM, Chris Read
        <[email protected]
        
<http://lists.schedmd.com/cgi-bin/dada/mail.cgi/r/slurmdev/062153156925/>>
        wrote:

            Greetings...

            Just stumbled across some strange behaviour on the
            accounting side of things: we have a collection of jobs
            that have duplicate records in the grid_job_table. The
            visible symptoms of this are that sacct shows the jobs as
            still pending when they are not.

            In all of the cases I can find:

            - there is no information available in the slurmdbd.log
            - the slurmctld.log shows the jobs have been canceled
            - the job_db_inx for the entry in PENDING state is > the
            job_db_inx for the entry in CANCELLED

            We're currently on 2.5.1.

            Anyone have any idea how these got there?

            Chris






Reply via email to