Dear Nate and Curtis,

I read a bit in the documentation of the python and underlying C library.. and played a bit around.

I can't speak for GridEngine's specific behavior because I haven't used it in a long time, but it's not surprising that jobs "disappear" as soon as they've exited. Unfortunately, Galaxy uses periodic polling rather than waiting on completion. We'd need to create a thread-per-submitted job unless you can still get job exit details by looping over jobs with a timeout wait.
> > You can gain some control over how Galaxy handles InvalidJobException
exceptions with drmaa job runner plugin params, see here:

However, if normally finished jobs also result in InvalidJobException, that probably won't help. Alternatively, you could create a DRMAAJobRunner subclass for GridEngine like we've done for Slurm that does some digging to learn more about terminal jobs.

I found a way to get the information. The problem in my script was that wait() "reaps" (that the term used by python-drmaa) all information on the job from the session(?). Hence calls to jobStatus after wait() will fail. The solution here would be to use synchronize() with parameter dispose=False, see attached file -- alternatively one can also wait until job status is DONE or FAILED.

But this seems not to be the source of the problem within Galaxy, since it never calls wait(..). The problem seems to be that an external python script submits the job in another session (when jobs are submitted as real user). The problem is then that jobs created in another session can not be queried (The documentation is a bit vague here: in the C library documentation of SGE I read that it is definitely not possible to query finished jobs).

I tried if it is possible to pickle the session -- without success.

Does anyone has an idea how one could get the active drmaa session from galaxy to the external script?

> We have had this problem on our SGE based installation for years. We
> referred to it as the "green screen of death" - as it would allow a
> biologist to continue analysis using output that was partial, at best,
> often resulting in seemingly successful completion of the entire
> analysis, but completely bogus results (say, cuffdiff killed half way
> through the genome, but it's green in galaxy, so no transcripts on the
> smaller chromosomes, but no error, either).

Did you use submission as real user? Or does the problem also appear if jobs are submitted as the single user running galaxy.

> We ended up implementing an external reaper that detected these killed
> jobs from SGE, and notified the user and galaxy post-hoc. It was not a
> very satisfactory solution. We are currently moving to SLURM for other
> reasons and hope the problem will not be present there.

I was also thinking about not using the python library at all, but using system calls to qsub, qkill, qacct, etc? Any opinions on this idea?

I guess your reaper could be of interest also for others. Is this available somewhere?



On Thu, Jun 15, 2017 at 10:27 AM, Matthias Bernt < <>> wrote:

    Dear list,

    I have two question for all DRMAA users. Here is the first one.

    I was checking how our queuing system (univa GridEngine) and Galaxy
    react if jobs are submitted that exceed run time or memory limits.

    I found out that the python drmaa library cannot query the job
    status after the job is finished (for both successful and
    unsuccessful jobs).

    In lib/galaxy/jobs/runners/ the call gives an exception
         self.ds.job_status( external_job_id )

    Is this always the case? Or might this be a problem with our GridEngine?

    I have attached some code for testing. Here the first call to
    s.jobStatus(jobid) works, but the second after s.wait(...) doesn't.
    But I get "drmaa.errors.InvalidJobException: code 18: The job
    specified by the 'jobid' does not exist."

    The same error pops up in the Galaxy logs. The consequence is that
    jobs that reached the limits are shown as completed successfully in

    Interestingly, quite a bit of information can be obtained from the
    return value of s.wait. I was wondering if this can be used to
    differentiate successful from failed jobs. In particular hasExited,
    hasSignal, and terminateSignal are different in the two cases.


    Please keep all replies on the list by using "reply all"
    in your mail client.  To manage your subscriptions to this
    and other Galaxy lists, please use the interface at: <>

    To search Galaxy mailing lists use the unified search at: <>

#!/usr/bin/env python

from __future__ import print_function
import drmaa
import os
import time

def main():
   """Submit a job.
   Note, need file called in current directory.
   with drmaa.Session() as s:
       print('Creating job template')
       jt = s.createJobTemplate()
       jt.jobName = "foo"
       jt.workingDirectory = "/homes/bierdepot/maze/tmp/"
       jt.remoteCommand = '/homes/bierdepot/maze/tmp/'
       jt.args = ['10','Simon_says']
       jt.nativeSpecification = "-l h_rt=10 -l h_vmem=1G -pe smp 2 -w n"
       jobid = s.runJob(jt)
       print('Your job has been submitted with id ' + jobid)

       decodestatus = {drmaa.JobState.UNDETERMINED: 'process status cannot be determined',
                        drmaa.JobState.QUEUED_ACTIVE: 'job is queued and active',
                        drmaa.JobState.SYSTEM_ON_HOLD: 'job is queued and in system hold',
                        drmaa.JobState.USER_ON_HOLD: 'job is queued and in user hold',
                        drmaa.JobState.USER_SYSTEM_ON_HOLD: 'job is queued and in user and system hold',
                        drmaa.JobState.RUNNING: 'job is running',
                        drmaa.JobState.SYSTEM_SUSPENDED: 'job is system suspended',
                        drmaa.JobState.USER_SUSPENDED: 'job is user suspended',
                        drmaa.JobState.DONE: 'job finished normally',
                        drmaa.JobState.FAILED: 'job finished, but failed'}

       s.synchronize([jobid], timeout=drmaa.Session.TIMEOUT_WAIT_FOREVER, dispose=False)

       retval = s.wait(jobid, drmaa.Session.TIMEOUT_WAIT_FOREVER)
       print('Job: {0} finished with status {1}'.format(retval.jobId, retval.hasExited))
       print("exitStatus {0}\nhasCoreDump {1}\nhasExited {2}\nhasSignal {3}\njobId {4}\nresourceUsage {5}\nterminatedSignal {6}wasAborted {7}\n".format(retval.exitStatus, retval.hasCoreDump, retval.hasExited, retval.hasSignal, retval.jobId, retval.resourceUsage, retval.terminatedSignal, retval.wasAborted))

       print('Cleaning up')
if __name__=='__main__':
Please keep all replies on the list by using "reply all"
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

To search Galaxy mailing lists use the unified search at:

Reply via email to