Thanks Nate for your prompt response.
Here is what I tried.
Got the jobid (e.g. 02b1cfb59f716d61) from the file download link (could not
find the "display URL" as you described).
Then tried to use the helper script but got the following error:
python ./helper.py -d 02b1cfb59f716d61
Traceback (most recent call last):
File "/path_to/galaxy-prod/galaxy-dist/scripts/helper.py", line 40, in ?
from galaxy.web import security
File "lib/galaxy/web/__init__.py", line 5, in ?
from framework import expose, json, json_pretty, require_login,
require_admin, url_for, error, form, FormBuilder, expose_api
File "lib/galaxy/web/framework/__init__.py", line 8, in ?
pkg_resources.require( "Cheetah" )
File "lib/galaxy/eggs/__init__.py", line 415, in require
raise EggNotFetchable( str( [ egg.name for egg in e.eggs ] ) )
From: Nate Coraor [mailto:n...@bx.psu.edu]
Sent: Friday, May 11, 2012 3:54 PM
To: Kelkar, Hemant
Subject: Re: [galaxy-dev] Problem related to a job that "failed"
On May 11, 2012, at 3:39 PM, Kelkar, Hemant wrote:
> This question pertains to a local galaxy install where the jobs are being
> submitted to a cluster running LSF.
> I periodically get an error from a galaxy job ("Job output not returned from
> cluster") even though the job completes properly on the cluster. In
> researching this issue our systems administrator discovered that the error
> seems to be happening because of NFS caching. The problem is that the job
> finishes on the job node, but the Galaxy server doesn't see that reflected in
> the output file because of delays in the cache update in NFS. The only
> solution he discovered was turning the caching off on the file system. In our
> case it is not possible because it will lead to a performance hit that would
> be not acceptable on the shared cluster. Most other file systems on the
> cluster are "NFS" so moving the database/pbs folder to another file system is
> not an option.
You wouldn't need to change attribute caching on the entire cluster, just on
the filesystem in which Galaxy is stored on the server on which the Galaxy
application runs. If that's not possible either, see the
'retry_job_output_collection' option in universe_wsgi.ini.
> I generally work around the NFS cache issue but now it is leading to a second
> problem. Since the job appears to be in some failed state to galaxy (it shows
> up as red in history) I can't seem to use the output file (even though it is
> there and I can see it using the "eye" icon) to move to the next step. The
> file attribute is set right.
> I assume a possible solution may be to reset the "failed" flag on the history
> item. Would this need to be done in the database? Downloading and then
> re-uploading the result file (a 25+ GB SAM file in this case) may be a
> workaround but it is not very practical.
Yes, this would be something to update in the database. The process is
1a. Collect the encoded history_dataset_association ID from the errored
dataset's display URL.
1b. Use galaxy-dist/scripts/helper.py -d <encoded_hda_id> to decode the id.
1c. In the database, 'select dataset_id from history_dataset_association where
1. Collect the output dataset id from the command line that was executed and
logged in the Galaxy server log
2. In the database, 'update dataset set state='ok' where id=<dataset_id>'
> Any ideas/suggestions?
> Please keep all replies on the list by using "reply all"
> in your mail client. To manage your subscriptions to this and other
> Galaxy lists, please use the interface at:
Please keep all replies on the list by using "reply all"
in your mail client. To manage your subscriptions to this
and other Galaxy lists, please use the interface at: