Re: [galaxy-dev] recovery: galaxy restart

2011-06-02 Thread Harendra chawla
Hi Nate,

I have used the local.py logic for submission of my jobs, for a grid based
architecture. But local.py does not have the recover() function so I was
trying to use the concept of recover() from other modules. My job submission
also uses the drmaa api's, so I thought of using the drmaa.py module.
Moreover the local.py doesn't have the monitor() or check_watched_items()
functions which means it doesn't use the concept of monitor_queue. So how
can I create a recover() function in such a case.
Can you suggest anything regarding this issue.

Thanks
Harendra

On Wed, Jun 1, 2011 at 10:52 PM, Nate Coraor n...@bx.psu.edu wrote:

 Harendra chawla wrote:
  Hi Nate,
 
  Ya I have seen the function  __check_jobs_at_startup(), it calls the
 recover
  function after checking the state of the job from the database and
 updating
  the jobWrapper. But what it does after calling the recover function and
 one
  more thing, does it use the monitor() and the check_watched_item() in the
  drmaa.py.

 Yes, notice that the last function of the recover() method is to insert
 the job state object into the DRM monitor queue with:

  self.monitor_queue.put( drm_job_state )

 Which is the same as the final step of the submission process, the
 queue_job() method.  Once this happens, the monitor thread will pick up
 the job state from monitor_queue (a Python Queue instance) and monitor
 it with monitor()/check_watched_items().

 --nate

 
  Thanks
  Harendra
 
  On Wed, Jun 1, 2011 at 9:10 PM, Nate Coraor n...@bx.psu.edu wrote:
 
   Harendra chawla wrote:
Hi Nate,
   
I got your point but which part of the code is doing all these
 things, I
mean how exactly this is done.
Is it using any other function apart from recover?
  
   Yes, see __check_jobs_at_startup() in lib/galaxy/jobs/__init__.py
  
   --nate
  
   
Regards
Harendra
   
On Wed, Jun 1, 2011 at 8:56 AM, Nate Coraor n...@bx.psu.edu wrote:
   
 Harendra chawla wrote:
  Hi everyone,
 
  I am trying to modify the *recover* function from the drmaa.py
  (/galaxy_central/lib/galaxy/job/runners/drmaa.py) as per my
   requirements.
  But I am not ale to understand the flow of that function.
 
  The recover function is called when the galaxy server is
 restarted.
   It
 first
  looks for the running jobs from the database. Then my problem is
 how
   it
  regains the same old state of the galaxy (specially the GUI)
 which
   was
  before the galaxy got restarted.
  Can anyone explain me the flow of the recover function and how
 the
   old
 state
  is regained.

 Hi Harendra,

 I'm not sure I understand what you mean by old state and the GUI -
 all
 that's really necessary here is to determine what Galaxy considers
 to
   be
 the state of the job (new, queued, running), recreate the in-memory
 job
 components (the JobWrapper), and place the job back in Galaxy's
 DRM-monitoring queue, which will then proceed with the process of
 finishing the job if it's finished in the DRM or waiting for it to
 finish if it's still queued or running in the DRM.

 --nate

 
  Regards
  Harendra

  ___
  Please keep all replies on the list by using reply all
  in your mail client.  To manage your subscriptions to this
  and other Galaxy lists, please use the interface at:
 
http://lists.bx.psu.edu/


  

___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/

Re: [galaxy-dev] Fwd: Request for pulling the updated FASTQ (de-)interlacer tool

2011-06-02 Thread Florent Angly

Hi Kanwei,
I think we'll get more feedback once it when it is added to Galaxy. I 
have run small tests (included in the functional tests) and real large 
datasets and have had no issues with the scripts and wrappers.

Regards,
Florent

On 27/05/11 18:49, Kanwei Li wrote:

Has anyone on the list tried this patch? Feedback would be appreciated.

Thanks,

K

On Fri, May 27, 2011 at 1:27 AM, Florent Angly 
florent.an...@gmail.com mailto:florent.an...@gmail.com wrote:


Can anyone comment on my email please?
Florent



On 22/05/11 14:03, Florent Angly wrote:

Let me know if you have any questions about this.
Thanks,
Florent

 Original Message 
Subject:Request for pulling the updated FASTQ (de-)interlacer tool
Date:   Tue, 17 May 2011 17:47:13 +1000
From:   Florent Angly florent.an...@gmail.com
mailto:florent.an...@gmail.com
To: galaxy-dev@lists.bx.psu.edu
mailto:galaxy-dev@lists.bx.psu.edu
galaxy-dev@lists.bx.psu.edu mailto:galaxy-dev@lists.bx.psu.edu



Dear Galaxy team,
Please consider pulling these changes I made to the
interlacer/de-interlacer tools into the main Galaxy repository:
https://bitbucket.org/fangly/galaxy-central/changeset/7f7ec8f05d36
While the tools were handling mate pairs fine, I now added the ability
to also keep single reads (reads that miss their mate) in a separate
file. The functional tests were updated and pass.
Thanks,
Florent




___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

http://lists.bx.psu.edu/




___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/

Re: [galaxy-dev] Blast2GO in Galaxy

2011-06-02 Thread Peter Cock
On Tue, May 31, 2011 at 11:26 AM, Peter Cock p.j.a.c...@googlemail.com wrote:
 Hi all,

 As I mentioned on Twitter, at the end of last week I wrapped Blast2GO
 for Galaxy, using the b2g4pipe program (Blast2GO for pipelines). See
 http://blast2go.org/

 Currently current code is on bitbucket under my tools branch,
 https://bitbucket.org/peterjc/galaxy-central/src/tools/

 Specifically files tools/ncbi_blast_plus/blast2go.* viewable here:
 https://bitbucket.org/peterjc/galaxy-central/src/tools/tools/ncbi_blast_plus/

 I've using a Galaxy location file, tool-data/blast2go.loc, to offer
 one or more Blast2GO configurations (properties files), mapping this
 to the -prop argument. This way you could have for example the Spanish
 Blast2GO server with its current database (May 2010), and a local
 Blast2GO database. I want to setup a local database and try this
 before submitting the wrapper to the Tool Shed.

I've done that using the current latest data from May 2011, so one
year newer than the current default public Blast2GO database provided
by the Blast2GO developers (dated May 2011). This requires downloaded
some large data files and importing them into a MySQL database, and
that took about 24 hours to process in all. The last step is done via
their Java tool and took most of the time. For more details, see:
http://blast2go.org/localgodb

However, the end result is worth the effort as running Blast2GO is now
much much faster. I need to try it on some larger files, but we're
talking at least an order of magnitude quicker, maybe two.

 The input to the tool is a BLAST XML file, specifically blasting
 against a protein database like NR (so blastp or blastx, not blastn
 etc). I want to try some very large BLAST XML files to confirm
 b2g4pipe copes with the current BLAST+ output files - I gather there
 were some problems in this area in the past, so having the wrapper
 script fragment the XML might be a workaround. Currently the only real
 function of the wrapper script is to rename the output file - b2g4pipe
 insists on using the *.annot extension.

Now that I have a local Blast2GO database, I'll be able to try out
b2g4pipe on some bigger XML files (without having to wait ages).

 Right now the only output is a tabular three column *.annot file,
 which can be loaded into the Blast2GO GUI. For analysis within Galaxy,
 I'm wondering about an option to split the first column (which holds
 the original FASTA query's identifier and any description) in two.
 i.e. Split at the first white space to give the FASTA identifier, and
 any optional description as a separate column. That would make
 linking/joining/filtering on the ID much easier.

 If anyone has any comments or feedback now, that would be welcome.

I've submitted the initial version to the Galaxy Tool Shed now, but
would still welcome feedback and would even consider non-backwards
compatible changes in the short term.

Peter
___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


[galaxy-dev] Updates to Galaxy Community Conference website

2011-06-02 Thread Peter Cock
Hi all,

I have a couple of minor suggestions for improving the website
http://galaxy.psu.edu/gcc2011/Home.html

First, please fill in a meaningful HTML titles since that will be used
as the default caption in a bookmark, tab or window title.

Second, please add a link to the presentation slides and videos,
e.g. from the programme page and main page?

http://wiki.g2.bx.psu.edu/GCC2011

I realise this isn't complete yet, but it is already a useful resource.
Also once you have all the slides, could the links in the wiki table
be made to point straight to the file?

Thanks,

Peter
___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


Re: [galaxy-dev] sqlite to postgres

2011-06-02 Thread Hans-Rudolf Hotz

Hi Harendra

I am not aware of anybody who made this kind of transition without 
running into all kind of problems. I suggest the following workaround:


- set up a new, independent Galaxy server using PostgreSQL.
- transfer the histories you want to keep to the new server using the
  'Export to File' option on the old server followed by the 'Import to
  File' option on the new server.


Regards, Hans


On 06/02/2011 12:01 PM, Harendra chawla wrote:

Hi,

I want to convert my database from sqlite to postgres. I am able to do it as
far as schema is concerned, but I want to preserve the data as well. When I
changed the database to postgres, all my history got lost. Is there any way
I can preserve the data in sqlite database while converting to postgres.

Regards
Harendra




___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

   http://lists.bx.psu.edu/

___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

 http://lists.bx.psu.edu/


Re: [galaxy-dev] problems previewing certain files, downloading files and with login

2011-06-02 Thread Nate Coraor
Matthew Conte wrote:
 Hi Nate,
 
 I was able to finally track down the login issue.  It had to do with the
 following setting in my universe_wsgi.ini:
 *
 *
 *cookie_path = /galaxy*
 
 Removing this out fixed the problem and I should be fine leaving it out out
 since I don't need to run more than one instance of Galaxy.  I probably
 shouldn't have had this setting in the first place, though I'm not exactly
 sure what this caused the problem.
 
 Anyways, thanks for all the help tracking these issues down.

Hi Matt,

I was about to get back to this now that the conference is over and I
was also wondering if that option was related - weird though, I'm not
sure why it should have an effect, but I'm glad you've figured out
what's up.  I guess there's a bug with the cookie path I'll need to
check out.

--nate

 
 -Matt
 
 On Mon, May 16, 2011 at 5:26 PM, Matthew Conte mco...@umd.edu wrote:
 
  Yep.
  On May 16, 2011 7:14 AM, Nate Coraor n...@bx.psu.edu wrote:
   Matthew Conte wrote:
   On Fri, May 13, 2011 at 11:18 AM, Nate Coraor n...@bx.psu.edu wrote:
  
That's correct, although it should've been fixed. Can you remove the
contents of galaxy-dist/database/compiled_templates, clear cache,
reload, and see if you still get the same results?
   
Thanks,
--nate
   
  
   I've removed the contents of that folder, cleared the browser cache,
   reloaded and still get the same results (nothing different in the log
   either).
  
   I'm not sure if this will help, but maybe I should mention that when I
  try
   to log in I do get the following screen to flash for a brief second
  before
   automatically taking me back to the welcome screen:
  
   [image: Screen shot 2011-05-13 at 1.44.25 PM.png]
  
   I'm going to try to set up a test environment to replicate this shortly.
   Does this happen with both Apache and nginx?
  
   --nate
 
___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


Re: [galaxy-dev] recovery: galaxy restart

2011-06-02 Thread Nate Coraor
Harendra chawla wrote:
 Hi Nate,
 
 I have used the local.py logic for submission of my jobs, for a grid based
 architecture. But local.py does not have the recover() function so I was
 trying to use the concept of recover() from other modules. My job submission
 also uses the drmaa api's, so I thought of using the drmaa.py module.
 Moreover the local.py doesn't have the monitor() or check_watched_items()
 functions which means it doesn't use the concept of monitor_queue. So how
 can I create a recover() function in such a case.
 Can you suggest anything regarding this issue.

Hi Harendra,

If you're submitting jobs to a cluster, I'd suggest using drmaa.py.  You
could make a copy of it and make any site changes necessary and load
that as a new job runner module.  Hacking cluster support on to the back
of local.py sort of defeats the purpose of Galaxy's cluster integration
and prevents stuff like exactly what it is you're trying to do.

Thanks,
--nate

 
 Thanks
 Harendra
 
 On Wed, Jun 1, 2011 at 10:52 PM, Nate Coraor n...@bx.psu.edu wrote:
 
  Harendra chawla wrote:
   Hi Nate,
  
   Ya I have seen the function  __check_jobs_at_startup(), it calls the
  recover
   function after checking the state of the job from the database and
  updating
   the jobWrapper. But what it does after calling the recover function and
  one
   more thing, does it use the monitor() and the check_watched_item() in the
   drmaa.py.
 
  Yes, notice that the last function of the recover() method is to insert
  the job state object into the DRM monitor queue with:
 
   self.monitor_queue.put( drm_job_state )
 
  Which is the same as the final step of the submission process, the
  queue_job() method.  Once this happens, the monitor thread will pick up
  the job state from monitor_queue (a Python Queue instance) and monitor
  it with monitor()/check_watched_items().
 
  --nate
 
  
   Thanks
   Harendra
  
   On Wed, Jun 1, 2011 at 9:10 PM, Nate Coraor n...@bx.psu.edu wrote:
  
Harendra chawla wrote:
 Hi Nate,

 I got your point but which part of the code is doing all these
  things, I
 mean how exactly this is done.
 Is it using any other function apart from recover?
   
Yes, see __check_jobs_at_startup() in lib/galaxy/jobs/__init__.py
   
--nate
   

 Regards
 Harendra

 On Wed, Jun 1, 2011 at 8:56 AM, Nate Coraor n...@bx.psu.edu wrote:

  Harendra chawla wrote:
   Hi everyone,
  
   I am trying to modify the *recover* function from the drmaa.py
   (/galaxy_central/lib/galaxy/job/runners/drmaa.py) as per my
requirements.
   But I am not ale to understand the flow of that function.
  
   The recover function is called when the galaxy server is
  restarted.
It
  first
   looks for the running jobs from the database. Then my problem is
  how
it
   regains the same old state of the galaxy (specially the GUI)
  which
was
   before the galaxy got restarted.
   Can anyone explain me the flow of the recover function and how
  the
old
  state
   is regained.
 
  Hi Harendra,
 
  I'm not sure I understand what you mean by old state and the GUI -
  all
  that's really necessary here is to determine what Galaxy considers
  to
be
  the state of the job (new, queued, running), recreate the in-memory
  job
  components (the JobWrapper), and place the job back in Galaxy's
  DRM-monitoring queue, which will then proceed with the process of
  finishing the job if it's finished in the DRM or waiting for it to
  finish if it's still queued or running in the DRM.
 
  --nate
 
  
   Regards
   Harendra
 
   ___
   Please keep all replies on the list by using reply all
   in your mail client.  To manage your subscriptions to this
   and other Galaxy lists, please use the interface at:
  
 http://lists.bx.psu.edu/
 
 
   
 
___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


Re: [galaxy-dev] Sharing tool definitions between XML files

2011-06-02 Thread Peter Cock
On Wed, Oct 6, 2010 at 11:16 AM, Peter pe...@maubp.freeserve.co.uk wrote:
 On Tue, Sep 28, 2010 at 12:04 PM, Peter pe...@maubp.freeserve.co.uk wrote:
 Hi all,

 I'm wondering if there is any established way to share parameter
 definitions between tool wrapper XML files?

 For example, I am currently working on NCBI BLAST+ wrappers, and these
 tools have a lot in common. For example, the output format option will
 be a select parameter, and it will be the same for blastn, blastp,
 tblastn, etc. I can just cut and paste the definition but this seems
 inelegant and will be a long term maintenance burden if it ever needs
 updating (lots of places would need to be updated the same way). I'd
 like to have a shared XML snippet defining this parameter which I
 could then reference/include from each tool wrapper than needs it.

 I'm thinking of something like XML's !ENTITY ... might work. Is this
 possible?

 Peter


 Alex Bossers replied on the BLAST+ thread to express interest in
 sharing definitions between wrapper XML files:

 http://lists.bx.psu.edu/pipermail/galaxy-dev/2010-September/003376.html

 As I add more parameters to my BLAST+ wrappers, keeping all
 the files in sync is getting a bit tedious...

 Peter


Hi all,

I'm replying to this old thread after finding this old (fixed) issue
on the bug tracker,
https://bitbucket.org/galaxy/galaxy-central/issue/131/allow-xml-includes-in-all-xml-files

Does this offer a solution to sharing definitions in tool XML files?
Has anyone got an example using this?

Peter
___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


Re: [galaxy-dev] Error when adding datasets

2011-06-02 Thread Nate Coraor
Louise-Amélie Schmitt wrote:
 Hello everyone
 
 I have an issue when trying to import new datasets or when putting a
 dataset into a history. I saw Edward Kirton had the same problem but he
 got no answer:
 http://lists.bx.psu.edu/pipermail/galaxy-dev/2010-May/002732.html
 
 Here is the error message I get when clicking the Add datasets button
 in a library, in the admin's Manage data libraries panel:
 
 
 UnmappedInstanceError: Class '__builtin__.NoneType' is not mapped
 
 URL:
 http://manni/galaxy/library_common/upload_library_dataset?library_id=f2db41e1fa331b3eshow_deleted=Falsecntrller=library_adminfolder_id=f2db41e1fa331b3euse_panels=False
 File
 '/g/funcgen/galaxy/eggs/WebError-0.8a-py2.6.egg/weberror/evalexception/middleware.py',
  line 364 in respond
   app_iter = self.application(environ, detect_start_response)
 File '/g/funcgen/galaxy/eggs/Paste-1.6-py2.6.egg/paste/debug/prints.py',
 line 98 in __call__
   environ, self.app)
 File '/g/funcgen/galaxy/eggs/Paste-1.6-py2.6.egg/paste/wsgilib.py', line
 539 in intercept_output
   app_iter = application(environ, replacement_start_response)
 File '/g/funcgen/galaxy/eggs/Paste-1.6-py2.6.egg/paste/recursive.py',
 line 80 in __call__
   return self.application(environ, start_response)
 File
 '/g/funcgen/galaxy/lib/galaxy/web/framework/middleware/remoteuser.py',
 line 109 in __call__
   return self.app( environ, start_response )
 File
 '/g/funcgen/galaxy/eggs/Paste-1.6-py2.6.egg/paste/httpexceptions.py',
 line 632 in __call__
   return self.application(environ, start_response)
 File '/g/funcgen/galaxy/lib/galaxy/web/framework/base.py', line 145 in
 __call__
   body = method( trans, **kwargs )
 File '/g/funcgen/galaxy/lib/galaxy/web/controllers/library_common.py',
 line 907 in upload_library_dataset
   trans.sa_session.refresh( history )
 File
 '/g/funcgen/galaxy/eggs/SQLAlchemy-0.5.6_dev_r6498-py2.6.egg/sqlalchemy/orm/scoping.py',
  line 127 in do
   return getattr(self.registry(), name)(*args, **kwargs)
 File
 '/g/funcgen/galaxy/eggs/SQLAlchemy-0.5.6_dev_r6498-py2.6.egg/sqlalchemy/orm/session.py',
  line 925 in refresh
   raise exc.UnmappedInstanceError(instance)
 UnmappedInstanceError: Class '__builtin__.NoneType' is not mapped
 
 
 Now when does it occur:
 
 I have two databases. One test database I created a month ago which
 works fine, even now. The other one I created recently which is supposed
 to be the final database. But the latter keeps triggering the above
 message, even when I drop it and create it all over again. I even tried
 to create a third one, all clean and new but the problem remains. I
 tried to trash all the eggs so Galaxy gets fresh new eggs, with no
 effect at all. The error's still there.
 
 If you have any clue, I'll be forever grateful.

Okay, the problem is that the exact process you're taking after
recreating the database is resulting in your user not having a history.
This is pretty rare, but does occasionally happen.  I just committed a
fix for it in 5616:1f2870dc3ab0 and 5617:83103cff8757, but you can work
around it without these changes pretty easily by hitting the Analysis
view before uploading to make sure you have a history.

--nate

 
 Cheers,
 L-A
 
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
 
   http://lists.bx.psu.edu/
___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


Re: [galaxy-dev] Error: unable to read Galaxy config

2011-06-02 Thread Nate Coraor
Lewis, Brian Andrew wrote:
 After following the instructions in the wiki for setting up scaling/load 
 balancing, I was testing out submitting a job and got this error when trying 
 to pull data from USCS Main:
 
 Traceback (most recent call last):
   File /usr/local/galaxy-dist/tools/data_source/data_source.py, line 6, in
 from galaxy.util import gzip_magic
   File /usr/local/galaxy-dist/lib/galaxy/util/__init__.py, line 20, in
 pkg_resources.require( 'docutils' )
   File /usr/local/galaxy-dist/lib/galaxy/eggs/__init__.py, line 400, in 
 require
 c = Crate()
   File /usr/local/galaxy-dist/lib/galaxy/eggs/__init__.py, line 259, in 
 __init__
 self.galaxy_config = GalaxyConfig()
   File /usr/local/galaxy-dist/lib/galaxy/eggs/__init__.py, line 358, in 
 __init__
 raise Exception( error: unable to read Galaxy config from %s % 
 GalaxyConfig.config_file )
 Exception: error: unable to read Galaxy config from /usr/local/galaxy-dist/0
 
 I'm running Galaxy on a RHEL6 server and I have one web server and one runner 
 set up using Apache as a proxy server.

Hi Brian,

Due to some hardcoding that shouldn't exist, you need to leave a copy of
universe_wsgi.ini behind in addition to the two new configs.  I'll work
on a fix for this ASAP.

--nate

 
 Thanks,
 Brian
 

 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
 
   http://lists.bx.psu.edu/

___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


Re: [galaxy-dev] Support for subdirs in dataset extra_files_path

2011-06-02 Thread Daniel Blankenberg
Patch Imported in 5618:b9fdb88da530.

Thanks!

Dan


On Mar 16, 2011, at 11:50 AM, Jim Johnson wrote:

 Request is issue#494   
 https://bitbucket.org/galaxy/galaxy-central/issue/494/support-sub-dirs-in-extra_files_path-patch
 
 I'm finding that some qiime metagenomics applications build HTML results with 
 an inherent directory structure.   For some other applications, e.g. FastQC, 
 I've been able to flatten the hierarchy and edit the html, but that appears 
 problematic for qiime.
 
 Galaxy hasn't supported a dataset extra_files_path hierarchy, though the 
 developers don't seem opposed to the idea: 
 http://lists.bx.psu.edu/pipermail/galaxy-dev/2010-October/003605.html
 
 I added a route  in  lib/galaxy/web/buildapp.pyand modified the dataset 
 download code to traverse a hierarchy in lib/galaxy/web/controllers/dataset.py
 I don't think these add any security vulnerabilities, (I tried the obvious 
 ../../ ).
 
 $ hg diff lib/galaxy/web/buildapp.py
 diff -r 6ae06d89fec7 lib/galaxy/web/buildapp.py
 --- a/lib/galaxy/web/buildapp.pyWed Mar 16 09:01:57 2011 -0400
 +++ b/lib/galaxy/web/buildapp.pyWed Mar 16 10:24:13 2011 -0500
 @@ -94,6 +94,8 @@
 webapp.add_route( '/async/:tool_id/:data_id/:data_secret', 
 controller='async', action='index', tool_id=None, data_id=None, 
 data_secret=None )
 webapp.add_route( '/:controller/:action', action='index' )
 webapp.add_route( '/:action', controller='root', action='index' )
 +# allow for subdirectories in extra_files_path
 +webapp.add_route( '/datasets/:dataset_id/display/{filename:.+?}', 
 controller='dataset', action='display', dataset_id=None, filename=None)
 webapp.add_route( '/datasets/:dataset_id/:action/:filename', 
 controller='dataset', action='index', dataset_id=None, filename=None)
 webapp.add_route( 
 '/display_application/:dataset_id/:app_name/:link_name/:user_id/:app_action/:action_param',
  controller='dataset', action='display_application', dataset_id=None, 
 user_id=None, app_name = None, link_name = None, app_action = None, 
 action_param = None )
 webapp.add_route( '/u/:username/d/:slug', controller='dataset', 
 action='display_by_username_and_slug' )
 $
 $ hg diff lib/galaxy/web/controllers/dataset.py
 diff -r 6ae06d89fec7 lib/galaxy/web/controllers/dataset.py
 --- a/lib/galaxy/web/controllers/dataset.py Wed Mar 16 09:01:57 2011 -0400
 +++ b/lib/galaxy/web/controllers/dataset.py Wed Mar 16 10:24:29 2011 -0500
 @@ -266,17 +266,18 @@
 log.exception( Unable to add composite parent %s to 
 temporary library download archive % data.file_name)
 msg = Unable to create archive for download, please 
 report this error
 messagetype = 'error'
 -flist = glob.glob(os.path.join(efp,'*.*')) # glob returns 
 full paths
 -for fpath in flist:
 -efp,fname = os.path.split(fpath)
 -try:
 -archive.add( fpath,fname )
 -except IOError:
 -error = True
 -log.exception( Unable to add %s to temporary 
 library download archive % fname)
 -msg = Unable to create archive for download, please 
 report this error
 -messagetype = 'error'
 -continue
 +for root, dirs, files in os.walk(efp):
 +for fname in files:
 +fpath = os.path.join(root,fname)
 +rpath = os.path.relpath(fpath,efp)
 +try:
 +archive.add( fpath,rpath )
 +except IOError:
 +error = True
 +log.exception( Unable to add %s to temporary 
 library download archive % rpath)
 +msg = Unable to create archive for download, 
 please report this error
 +messagetype = 'error'
 +continue
 if not error:
 if params.do_action == 'zip':
 archive.close()
 
 
 
 
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
 
 http://lists.bx.psu.edu/


___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


Re: [galaxy-dev] Migration error: fields in MySQL

2011-06-02 Thread Nate Coraor
John Eppley wrote:
 
 I had an error upgrading my galaxy instance. I got the following exception 
 while migrating the db (during step 64-65):
 
 sqlalchemy.exc.ProgrammingError: (ProgrammingError) (1064, You have an error 
 in your SQL syntax; check the manual that corresponds to your MySQL server 
 version for the right syntax to use near 'fields FROM form_definition' at 
 line 1) u'SELECT id, fields FROM form_definition' []
 
 It seems my version (4.1.22-log) of MySQL did not like 'fields' as a column 
 name. If I alias the formdefinition as f and us f.fields, the error goes 
 away. I also had to modify migration 76 for the same reason.

Hi John,

Thanks for the patch, I've committed it as 5619:b6689fb6532e.

--nate

 
 Here is my diff of the migrations dir:
 diff -r 50e249442c5a 
 lib/galaxy/model/migrate/versions/0065_add_name_to_form_fields_and_values.py
 --- 
 a/lib/galaxy/model/migrate/versions/0065_add_name_to_form_fields_and_values.py
Thu Apr 07 08:39:07 2011 -0400
 +++ 
 b/lib/galaxy/model/migrate/versions/0065_add_name_to_form_fields_and_values.py
Fri Apr 15 11:09:26 2011 -0400
 @@ -39,7 +39,7 @@
  return ''
  # Go through the entire table and add a 'name' attribute for each field
  # in the list of fields for each form definition
 -cmd = SELECT id, fields FROM form_definition
 +cmd = SELECT f.id, f.fields FROM form_definition f
  result = db_session.execute( cmd )
  for row in result:
  form_definition_id = row[0]
 @@ -53,7 +53,7 @@
  field[ 'helptext' ] = field[ 'helptext' ].replace(', 
 '').replace('', )
  field[ 'label' ] = field[ 'label' ].replace(', '')
  fields_json = to_json_string( fields_list )
 -cmd = UPDATE form_definition SET fields='%s' WHERE id=%i %( 
 fields_json, form_definition_id )
 +cmd = UPDATE form_definition f SET f.fields='%s' WHERE f.id=%i 
 %( fields_json, form_definition_id )
  db_session.execute( cmd )
  # replace the values list in the content field of the form_values table 
 with a name:value dict
  cmd = SELECT form_values.id, form_values.content, 
 form_definition.fields \
 @@ -112,7 +112,7 @@
  cmd = UPDATE form_values SET content='%s' WHERE id=%i %( 
 to_json_string( values_list ), form_values_id )
  db_session.execute( cmd )
  # remove name attribute from the field column of the form_definition 
 table
 -cmd = SELECT id, fields FROM form_definition
 +cmd = SELECT f.id, f.fields FROM form_definition f
  result = db_session.execute( cmd )
  for row in result:
  form_definition_id = row[0]
 @@ -124,5 +124,5 @@
  for index, field in enumerate( fields_list ):
  if field.has_key( 'name' ):
  del field[ 'name' ]
 -cmd = UPDATE form_definition SET fields='%s' WHERE id=%i %( 
 to_json_string( fields_list ), form_definition_id )
 +cmd = UPDATE form_definition f SET f.fields='%s' WHERE id=%i 
 %( to_json_string( fields_list ), form_definition_id )
  db_session.execute( cmd )
 diff -r 50e249442c5a 
 lib/galaxy/model/migrate/versions/0076_fix_form_values_data_corruption.py
 --- 
 a/lib/galaxy/model/migrate/versions/0076_fix_form_values_data_corruption.py   
Thu Apr 07 08:39:07 2011 -0400
 +++ 
 b/lib/galaxy/model/migrate/versions/0076_fix_form_values_data_corruption.py   
Fri Apr 15 11:09:26 2011 -0400
 @@ -32,7 +32,7 @@
  def upgrade():
  print __doc__
  metadata.reflect()
 -cmd = SELECT form_values.id as id, form_values.content as field_values, 
 form_definition.fields as fields  \
 +cmd = SELECT form_values.id as id, form_values.content as field_values, 
 form_definition.fields as fdfields  \
+  FROM form_definition, form_values  \
+  WHERE form_values.form_definition_id=form_definition.id  \
+  ORDER BY form_values.id
 @@ -46,7 +46,7 @@
  except Exception, e:
  corrupted_rows = corrupted_rows + 1
  # content field is corrupted
 -fields_list = from_json_string( _sniffnfix_pg9_hex( str( 
 row['fields'] ) ) )
 +fields_list = from_json_string( _sniffnfix_pg9_hex( str( 
 row['fdfields'] ) ) )
  field_values_str = _sniffnfix_pg9_hex( str( row['field_values'] 
 ) )
  try:
  #Encoding errors?  Just to be safe.
 
 -j
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
 
   http://lists.bx.psu.edu/
___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


Re: [galaxy-dev] Error when adding datasets

2011-06-02 Thread Louise-Amélie Schmitt
On Thu, 2 Jun 2011 11:48:55 -0400, Nate Coraor n...@bx.psu.edu wrote:
 Greg Von Kuster wrote:
 Hello Louise,
 
 I've CC'd Nate on this as he may be able to help - although no
 guarantees.  I'm not expert enough in this area to know where to look
for
 the cause.  Perhaps someone in the community can help as well.
 
 It is likely that LDAP is playing a role in this behavior.
 
 
 On Apr 14, 2011, at 1:00 PM, Louise-Amélie Schmitt wrote:
 
  The thing is, we use LDAP logging so we can't even access the website
  without logging in.
  Moreover, when I logged in, I arrived on the data analysis page where
  the
  automatic unnamed history was obviously created in the history
panel.
  
  I forgot to mention we have issues creating and deleting histories,
  like,
  we can't access some histories and everytime we delete histories, two
  extra
  unnamed histories are created. As I mentioned before, it is also
  impossible
  to load a dataset in any history, existing or not.
 
 Okay, somehow I missed all the followup on this question when I
 committed a fix a few minutes ago.  Could you provide some details about
 the inaccessible histories (what happens when you try to access them,
 traceback, etc?) and loading data - you mean the upload tool fails?  Is
 there also a traceback here?

ooh it's been a while I haven't seen the issue since I not-so-elegantly
avoided the problem making our two instances run on separate machines.

But I'll try to answer your question as accurately as I can.

I experienced a similar error (pretty much the same traceback as the one I
pasted in a previous message, just yelling at me saying that the history is
nonexistent) whenever I tried to:
- upload data in the manage libraries admin panel and in the analyse data
panel (upload tool)
- load an already uploaded dataset in any kind of history (saved or new)
- probably do something else, I can't really remember

I also had issues dealing with histories but not getting any traceback,
like I mentioned a little lower: Everytime I tried to delete any number of
histories, the histories were deleted but two were created.

I hope this will help, sorry I can't really be more precise.
L-A


 
 Thanks,
 --nate
 
  
  Do you think it could be linked to our using LDAP?
  
  Thanks
  L-A
  
  
  On Thu, 14 Apr 2011 12:06:55 -0400, Greg Von Kuster g...@bx.psu.edu
  wrote:
  On Apr 14, 2011, at 11:49 AM, Louise-Amélie Schmitt wrote:
  
  Here is the result I got from the debug statements:
  
  galaxy.web.controllers.library_common DEBUG 2011-04-14 17:46:02,286
  ###
  history: None
  
  This is the problem - when you registered normally, a history would
  have
  been automatically created for you.  But somehow you don't have a
  history -
  no idea why / how this happened, but it is very unlikely this is a
  result
  of a bug within Galaxy.  Try logging out and logging in again and a
  new
  history should be created for you.
  
  
  galaxy.web.controllers.library_common DEBUG 2011-04-14 17:46:02,286
  ###
  trans: galaxy.web.framework.GalaxyWebUITransaction object at
  0x29f40950
  galaxy.web.controllers.library_common DEBUG 2011-04-14 17:46:02,286
  ###
  trans.sa_session: sqlalchemy.orm.scoping.ScopedSession object at
  0x19ab21d0
  
  Thanks again
  L-A
  
  
  Le jeudi 14 avril 2011 à 11:05 -0400, Greg Von Kuster a écrit :
  I assume when you dropped the old database and recreated the new,
  empty database, you made sure the database connection string in
  universe_wsgi.ini was correctly set.  if so, when you started up
  Galaxy, it would have created all of the required tables in the
new
  database, and they would all be empty.  When you first registered
as
  the admin user, it would have automatically populated several
tables
  with data, including the history table.  One or more of these
things
  must not have been successful.
  
  
  To attempt to narrow down the problem, I'll need you to do the
  following.  Here are lines # 905 - 907 in
  ~/lib/galaxy/web/controllers/library-common.py
  
  
# Send the current history to the form to enable importing
  datasets from history to library
history = trans.get_history()
trans.sa_session.refresh( history )
  
  
  Please add the following debug statements:
  
  
# Send the current history to the form to enable importing
  datasets from history to library
history = trans.get_history()
log.debug(### history: %s % str( history ))
log.debug(### trans: %s % str( trans ))
log.debug(### trans.sa_session: %s %
  str( trans.sa_session ))
trans.sa_session.refresh( history )
  
  
  Stop and restart your Galaxy server after making the above
changes,
  and reply back with the output of the debug statements.  Assuming
  you
  start your Galaxy instance with:
  
  
  sh run.sh
  
  
  you'll see the results of the debug statements in the log
scrolling
  in
  the window in which you started Galaxy.
  
  
  Thanks
  
  
  On Apr 14, 2011, at 10:46 

[galaxy-dev] Defining new file formats in Galaxy (for new tool wrappers)

2011-06-02 Thread Peter Cock
Hi all,

Something I've not needed to do until now is define a new file format
in Galaxy. I understand the basic principle and defining a subclass in
Python... however, how does this work with new tools on the Tool Shed?
In particular, if an output format is likely to be used by more than
one tool, can we get it added to the Galaxy core?

As an example, the basic functionality of the Blast2GO for pipelines
tool (b2g4pipe) takes a BLAST XML input file, and gives a tab
separated annotation output file. Galaxy already has 'blastxml' and
'tabular' file formats defined, so I didn't need to do anything extra.
However, the tool can also take (a directory of) InterProScan XML
files as input, so here a new 'interproscanxml' format would useful.
Then any wrapper using or producing InterProScan XML could take
advantage of this. e.g. Konrad's InterProScan wrapper could then offer
the XML output as an option in addition to or instead of the tabular
output.

Related to this example, why isn't there a generic base class for XML
formats in general?
https://bitbucket.org/galaxy/galaxy-central/issue/568/missing-xml-datatype-base-class

Regards,

Peter
___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


Re: [galaxy-dev] Error: unable to read Galaxy config

2011-06-02 Thread Nate Coraor
Lewis, Brian Andrew wrote:
 Nate -
 
 I checked and I still have a copy of universe_wsgi.ini within my galaxy-dist 
 directory.  I did also make a change in 
 /usr/local/galaxy-dist/lib/galaxy/eggs/__init__.py according to the post 
 here: http://lists.bx.psu.edu/pipermail/galaxy-dev/2010-May/002637.html  

Ah, this would probably be problematic - when
tools/data_source/data_source.py runs, its argv[2] is not going to be
the universe_wsgi.ini file.  What happens if you change it back to the
original?

--nate

 
 ~ Brian
 
 
 -Original Message-
 From: Nate Coraor [mailto:n...@bx.psu.edu] 
 Sent: Thursday, June 02, 2011 11:51 AM
 To: Lewis, Brian Andrew
 Cc: galaxy-...@bx.psu.edu
 Subject: Re: [galaxy-dev] Error: unable to read Galaxy config
 
 Lewis, Brian Andrew wrote:
  After following the instructions in the wiki for setting up scaling/load 
  balancing, I was testing out submitting a job and got this error when 
  trying to pull data from USCS Main:
  
  Traceback (most recent call last):
File /usr/local/galaxy-dist/tools/data_source/data_source.py, line 6, in
  from galaxy.util import gzip_magic
File /usr/local/galaxy-dist/lib/galaxy/util/__init__.py, line 20, in
  pkg_resources.require( 'docutils' )
File /usr/local/galaxy-dist/lib/galaxy/eggs/__init__.py, line 400, in 
  require
  c = Crate()
File /usr/local/galaxy-dist/lib/galaxy/eggs/__init__.py, line 259, in 
  __init__
  self.galaxy_config = GalaxyConfig()
File /usr/local/galaxy-dist/lib/galaxy/eggs/__init__.py, line 358, in 
  __init__
  raise Exception( error: unable to read Galaxy config from %s % 
  GalaxyConfig.config_file )
  Exception: error: unable to read Galaxy config from 
  /usr/local/galaxy-dist/0
  
  I'm running Galaxy on a RHEL6 server and I have one web server and one 
  runner set up using Apache as a proxy server.
 
 Hi Brian,
 
 Due to some hardcoding that shouldn't exist, you need to leave a copy of 
 universe_wsgi.ini behind in addition to the two new configs.  I'll work on a 
 fix for this ASAP.
 
 --nate
 
  
  Thanks,
  Brian
  
 
  ___
  Please keep all replies on the list by using reply all
  in your mail client.  To manage your subscriptions to this and other 
  Galaxy lists, please use the interface at:
  
http://lists.bx.psu.edu/
 
 
___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


Re: [galaxy-dev] Error: unable to read Galaxy config

2011-06-02 Thread Lewis, Brian Andrew
Okay, yeah, when I changed it back to the original code it worked.  

~ Brian


-Original Message-
From: Nate Coraor [mailto:n...@bx.psu.edu] 
Sent: Thursday, June 02, 2011 12:47 PM
To: Lewis, Brian Andrew
Cc: galaxy-...@bx.psu.edu
Subject: Re: [galaxy-dev] Error: unable to read Galaxy config

Lewis, Brian Andrew wrote:
 Nate -
 
 I checked and I still have a copy of universe_wsgi.ini within my 
 galaxy-dist directory.  I did also make a change in 
 /usr/local/galaxy-dist/lib/galaxy/eggs/__init__.py according to the 
 post here: 
 http://lists.bx.psu.edu/pipermail/galaxy-dev/2010-May/002637.html

Ah, this would probably be problematic - when tools/data_source/data_source.py 
runs, its argv[2] is not going to be the universe_wsgi.ini file.  What happens 
if you change it back to the original?

--nate

 
 ~ Brian
 
 
 -Original Message-
 From: Nate Coraor [mailto:n...@bx.psu.edu]
 Sent: Thursday, June 02, 2011 11:51 AM
 To: Lewis, Brian Andrew
 Cc: galaxy-...@bx.psu.edu
 Subject: Re: [galaxy-dev] Error: unable to read Galaxy config
 
 Lewis, Brian Andrew wrote:
  After following the instructions in the wiki for setting up scaling/load 
  balancing, I was testing out submitting a job and got this error when 
  trying to pull data from USCS Main:
  
  Traceback (most recent call last):
File /usr/local/galaxy-dist/tools/data_source/data_source.py, line 6, in
  from galaxy.util import gzip_magic
File /usr/local/galaxy-dist/lib/galaxy/util/__init__.py, line 20, in
  pkg_resources.require( 'docutils' )
File /usr/local/galaxy-dist/lib/galaxy/eggs/__init__.py, line 400, in 
  require
  c = Crate()
File /usr/local/galaxy-dist/lib/galaxy/eggs/__init__.py, line 259, in 
  __init__
  self.galaxy_config = GalaxyConfig()
File /usr/local/galaxy-dist/lib/galaxy/eggs/__init__.py, line 358, in 
  __init__
  raise Exception( error: unable to read Galaxy config from %s % 
  GalaxyConfig.config_file )
  Exception: error: unable to read Galaxy config from
  /usr/local/galaxy-dist/0
  
  I'm running Galaxy on a RHEL6 server and I have one web server and one 
  runner set up using Apache as a proxy server.
 
 Hi Brian,
 
 Due to some hardcoding that shouldn't exist, you need to leave a copy of 
 universe_wsgi.ini behind in addition to the two new configs.  I'll work on a 
 fix for this ASAP.
 
 --nate
 
  
  Thanks,
  Brian
  
 
  ___
  Please keep all replies on the list by using reply all
  in your mail client.  To manage your subscriptions to this and other 
  Galaxy lists, please use the interface at:
  
http://lists.bx.psu.edu/
 
 

___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


[galaxy-dev] Separate histories between localhost and 127.0.0.1

2011-06-02 Thread Lewis, Brian Andrew
For some odd reason, when I try to import data using the URL 
http://localhost:8081 or http://localhost:8081/galaxy none of my jobs will even 
show up.  However if I go to http://127.0.0.1 or http://127.0.0.1/galaxy the 
jobs run fine.  Here's a snip from my httpd.conf file:

RewriteEngine on
RewriteRule ^/static/style/(.*) 
/usr/local/galaxy-dist/static/june_2007_style/blue/$1 [L]
RewriteRule ^/static/scripts/(.*) 
/usr/local/galaxy-dist/static/scripts/packed/$1 [L]
RewriteRule ^/static/(.*) /usr/local/galaxy-dist/static/$1 [L]
RewriteRule ^/favicon.ico /usr/local/galaxy-dist/static/favicon.ico [L]
RewriteRule ^/robots.txt /usr/local/galaxy-dist/static/robots.txt [L]
RewriteRule ^(.*) http://localhost:8081$1 [P]

The host lines are commented out in my universe_wsgi.ini files, so I'm guessing 
they should be set to the default localhost.

~ Brian
___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/

Re: [galaxy-dev] Defining new file formats in Galaxy (for new tool wrappers)

2011-06-02 Thread Greg Von Kuster

On Jun 2, 2011, at 1:29 PM, Nate Coraor wrote:

 Peter Cock wrote:
 Hi all,
 
 Something I've not needed to do until now is define a new file format
 in Galaxy. I understand the basic principle and defining a subclass in
 Python... however, how does this work with new tools on the Tool Shed?
 In particular, if an output format is likely to be used by more than
 one tool, can we get it added to the Galaxy core?
 
 I think people have provided the new subclass as a patch with the tool,
 but probably many of them, if well written, could be added to the core.
 
 As an example, the basic functionality of the Blast2GO for pipelines
 tool (b2g4pipe) takes a BLAST XML input file, and gives a tab
 separated annotation output file. Galaxy already has 'blastxml' and
 'tabular' file formats defined, so I didn't need to do anything extra.
 However, the tool can also take (a directory of) InterProScan XML
 files as input, so here a new 'interproscanxml' format would useful.
 Then any wrapper using or producing InterProScan XML could take
 advantage of this. e.g. Konrad's InterProScan wrapper could then offer
 the XML output as an option in addition to or instead of the tabular
 output.


We will certainly include support for new data formats into the Galaxy core.  
In case you haven't seen it, details for adding new formats is available in our 
wiki at https://bitbucket.org/galaxy/galaxy-central/wiki/AddingDatatypes.  It's 
fairly straightforward.  However, glancing at the wiki, it looks like there is 
no mention of functional tests for the new format.  If we could get a patch 
that includes a functional test for uploading the format as new method(s) in 
~/test/functional/test_get_data.py, it would be great.


 
 Related to this example, why isn't there a generic base class for XML
 formats in general?
 https://bitbucket.org/galaxy/galaxy-central/issue/568/missing-xml-datatype-base-class
 
 It just hadn't been necessary in the past and no one had the time to
 write it, I agree it could be helpful since there are other more
 specific XML types.

Yes, XML formats have not yet been abstracted, and certainly can be.  Just a 
matter of bandwidth...

 
 --nate
 
 
 Regards,
 
 Peter
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
 
  http://lists.bx.psu.edu/
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
 
  http://lists.bx.psu.edu/

Greg Von Kuster
Galaxy Development Team
g...@bx.psu.edu




___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


Re: [galaxy-dev] Problem with Torque/Maui

2011-06-02 Thread Nate Coraor
Hi Marco,

Thanks for all of the details, they make a big difference when
troubleshooting.  Sorry for the delay in response.

Can you ensure that you don't have any of the pbs_* options set in
universe_wsgi.ini?  I noticed that there are stagein/stageouts
set on the job.

You may also need to set $usecp in mom_priv/config on your execution
hosts to prevent pbs_mom from trying to rcp/scp the error and output
files back to galaxy1.  $usecp instructs pbs_mom to consider /path on
the execution host to be the same filesystem as on the submission host,
e.g.:

$usecp *:/mnt/equallogic1 /mnt/equallogic1

--nate

Marco Moretto wrote:
 Hi all,
 first of all, thanks to the Galaxy Team for this really useful software.
 Actually I don't really know if my problem is related with Galaxy or with
 Torque/Maui but I didn't find any solution looking in both Torque and Maui
 user lists, so I hope that some of you with more experience could give me
 some good advices. I'm trying to set up Galaxy in a small local virtual
 environment in order to test it. I started with 2 virtual ubuntu servers
 called galaxy1 and galaxy2. On galaxy1 I succesfully installed Galaxy,
 Apache, Torque and Maui. I'm using Postgres ad DBMS. It is installed on
 another real DB server. The virtual server galaxy2 is used as node. Galaxy
 is working like a charm locally but when I try to use Torque problems arise.
 Torque alone works correctly. That means that I can submit a job with qsub
 and everything works. The 2 virtual server (galaxy1 and galaxy2) share a
 directory (through NFS) in which I installed Galaxy following the unified
 method from the documentation.
 Now, as I said, Galaxy alone works, Torque/Maui alone works. When I put the
 two together nothing works.
 As a test I upload (using local runner) a gff file. Then I try to make a
 filter to the gff using Filter and sort - Extract features. When I run
 this tool the corresponding job on the Torque queue runs forever in Hold
 state. I report some output from diagnose programs:
 The diagnose -j reports the following:
 
 Name  State Par Proc QOS WCLimit R  Min User
  Group  Account  QueuedTime  Network  Opsys   ArchMem   Disk  Procs
   Class Features
 
 29 Hold DEF1 DEF 1:00:00 01   galaxy
 galaxy-00:02:36   [NONE] [NONE] [NONE]=0=0NC0
 [batch:1] [NONE]
 
 While the showq command reports
 
 ACTIVE JOBS
 JOBNAMEUSERNAME  STATE  PROC   REMAINING
  STARTTIME
 
 
  0 Active Jobs   0 of1 Processors Active (0.00%)
 
 IDLE JOBS--
 JOBNAMEUSERNAME  STATE  PROC WCLIMIT
  QUEUETIME
 
 
 0 Idle Jobs
 
 BLOCKED JOBS
 JOBNAMEUSERNAME  STATE  PROC WCLIMIT
  QUEUETIME
 
 29   galaxy   Hold 1 1:00:00  Wed May  4
 03:56:40
 
 
 The checkjob reports:
 checking job 29
 
 State: Hold
 Creds:  user:galaxy  group:galaxy  class:batch  qos:DEFAULT
 WallTime: 00:00:00 of 1:00:00
 SubmitTime: Wed May  4 03:56:40
   (Time Queued  Total: 00:03:07  Eligible: 00:00:01)
 
 The qstat -f reports
 
 Job Id: 33.galaxy1.research.intra.ismaa.it
 Job_Name = 27_extract_features1_marco.more...@iasma.it
 Job_Owner = gal...@galaxy1.research.intra.ismaa.it
 job_state = W
 queue = batch
 server = galaxy1.research.intra.ismaa.it
 ctime = Wed May  4 04:56:36 2011
 Error_Path =
 galaxy1:/mnt/equallogic1/galaxy/galaxy-dist/database/pbs/27.e
 
 exec_host = galaxy2/0
 exec_port = 15003
 Execution_Time = Wed May  4 05:26:41 2011
 mtime = Wed May  4 04:56:37 2011
 Output_Path =
 galaxy1:/mnt/equallogic1/galaxy/galaxy-dist/database/pbs/27.
 o
 qtime = Wed May  4 04:56:36 2011
 Resource_List.neednodes = 1
 Resource_List.nodect = 1
 Resource_List.nodes = 1
 Resource_List.walltime = 01:00:00
 stagein = /mnt/equallogic1/galaxy/tmp/dataset_18.dat@galaxy1
 :/mnt/equallog
 ic1/galaxy/galaxy-dist/database/files/000/dataset_18.dat,
 /mnt/equallogic1/galaxy/tmp/dataset_30.dat@galaxy1
 :/mnt/equallogic1/g
 alaxy/galaxy-dist/database/files/000/dataset_30.dat
 stageout =
 /mnt/equallogic1/galaxy/galaxy-dist/database/files/000/dataset_
 30.dat@galaxy1
 :/mnt/equallogic1/galaxy/galaxy-dist/database/files/000/
 dataset_30.dat
 substate = 37
 Variable_List = PBS_O_QUEUE=batch,
 PBS_O_HOST=galaxy1.research.intra.ismaa.it
 euser = galaxy
 egroup = galaxy
 hashname = 33.galaxy1.research.intra.ismaa.it
 queue_rank = 33
 queue_type = E
 
 StartDate: -00:03:06  Wed May  4 03:56:41
 Total Tasks: 1
 
 Req[0]  TaskCount: 1  Partition: DEFAULT
 Network: [NONE]  Memory = 0  Disk = 0  Swap = 0
 Opsys: [NONE]  Arch: [NONE]  Features: [NONE]
 
 
 IWD: [NONE]  Executable:  [NONE]
 Bypass: 0  StartCount: 1
 PartitionMask: [ALL]
 PE:  1.00  StartPriority:  1
 cannot select job 29 for 

Re: [galaxy-dev] suggestion for multithreading

2011-06-02 Thread Assaf Gordon
(moved to galaxy-dev)

Nate Coraor wrote, On 06/02/2011 01:31 PM:
 Peter Cock wrote:
 On Thu, Jun 2, 2011 at 6:23 PM, Nate Coraor n...@bx.psu.edu wrote:

 pbs.py then knows to translate 'resource type=cores8/resource' to
 '-l nodes=1:ppn=8'.

 Your tool can access that value a bunch, like $__resources__.cores.

 The same should be possible for other consumables.


Just a thought here:

The actual parameters that are passed to the scheduler are not necessarily 
hard-coded.
Meaning, at least with SGE, specifying the number of cores can be:
 qsub -pe threads=8
or
 qsub -pe cores=8
or
 qsub -pe jiffies=8

and same thing for memory limitation (e.g. -l virtual_free=800M).

The reason is that those resources (e.g. threads, cores, virtual_free) 
are just identifiers, and they are created and configured by whomever installed 
SGE - they are not built-in or hard-coded).

So just be careful in your design/implementation when automatically translating 
XML resources to hard-coded parameters.

If you do hard-code them, just make sure the specifically document it (i.e. 
Galaxy expect the SGE threads parameter to be -pe threads=8 and nothing else).

-gordon
___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


Re: [galaxy-dev] Appropriate place to ask questions about your NGLIMS for Galaxy

2011-06-02 Thread Nate Coraor
scott.cou...@monash.edu wrote:
 Hi,
 
 I'm a newbie when it comes to this kind of thing, so I was wondering whether 
 you can tell me the most appropriate place to ask questions about making some 
 modifications to your nglims extension for galaxy? Should they be posted to 
 the normal galaxy development list? Should I ask you directly?

Hi Scott,

It seems like maybe you intended to send this to Brad directly, but
regardless, posting nglims questions here is completely fine.  I don't
think he has a seperate mailing list for it, and we're planning to
integrate it with the Galaxy trunk code anyway.

--nate

 
 Cheers,
 
 Scott.
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
 
   http://lists.bx.psu.edu/
___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


Re: [galaxy-dev] Error when adding datasets

2011-06-02 Thread Nate Coraor
Louise-Amélie Schmitt wrote:
 On Thu, 2 Jun 2011 11:48:55 -0400, Nate Coraor n...@bx.psu.edu wrote:
  Greg Von Kuster wrote:
  Hello Louise,
  
  I've CC'd Nate on this as he may be able to help - although no
  guarantees.  I'm not expert enough in this area to know where to look
 for
  the cause.  Perhaps someone in the community can help as well.
  
  It is likely that LDAP is playing a role in this behavior.
  
  
  On Apr 14, 2011, at 1:00 PM, Louise-Amélie Schmitt wrote:
  
   The thing is, we use LDAP logging so we can't even access the website
   without logging in.
   Moreover, when I logged in, I arrived on the data analysis page where
   the
   automatic unnamed history was obviously created in the history
 panel.
   
   I forgot to mention we have issues creating and deleting histories,
   like,
   we can't access some histories and everytime we delete histories, two
   extra
   unnamed histories are created. As I mentioned before, it is also
   impossible
   to load a dataset in any history, existing or not.
  
  Okay, somehow I missed all the followup on this question when I
  committed a fix a few minutes ago.  Could you provide some details about
  the inaccessible histories (what happens when you try to access them,
  traceback, etc?) and loading data - you mean the upload tool fails?  Is
  there also a traceback here?
 
 ooh it's been a while I haven't seen the issue since I not-so-elegantly
 avoided the problem making our two instances run on separate machines.

When these were running on the same machine, were they both served as
seperate directories under the same hostname?  If so, did you set
cookie_path in the config?

If not, your cookies for one instance were being used with another
instance, with unpredictable results.  Although from recent discussion
here on -dev, it looks like there might be a bug with cookie_path.

--nate

 
 But I'll try to answer your question as accurately as I can.
 
 I experienced a similar error (pretty much the same traceback as the one I
 pasted in a previous message, just yelling at me saying that the history is
 nonexistent) whenever I tried to:
 - upload data in the manage libraries admin panel and in the analyse data
 panel (upload tool)
 - load an already uploaded dataset in any kind of history (saved or new)
 - probably do something else, I can't really remember
 
 I also had issues dealing with histories but not getting any traceback,
 like I mentioned a little lower: Everytime I tried to delete any number of
 histories, the histories were deleted but two were created.
 
 I hope this will help, sorry I can't really be more precise.
 L-A
 
 
  
  Thanks,
  --nate
  
   
   Do you think it could be linked to our using LDAP?
   
   Thanks
   L-A
   
   
   On Thu, 14 Apr 2011 12:06:55 -0400, Greg Von Kuster g...@bx.psu.edu
   wrote:
   On Apr 14, 2011, at 11:49 AM, Louise-Amélie Schmitt wrote:
   
   Here is the result I got from the debug statements:
   
   galaxy.web.controllers.library_common DEBUG 2011-04-14 17:46:02,286
   ###
   history: None
   
   This is the problem - when you registered normally, a history would
   have
   been automatically created for you.  But somehow you don't have a
   history -
   no idea why / how this happened, but it is very unlikely this is a
   result
   of a bug within Galaxy.  Try logging out and logging in again and a
   new
   history should be created for you.
   
   
   galaxy.web.controllers.library_common DEBUG 2011-04-14 17:46:02,286
   ###
   trans: galaxy.web.framework.GalaxyWebUITransaction object at
   0x29f40950
   galaxy.web.controllers.library_common DEBUG 2011-04-14 17:46:02,286
   ###
   trans.sa_session: sqlalchemy.orm.scoping.ScopedSession object at
   0x19ab21d0
   
   Thanks again
   L-A
   
   
   Le jeudi 14 avril 2011 à 11:05 -0400, Greg Von Kuster a écrit :
   I assume when you dropped the old database and recreated the new,
   empty database, you made sure the database connection string in
   universe_wsgi.ini was correctly set.  if so, when you started up
   Galaxy, it would have created all of the required tables in the
 new
   database, and they would all be empty.  When you first registered
 as
   the admin user, it would have automatically populated several
 tables
   with data, including the history table.  One or more of these
 things
   must not have been successful.
   
   
   To attempt to narrow down the problem, I'll need you to do the
   following.  Here are lines # 905 - 907 in
   ~/lib/galaxy/web/controllers/library-common.py
   
   
 # Send the current history to the form to enable importing
   datasets from history to library
 history = trans.get_history()
 trans.sa_session.refresh( history )
   
   
   Please add the following debug statements:
   
   
 # Send the current history to the form to enable importing
   datasets from history to library
 history = trans.get_history()
 log.debug(### history: %s % str( history ))

Re: [galaxy-dev] Update of cufflinks

2011-06-02 Thread Assaf Gordon
Just FYI,

With the recent version of cufflinks (at least 1.0.2), the developers added a 
'feature' of automatic version check.
While the idea is nice, the implications are not:
cufflinks,cuffcompare and cuffdiff will connect to host 
cufflinks.cbcb.umd.edu every time you run them.

I'm sure the intentions were good, but I don't like programs calling home, 
basically reporting usage statistics and origins.

The check_update code is also not written with security in mind, and might be 
exploited (buffer overruns and such) - it's best to disable it.
This will also generate a message to STDERR, which is just annoying (and bad, 
with galaxy).

To disable it on the command line, add --no-update-check to every invocation.

If you compiled it from source code, edit the file ./src/common.cpp and 
change the line:
   bool no_update_check = false;
to:
   bool no_update_check = true;

Then recompile and re-install.

-gordon


Jeremy Goecks wrote, On 05/10/2011 09:02 AM:
 
 We recently updated the Cufflinks/compare/diff wrappers to be compatible with 
 v1.0.* ; the wrappers are available in galaxy-central and should be available 
 in galaxy-dist in the next couple weeks as we're planning another release 
 soon.
 
 New features in v1.0.* have not been implemented in these updates (and, in 
 fact, there are features from 0.9.* that are still not available); the 
 changes we made were simply to ensure that the wrappers' current 
 functionality works with the new versions. 
 
 In the next couple months we plan to extend the current wrappers to include 
 new functionality. However, community contributions that extend the current 
 wrappers to include new functionality would be most welcome, and we can 
 integrate them into the galaxy code base if/when they are available.
 
 Best,
 J.
 
___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


[galaxy-dev] Early registration for BOSC ends tomorrow, Friday June 3

2011-06-02 Thread Brad Chapman
If you haven't already registered for BOSC, now is your chance--after 
June 3, prices will go up!

Registration for BOSC is through the ISMB main conference website:
http://www.iscb.org/ismbeccb2011-registration#sigs . Since BOSC is a
two-day SIG, the price is 2x the one-day SIG price listed on the ISMB
website. You can register for BOSC without registering for the main ISMB
conference, if you want.

The preliminary BOSC schedule (subject to change) is now up at
http://www.open-bio.org/wiki/BOSC_2011_Schedule (more details will be
added soon). There is also a two day Codefest proceeding BOSC; please add
yourself to the list of attendees if you are interested:
http://www.open-bio.org/wiki/Codefest_2011

The BOSC talks have already been chosen, but we have spaces
for last-minute posters. If you'd like your poster abstract
to appear in the BOSC program, you should submit it now--see
http://www.open-bio.org/wiki/BOSC_2011#Abstract_Submission_Information

We hope to see you at BOSC!

Nomi Harris
Co-Chair, BOSC 2011
___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


Re: [galaxy-dev] how can i get all the output file names in local.py file

2011-06-02 Thread Nate Coraor
shashi shekhar wrote:
 Hi All,
 
   I am using standalone system for   galaxy and  am doing some modification
 in galaxy .i need the all output file names in local.py .How can i get all
 the output file names in local.py file.

From the job wrapper object in local.py, job_wrapper.get_output_fnames()

--nate

 
 Regards
 shashi shekhar

 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
 
   http://lists.bx.psu.edu/

___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


Re: [galaxy-dev] Separate histories between localhost and 127.0.0.1

2011-06-02 Thread Nate Coraor
Lewis, Brian Andrew wrote:
 For some odd reason, when I try to import data using the URL 
 http://localhost:8081 or http://localhost:8081/galaxy none of my jobs will 
 even show up.  However if I go to http://127.0.0.1 or http://127.0.0.1/galaxy 
 the jobs run fine.  Here's a snip from my httpd.conf file:
 
 RewriteEngine on
 RewriteRule ^/static/style/(.*) 
 /usr/local/galaxy-dist/static/june_2007_style/blue/$1 [L]
 RewriteRule ^/static/scripts/(.*) 
 /usr/local/galaxy-dist/static/scripts/packed/$1 [L]
 RewriteRule ^/static/(.*) /usr/local/galaxy-dist/static/$1 [L]
 RewriteRule ^/favicon.ico /usr/local/galaxy-dist/static/favicon.ico [L]
 RewriteRule ^/robots.txt /usr/local/galaxy-dist/static/robots.txt [L]
 RewriteRule ^(.*) http://localhost:8081$1 [P]
 
 The host lines are commented out in my universe_wsgi.ini files, so I'm 
 guessing they should be set to the default localhost.

Hi Brian,

Since localhost and 127.0.0.1 are considered different from the server
name perspective, the cookies will be different.  If you access
http://localhost or http://localhost:8081 you'll share cookies.
Likewise, if you access http://127.0.0.1 or http://127.0.0.1:8081, those
cookies will be shared.  Mixing localhost and 127.0.0.1 results in
seperate cookies being issues for both.

If you're logged in, you can view your history with the localhost URLs
by clicking Saved histories from the Options menu.

--nate

 
 ~ Brian

 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
 
   http://lists.bx.psu.edu/

___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/