Re: [galaxy-dev] cloud instance missing /opt/sge/default/common directory

2011-09-26 Thread Enis Afgan
Are you continuously running this instance and just restarting it or do you
start a clean instance each time?
I'd suggest ensuring the persistent_data.yaml in your cluster's bucket is
correct (see below), terminating this instance, and starting a new one with
the same user data as used for this cluster. That way you should get a clean
(i.e., standardized) start and that is something that should follow the
process correctly.

persistent_data.yaml contains the data required to reestablish an existing
cluster and must follow the following format but contain proper values
(especially if you customized your cluster via custom snapshots). This file
is in your cluster's bucket on S3 so you can download it, check it, and
upload if anything was changed or got out of sync.

data_filesystems:
  galaxyData:
  - size: 50
vol_id: vol-909342f8
services:
- service: SGE
- service: Postgres
- service: Galaxy
static_filesystems:
- filesystem: galaxyIndices
  size: 700
  snap_id: !!python/unicode 'snap-5b030634'
- filesystem: galaxyTools
  size: 2
  snap_id: !!python/unicode 'snap-1688b978'



On Fri, Sep 23, 2011 at 1:09 PM, Joseph Hargitai 
joseph.hargi...@einstein.yu.edu wrote:

  Hi,

 the mount is no longer there:

 ubuntu@ip-10-68-42-15:~$ more /etc/hosts
 127.0.0.1 localhost

 # The following lines are desirable for IPv6 capable hosts
 ::1 ip6-localhost ip6-loopback
 fe00::0 ip6-localnet
 ff00::0 ip6-mcastprefix
 ff02::1 ip6-allnodes
 ff02::2 ip6-allrouters
 ff02::3 ip6-allhosts
 ubuntu@ip-10-68-42-15:~$ df -h
 FilesystemSize  Used Avail Use% Mounted on
 /dev/sda1  15G  9.5G  5.6G  63% /
 devtmpfs  7.3G  128K  7.3G   1% /dev
 none  7.6G 0  7.6G   0% /dev/shm
 none  7.6G   96K  7.6G   1% /var/run
 none  7.6G 0  7.6G   0% /var/lock
 none  7.6G 0  7.6G   0% /lib/init/rw
 /dev/sdb  414G  201M  393G   1% /mnt
 ubuntu@ip-10-68-42-15:~$ cd /mnt
 ubuntu@ip-10-68-42-15:/mnt$ ls
 cm  lost+found



 ubuntu@ip-10-68-42-15:/mnt$ more /etc/fstab
 # /etc/fstab: static file system information.
 # file system mount point   type
 options
dump  pass
 proc/proc   proc
 nodev,no
 exec,nosuid 0   0
 /dev/sda1   /   xfs
 defaults
 0   0
 /dev/sdb/mntautodefaults,nobootwait,comment=cloudconfig0

 0


  --
 *From:* Enis Afgan [eaf...@emory.edu]
 *Sent:* Thursday, September 22, 2011 8:48 AM

 *To:* Joseph Hargitai
 *Cc:* galaxy-dev@lists.bx.psu.edu
 *Subject:* Re: [galaxy-dev] cloud instance missing /opt/sge/default/common
 directory

   Hi Joe,
 And this is happening on a freshly booted instance (from
 a previously existing cluster) using the same AMI? The order of execution
 seems a bit odd, seeing Galaxy being removed before SGE is setup; SGE should
 be the first thing that gets setup so I'm wondering...

  If you log into the instance, what is in /etc/hosts? Does it match the
 instance DNS?

  And if you try executing that same command (cd /opt/sge; ./inst_sge -m -x
 -auto /opt/sge/galaxyEC2.conf) by hand (as root), is any more info
 produced? Also, qmaster log should be available under /opt/sge/ge6 (or
 something like this) /default/spool/qmaster/ so please take a look there as
 well and see if more info is available.



 On Thu, Sep 22, 2011 at 12:46 AM, Joseph Hargitai 
 joseph.hargi...@einstein.yu.edu wrote:

  the error is

 '
 [DEBUG] galaxy:139 2011-09-22 00:03:21,055: Galaxy UI does not seem to
 be accessible.
 [DEBUG] master:1491 2011-09-22 00:03:21,055: SS: SGE..Shut down;
 FS-galaxyIndices..OK; FS-galaxyTools..OK; FS-galaxyData..OK; Postgres..OK;
 Galaxy..Starting;
 [DEBUG] root:354 2011-09-22 00:03:24,724: Managing services: []
 [INFO] galaxy:30 2011-09-22 00:03:24,724: Removing 'Galaxy' service
 [INFO] galaxy:122 2011-09-22 00:03:24,724: Shutting down Galaxy...
 [DEBUG] misc:511 2011-09-22 00:03:26,067: Successfully stopped Galaxy.
 [DEBUG] root:354 2011-09-22 00:03:33,936: Managing services: []
 [DEBUG] sge:61 2011-09-22 00:03:33,937: Unpacking SGE from
 '/opt/galaxy/pkg/ge6.2u5'
 [DEBUG] sge:76 2011-09-22 00:03:33,937: Cleaning '/opt/sge' directory.
 [DEBUG] sge:82 2011-09-22 00:03:34,117: Unpacking SGE to '/opt/sge'.
 [INFO] sge:96 2011-09-22 00:03:35,557: Configuring SGE...
 [DEBUG] sge:104 2011-09-22 00:03:35,558: Created SGE install template as
 file '/opt/sge/galaxyEC2.conf'
 [DEBUG] sge:112 2011-09-22 00:03:35,558: Setting up SGE.
 [ERROR] misc:514 2011-09-22 00:03:35,651: Setting up SGE did not go
 smoothly, running command 'cd /opt/sge; ./inst_sge -m -x -auto
 /opt/sge/galaxyEC2.conf' returned code '2' and following stderr: '[: 359:
 11: unexpected operator
 [: 359: 11: unexpected operator
 [: 359: 11: unexpected operator
 [: 359: 11: unexpected operator
 error resolving local host: can't 

Re: [galaxy-dev] Galaxy fronting multiple clusters

2011-09-26 Thread Luobin Yang
I have never setup a Galaxy instance fronting multiple clusters, but it's
something I would like to explore. I have a dedicated cluster to run Galaxy
jobs and I've got another a shared cluster to which I hope Galaxy can assign
jobs when the dedicated cluster is too busy.

From my understanding of Galaxy, the tool runner for each tool is hardcoded
in the universe.ini file and if you do not configure a tool runner for a
tool, Galaxy will use the default tool runner, which is determined by the
default_cluster_job_runner parameter. I believe you can configure multiple
job runners for a specific tool under the [galaxy:tool_runners] in
universe.ini, for instance, for each cluster, you have a different tool
runner for a specific tool,  however, Galaxy probably will just use one of
them, most likely the last one. So cluster selection for a tool is
determined by the job runner, which is hard coded in the universe.ini file.
As a result, the running of a workflow is determined by the tools in the
workflow. If each tool in the workflow is configured to use the same
cluster, then the workflow is run on the same cluster, otherwise, it will
span multiple clusters.

I think if you can configure the machine that runs Galaxy instance to be the
submit host of multiple clusters, then it's possible to have Galaxy front
multiple clusters. For me, the biggest hurdle is how to let two clusters
having a shared storage space and configure a machine in one cluster to be
the submit host of another cluster.

Thanks,
Luoin

On Mon, Sep 19, 2011 at 9:44 AM, Ann Black annbl...@eng.uiowa.edu wrote:

 Hello -

 I am working on standing up our own galaxy installation.  We would like to
 have galaxy front multiple clusters, and I have some questions I was hoping
 someone could help with.

 1)  From reading other forum posts on this subject, it seems I need to
 minimally do the following ... is this correct?:
   A) have galaxy server w/ sge register as a job submitting host to the
 head node of each cluster
   B) Configure multiple tool runners for each tool per remote cluster?

 2) When galaxy would submit a job, how would a backend remote cluster be
 selected?  When running workflows, would the same cluster be used to run the
 entire workflow - or could the workflow then span remote clusters?

 3) I am trying to understand some of the source code, where is the logic
 that would then dispatch the job and select a job runner to use?

 4) Other advice or steps needed in order to get galaxy to front multiple
 remote clusters?


 Thanks so much,

 Ann


 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/

___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/

Re: [galaxy-dev] picard error.

2011-09-26 Thread Ross
A fix for these warnings is in revision 6048 of galaxy-central -
again, thanks for pointing them out Michael.

For the record, the warnings were from the HSmetrics and GCbias tools
and arose because there is a user configurable maxheap parameter that
is not available by default - it is commented out in the tool form -
but it had not been commented out from the test - since it is not
enabled by default that generated a warning. The test parameter is now
also commented out by default and must be uncommented for sites
running into ram limits with very large datasets.

On Tue, Sep 27, 2011 at 8:04 AM, Ross ross.laza...@gmail.com wrote:
 Michael,

 Thanks for pointing these warning errors out. I think they're harmless
 and can be ignored but I'll definitely take a look later today and
 check in a fix.
 As I recall, we had some discussion some months ago about whether to
 allow the user to set the size of the java heap and decided that it
 was not necessary - these messages suggest the parameter is still
 somewhere in the tool tests - oddly enough I can't see it but I'm
 seeing the same messages!
 I'll let you know what I find once I've had a chance to take a look.


 On Mon, Sep 26, 2011 at 11:06 PM, michael burrell (TOC)
 michael.burr...@nbi.ac.uk wrote:
 Good afternoon,



 When starting galaxy I find the following errors logged in the paster.log



 galaxy.tools DEBUG 2011-09-26 13:58:21,577 Loaded tool: PicardASMetrics 0.03

 galaxy.tools.test DEBUG 2011-09-26 13:58:21,625 Error in add_param for
 maxheap: Unable to determine parameter type of test input 'maxheap'. Ensure
 that the parameter exists and that any container groups are defined first.



 galaxy.tools DEBUG 2011-09-26 13:58:21,703 Loaded tool: PicardInsertSize
 0.3.0

 galaxy.tools.test DEBUG 2011-09-26 13:58:21,737 Error in add_param for
 maxheap: Unable to determine parameter type of test input 'maxheap'. Ensure
 that the parameter exists and that any container groups are defined first.



 Our galaxy version is.



 galaxy@jic55666:~/software/galaxy-central$ hg tip

 changeset:   6042:2469c53051ea

 tag: tip

 user:    Daniel Blankenberg d...@bx.psu.edu

 date:    Sun Sep 25 13:27:14 2011 -0400

 summary: Some random lines tool tweaks.



 Thank you for your assistance.



 Michael.



 
 Please note, that my email address is no longer
 givenname.surn...@bbsrc.ac.uk. Please update your records to reflect my new
 email address of givenname.surn...@nbi.ac.uk as shown on this email.
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/




 --
 Ross Lazarus MBBS MPH;
 Associate Professor, Harvard Medical School;
 Director of Bioinformatics, Channing Lab; Tel: +1 617 505 4850;
 Head, Medical Bioinformatics, BakerIDI; Tel: +61 385321444;




-- 
Ross Lazarus MBBS MPH;
Associate Professor, Harvard Medical School;
Director of Bioinformatics, Channing Lab; Tel: +1 617 505 4850;
Head, Medical Bioinformatics, BakerIDI; Tel: +61 385321444;

___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/