We're regularly running BatchJobs itself on an LSF cluster. Works great.

On Thu, Jun 6, 2013 at 5:48 PM, Henrik Bengtsson <h...@biostat.ucsf.edu>wrote:

> Great - this looks promising already.
>
> What's your test system(s), beyond standard SSH and multicore
> clusters?  I'm on a Torque/PBS system.
>
> I'm happy to test, give feedback etc.  I don't see an 'Issues' tab on
> the GitHub page.  Michel, how do you prefer to get feedback?
>
> /Henrik
>
>
> On Thu, Jun 6, 2013 at 5:21 PM, Michael Lawrence
> <lawrence.mich...@gene.com> wrote:
> > And here is the on-going development of the backend:
> > https://github.com/mllg/BiocParallel/tree/batchjobs
> >
> > Not sure how well it's been tested.
> >
> > Kudos to Michel Lang for making so much progress so quickly.
> >
> > Michael
> >
> > On Thu, Jun 6, 2013 at 1:59 PM, Dan Tenenbaum <dtene...@fhcrc.org>
> wrote:
> >>
> >> On Thu, Jun 6, 2013 at 1:56 PM, Henrik Bengtsson <h...@biostat.ucsf.edu>
> >> wrote:
> >> > Hi, I'd like to pick up the discussion on a BatchJobs backend for
> >> > BiocParallel where it was left back in Dec 2012 (Bioc-devel thread
> >> > 'BiocParallel'
> >> > [https://stat.ethz.ch/pipermail/bioc-devel/2012-December/003918.html]
> ).
> >> >
> >> > Florian, would you mind sharing your BatchJobs backend code?  Is it
> >> > independent of BiocParallel and/or have you tried it with the most
> >> > recent BiocParallel implementation
> >> > [https://github.com/Bioconductor/BiocParallel/]?
> >> >
> >>
> >> You should be aware that there is  Google Summer of Code project in
> >> progress to address this.
> >>
> >> http://www.bioconductor.org/developers/gsoc2013/ (towards the bottom)
> >>
> >> Dan
> >>
> >>
> >> > /Henrik
> >> >
> >> > On Tue, Dec 4, 2012 at 12:38 PM, Henrik Bengtsson <
> h...@biostat.ucsf.edu>
> >> > wrote:
> >> >> Thanks.
> >> >>
> >> >> On Tue, Dec 4, 2012 at 3:47 AM, Vincent Carey
> >> >> <st...@channing.harvard.edu> wrote:
> >> >>> I have been booked up so no chance to deploy but I do have access to
> >> >>> SGE and
> >> >>> LSF so will try and will report ASAP.
> >> >>
> >> >> ...and I'll try it out on PBS (... but I most likely won't have time
> >> >> to do this until the end of the year).
> >> >>
> >> >> Henrik
> >> >>
> >> >>>
> >> >>>
> >> >>> On Tue, Dec 4, 2012 at 4:08 AM, Hahne, Florian
> >> >>> <florian.ha...@novartis.com>
> >> >>> wrote:
> >> >>>>
> >> >>>> Hi Henrik,
> >> >>>> I have now come up now with a relatively generic version of this
> >> >>>> SGEcluster approach. It does indeed use BatchJobs under the hood
> and
> >> >>>> should thus support all available cluster queues, assuming that the
> >> >>>> necessary batchJobs routines are available. I could only test this
> on
> >> >>>> our
> >> >>>> SGE cluster, but Vince wanted to try other queuing systems. Not
> sure
> >> >>>> how
> >> >>>> far he got. For now the code is wrapped in a little package called
> >> >>>> Qcluster with some documentation. If you want to I can send you a
> >> >>>> version
> >> >>>> in a separate mail. Would be good to test this on other systems,
> and
> >> >>>> I am
> >> >>>> sure there remain some bugs that need to be ironed out. In
> particular
> >> >>>> the
> >> >>>> fault tolerance you mentioned needs to be addressed properly.
> >> >>>> Currently
> >> >>>> the code may leave unwanted garbage if things fail in the wrong
> >> >>>> places
> >> >>>> because all the communication is file-based.
> >> >>>> Martin, I'll send you my updated version in case you want to
> include
> >> >>>> this
> >> >>>> in biocParallel for others to contribute.
> >> >>>> Florian
> >> >>>> --
> >> >>>>
> >> >>>>
> >> >>>>
> >> >>>>
> >> >>>>
> >> >>>>
> >> >>>> On 12/4/12 5:46 AM, "Henrik Bengtsson" <h...@biostat.ucsf.edu>
> wrote:
> >> >>>>
> >> >>>> >Picking up this thread in lack of other places (= were should
> >> >>>> >BiocParallel be discussed?)
> >> >>>> >
> >> >>>> >I saw Martin's updates on the BiocParallel - great.  Florian's SGE
> >> >>>> >scheduler was also mentioned; is that one built on top of
> BatchJobs?
> >> >>>> >If so I'd be interested in looking into that/generalizing that to
> >> >>>> > work
> >> >>>> >with any BatchJobs scheduler.
> >> >>>> >
> >> >>>> >I believe there is going to be a new release of BatchJobs rather
> >> >>>> > soon,
> >> >>>> >so it's probably worth waiting until that is available.
> >> >>>> >
> >> >>>> >The main use case I'm interested in is to launch batch jobs on a
> >> >>>> >PBS/Torque cluster, and then use multicore processing on each
> >> >>>> > compute
> >> >>>> >node.  It would be nice to be able to do this using the
> BiocParallel
> >> >>>> >model, but maybe it is too optimistic to get everything to work
> >> >>>> > under
> >> >>>> >same model.  Also, as Vince hinted, fault tolerance etc needs to
> be
> >> >>>> >addressed and needs to be addressed differently in the different
> >> >>>> >setups.
> >> >>>> >
> >> >>>> >/Henrik
> >> >>>> >
> >> >>>> >On Tue, Nov 20, 2012 at 6:59 AM, Ramon Diaz-Uriarte
> >> >>>> > <rdia...@gmail.com>
> >> >>>> >wrote:
> >> >>>> >>
> >> >>>> >>
> >> >>>> >>
> >> >>>> >> On Sat, 17 Nov 2012 13:05:29 -0800,"Ryan C. Thompson"
> >> >>>> >><r...@thompsonclan.org> wrote:
> >> >>>> >>
> >> >>>> >>> On 11/17/2012 02:39 AM, Ramon Diaz-Uriarte wrote:
> >> >>>> >>> > In addition to Steve's comment, is it really a good thing
> that
> >> >>>> >>> > "all
> >> >>>> >>>code
> >> >>>> >>> > stays the same."?  I mean, multiple machines vs. multiple
> cores
> >> >>>> >>> > are,
> >> >>>> >>> > often, _very_ different things: for instance, shared vs.
> >> >>>> >>> > distributed
> >> >>>> >>> > memory, communication overhead differences, whether or not
> you
> >> >>>> >>> > can
> >> >>>> >>>assume
> >> >>>> >>> > packages and objects to be automagically present in the
> >> >>>> >>> > slaves/child
> >> >>>> >>> > process, etc. So, given they are different situations, I
> think
> >> >>>> >>> > it
> >> >>>> >>> > sometimes makes sense to want to write different code for
> each
> >> >>>> >>>situation
> >> >>>> >>> > (I often do); not to mention Steve's hybrid cases ;-).
> >> >>>> >>> >
> >> >>>> >>> >
> >> >>>> >>> > Since BiocParallel seems to be a major undertaking, maybe it
> >> >>>> >>> > would
> >> >>>> >>> > be
> >> >>>> >>> > appropriate to provide a flexible approach, instead of hard
> >> >>>> >>> > wiring
> >> >>>> >>>the
> >> >>>> >>> > foreach approach.
> >> >>>> >>> Of course there are cases where the same code simply can't work
> >> >>>> >>> for
> >> >>>> >>>both
> >> >>>> >>> multicore and multi-machine situations, but those generally
> don't
> >> >>>> >>> fall
> >> >>>> >>> into the category of things that can be done using lapply.
> Lapply
> >> >>>> >>> and
> >> >>>> >>> all of its parallelized buddies like mclapply, parLapply, and
> >> >>>> >>> foreach
> >> >>>> >>> are designed for data-parallel operations with no
> interdependence
> >> >>>> >>> between results, and these kinds of operations generally
> >> >>>> >>> parallelize
> >> >>>> >>> as
> >> >>>> >>> well across machines as across cores, unless your network is
> not
> >> >>>> >>> fast
> >> >>>> >>> enough (in which case you would choose not to use multi-machine
> >> >>>> >>> parallelism). If you want a parallel algorithm for something
> like
> >> >>>> >>> the
> >> >>>> >>> disjoin method of GRanges, you might need to write some special
> >> >>>> >>> purpose
> >> >>>> >>> code, and that code might be very different for multicore vs
> >> >>>> >>>multi-machine.
> >> >>>> >>
> >> >>>> >>> So yes, sometimes there is a fundamental reason that you have
> to
> >> >>>> >>> change
> >> >>>> >>> the code to make it run on multiple machines, and neither
> foreach
> >> >>>> >>> nor
> >> >>>> >>> any other parallelization framework will save you from having
> to
> >> >>>> >>>rewrite
> >> >>>> >>> your code. But often there is no fundamental reason that the
> code
> >> >>>> >>> has
> >> >>>> >>>to
> >> >>>> >>> change, but you end up changing it anyway because of
> limitations
> >> >>>> >>> in
> >> >>>> >>>your
> >> >>>> >>> parallelization framework. This is the case that foreach saves
> >> >>>> >>> you
> >> >>>> >>>from.
> >> >>>> >>
> >> >>>> >>
> >> >>>> >>
> >> >>>> >> Hummm... I guess you are right, and we are talking about "often"
> >> >>>> >> or
> >> >>>> >>"most
> >> >>>> >> of the time", which is where all this would fit. Point taken.
> >> >>>> >>
> >> >>>> >>
> >> >>>> >> Best,
> >> >>>> >>
> >> >>>> >> R.
> >> >>>> >>
> >> >>>> >>
> >> >>>> >>
> >> >>>> >>
> >> >>>> >>
> >> >>>> >>
> >> >>>> >> --
> >> >>>> >> Ramon Diaz-Uriarte
> >> >>>> >> Department of Biochemistry, Lab B-25
> >> >>>> >> Facultad de Medicina
> >> >>>> >> Universidad Autónoma de Madrid
> >> >>>> >> Arzobispo Morcillo, 4
> >> >>>> >> 28029 Madrid
> >> >>>> >> Spain
> >> >>>> >>
> >> >>>> >> Phone: +34-91-497-2412
> >> >>>> >>
> >> >>>> >> Email: rdia...@gmail.com
> >> >>>> >>        ramon.d...@iib.uam.es
> >> >>>> >>
> >> >>>> >> http://ligarto.org/rdiaz
> >> >>>> >>
> >> >>>> >> _______________________________________________
> >> >>>> >> Bioc-devel@r-project.org mailing list
> >> >>>> >> https://stat.ethz.ch/mailman/listinfo/bioc-devel
> >> >>>>
> >> >>>
> >> >
> >> > _______________________________________________
> >> > Bioc-devel@r-project.org mailing list
> >> > https://stat.ethz.ch/mailman/listinfo/bioc-devel
> >
> >
>

        [[alternative HTML version deleted]]

_______________________________________________
Bioc-devel@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/bioc-devel

Reply via email to