0.8.1 we used branch 0.8 and pull request into our local repo. I remember we have to deal with few issues but once we are thought that its working great. On Mar 10, 2014 6:51 PM, "Mayur Rustagi" <mayur.rust...@gmail.com> wrote:
> Which version of Spark are you using? > > > Mayur Rustagi > Ph: +1 (760) 203 3257 > http://www.sigmoidanalytics.com > @mayur_rustagi <https://twitter.com/mayur_rustagi> > > > > On Mon, Mar 10, 2014 at 6:49 PM, abhinav chowdary < > abhinav.chowd...@gmail.com> wrote: > >> for any one who is interested to know about job server from Ooyala.. we >> started using it recently and been working great so far.. >> On Feb 25, 2014 9:23 PM, "Ognen Duzlevski" <og...@nengoiksvelzud.com> >> wrote: >> >>> In that case, I must have misunderstood the following (from >>> http://spark.incubator.apache.org/docs/0.8.1/job-scheduling.html). >>> Apologies. Ognen >>> >>> "Inside a given Spark application (SparkContext instance), multiple >>> parallel jobs can run simultaneously if they were submitted from separate >>> threads. By "job", in this section, we mean a Spark action (e.g. save, >>> collect) and any tasks that need to run to evaluate that action. >>> Spark's scheduler is fully thread-safe and supports this use case to enable >>> applications that serve multiple requests (e.g. queries for multiple >>> users). >>> >>> By default, Spark's scheduler runs jobs in FIFO fashion. Each job is >>> divided into "stages" (e.g. map and reduce phases), and the first job gets >>> priority on all available resources while its stages have tasks to launch, >>> then the second job gets priority, etc. If the jobs at the head of the >>> queue don't need to use the whole cluster, later jobs can start to run >>> right away, but if the jobs at the head of the queue are large, then later >>> jobs may be delayed significantly. >>> >>> Starting in Spark 0.8, it is also possible to configure fair sharing >>> between jobs. Under fair sharing, Spark assigns tasks between jobs in a >>> "round robin" fashion, so that all jobs get a roughly equal share of >>> cluster resources. This means that short jobs submitted while a long job is >>> running can start receiving resources right away and still get good >>> response times, without waiting for the long job to finish. This mode is >>> best for multi-user settings. >>> >>> To enable the fair scheduler, simply set the spark.scheduler.mode to >>> FAIR before creating a SparkContext:" >>> On 2/25/14, 12:30 PM, Mayur Rustagi wrote: >>> >>> fair scheduler merely reorders tasks .. I think he is looking to run >>> multiple pieces of code on a single context on demand from customers...if >>> the code & order is decided then fair scheduler will ensure that all tasks >>> get equal cluster time :) >>> >>> >>> >>> Mayur Rustagi >>> Ph: +919632149971 >>> h <https://twitter.com/mayur_rustagi>ttp://www.sigmoidanalytics.com >>> https://twitter.com/mayur_rustagi >>> >>> >>> >>> On Tue, Feb 25, 2014 at 10:24 AM, Ognen Duzlevski < >>> og...@nengoiksvelzud.com> wrote: >>> >>>> Doesn't the fair scheduler solve this? >>>> Ognen >>>> >>>> >>>> On 2/25/14, 12:08 PM, abhinav chowdary wrote: >>>> >>>> Sorry for not being clear earlier >>>> how do you want to pass the operations to the spark context? >>>> this is partly what i am looking for . How to access the active spark >>>> context and possible ways to pass operations >>>> >>>> Thanks >>>> >>>> >>>> >>>> On Tue, Feb 25, 2014 at 10:02 AM, Mayur Rustagi < >>>> mayur.rust...@gmail.com> wrote: >>>> >>>>> how do you want to pass the operations to the spark context? >>>>> >>>>> >>>>> Mayur Rustagi >>>>> Ph: +919632149971 >>>>> h <https://twitter.com/mayur_rustagi>ttp://www.sigmoidanalytics.com >>>>> https://twitter.com/mayur_rustagi >>>>> >>>>> >>>>> >>>>> On Tue, Feb 25, 2014 at 9:59 AM, abhinav chowdary < >>>>> abhinav.chowd...@gmail.com> wrote: >>>>> >>>>>> Hi, >>>>>> I am looking for ways to share the sparkContext, meaning i >>>>>> need to be able to perform multiple operations on the same spark context. >>>>>> >>>>>> Below is code of a simple app i am testing >>>>>> >>>>>> def main(args: Array[String]) { >>>>>> println("Welcome to example application!") >>>>>> >>>>>> val sc = new SparkContext("spark://10.128.228.142:7077", >>>>>> "Simple App") >>>>>> >>>>>> println("Spark context created!") >>>>>> >>>>>> println("Creating RDD!") >>>>>> >>>>>> Now once this context is created i want to access this to submit >>>>>> multiple jobs/operations >>>>>> >>>>>> Any help is much appreciated >>>>>> >>>>>> Thanks >>>>>> >>>>>> >>>>>> >>>>>> >>>>> >>>> >>>> >>>> -- >>>> Warm Regards >>>> Abhinav Chowdary >>>> >>>> >>>> >>> >>> >