We may introduce some paragraph dependency system to remove this limitation.

Indeed, in the InterpreterContext object we can introduce dependency link
between different paragraphs so that we guarantee a correct ordering of
execution

It would imply updating the AngularJS code to add a config section for
paragraph dependency

After the refactoring of ZeppelinContext, I might do a POC to see how it
can be done easily.

After that, all interpreters can leverage the ParallelScheduler

On Wed, Jul 15, 2015 at 3:45 PM, moon soo Lee <m...@apache.org> wrote:

> Hi,
>
> Thanks for asking question.
>
> The reason is simply because of it is running code statements. The
> statements can have order and dependency. Imagine i have two paragraphs
>
> %spark
> val a = 1
>
> %spark
> print(a)
>
> If they're not running one by one, that means they possibly runs in random
> order and the output will be always different. Either '1' or 'val a can not
> found'.
>
> This is the reason why. But if there are nice idea to handle this problem
> i agree using parallel scheduler would help a lot.
>
> Thanks,
> moon
>
> On 2015년 7월 14일 (화) at 오후 7:59 linxi zeng <linxizeng0...@gmail.com> wrote:
>
>> any one who have the same question with me? or this is not a question?
>>
>> 2015-07-14 11:47 GMT+08:00 linxi zeng <linxizeng0...@gmail.com>:
>>
>>> hi, Moon:
>>>    I notice that the getScheduler function in the SparkInterpreter.java
>>> return a FIFOScheduler which makes the spark interpreter run spark job one
>>> by one. It's not a good experience when couple of users do some work on
>>> zeppelin at the same time, because they have to wait for each other.
>>> And at the same time, SparkSqlInterpreter can chose what scheduler to
>>> use by "zeppelin.spark.concurrentSQL".
>>> My question is, what kind of consideration do you based on to make such
>>> a decision?
>>>
>>
>>

Reply via email to