Hi, Thanks for asking question.
The reason is simply because of it is running code statements. The statements can have order and dependency. Imagine i have two paragraphs %spark val a = 1 %spark print(a) If they're not running one by one, that means they possibly runs in random order and the output will be always different. Either '1' or 'val a can not found'. This is the reason why. But if there are nice idea to handle this problem i agree using parallel scheduler would help a lot. Thanks, moon On 2015년 7월 14일 (화) at 오후 7:59 linxi zeng <linxizeng0...@gmail.com> wrote: > any one who have the same question with me? or this is not a question? > > 2015-07-14 11:47 GMT+08:00 linxi zeng <linxizeng0...@gmail.com>: > >> hi, Moon: >> I notice that the getScheduler function in the SparkInterpreter.java >> return a FIFOScheduler which makes the spark interpreter run spark job one >> by one. It's not a good experience when couple of users do some work on >> zeppelin at the same time, because they have to wait for each other. >> And at the same time, SparkSqlInterpreter can chose what scheduler to use >> by "zeppelin.spark.concurrentSQL". >> My question is, what kind of consideration do you based on to make such >> a decision? >> > >