Hi,

You can pass variable from between spark and pyspark in this way.

%spark
z.put("var1", myVar)

%pyspark
myVar = z.get("var1")

The other way (pyspark->spark) also works.

For SparkSQL, one possible way to use variable is

%spark
z.show(sqlContext.sql(s"select .... ${myVar}"))

Hope this helps.

Thanks,
moon

On Thu, Jul 23, 2015 at 2:36 AM TEJA SRIVASTAV <tejasrivas...@gmail.com>
wrote:

> Hello all,
>
> Is it possible to pass different variables or different interpreters
> thoroughout a notebook
>
> Like i should be able to access the variables between the paragraph use
> the data for visualization in d3 via anuglar or scala
>
> and then pass those variables for processing in SQL or python
>
>
> --
> With regards,
> Teja Srivastav
> +91 9066 82 32 80
>

Reply via email to