I just did that using the RC3 version of Toree for the 0.1.x branch. If
you're on master, maybe it doesn't require _jvm_kernel. I just saw that was
needed for our RC3.

On Wed, Nov 2, 2016 at 12:12 PM <[email protected]> wrote:

> That is not working for me in the release I have 0.1.0…
>
> %%pyspark
> print dir(kernel._jvm_kernel)
>
>
> ['__call__', '__class__', '__delattr__', '__dict__', '__doc__',
> '__format__', '__getattribute__', '__hash__', '__init__', '__module__',
> '__new__', '__reduce__', '__reduce_ex__', '__repr__', '__setattr__',
> '__sizeof__', '__str__', '__subclasshook__', '__weakref__', '_get_args',
> 'command_header', 'container', 'converters', 'gateway_client', 'name',
> 'pool', 'target_id’]
>
>
> Any ideas?
>
> Thanks,
> Ian
>
>
>
>
>
>
>
> On 11/2/16, 12:38 PM, "Chip Senkbeil" <[email protected]> wrote:
>
> >When running Scala as the default interpreter in a notebook:
> >
> >Cell 1:
> >val x = 3
> >kernel.data.put("x", x)
> >
> >Cell2:
> >%%pyspark
> >x = kernel._jvm_kernel.data().get("x")
> >kernel._jvm_kernel.data().put("x", x + 1)
> >
> >Cell3:
> >println(kernel.data.get("x"))
> >
> >On Wed, Nov 2, 2016 at 11:27 AM <[email protected]> wrote:
> >
> >> Thanks Chip, now, I understand how to work with it from the JVM side.
> >>Any
> >> chance you have a snippet of how to get a value from the map in python?
> >>
> >> Ian
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >> On 11/2/16, 11:39 AM, "Chip Senkbeil" <[email protected]> wrote:
> >>
> >> >While it isn't supported (we don't test its use in this case), you can
> >> >store objects in a shared hashmap under the kernel object that is made
> >> >available in each interpreter. The map is exposed as `kernel.data`, but
> >> >the
> >> >way you access and store data is different per language.
> >> >
> >> >The signature of the data map on the kernel is `val data:
> >>java.util.Map[
> >> >String, Any]` and we use a concurrent hashmap, so it can handle being
> >> >accessed from different threads.
> >> >
> >> >On Wed, Nov 2, 2016 at 10:28 AM <[email protected]> wrote:
> >> >
> >> >> Hi,
> >> >>
> >> >> I¹m working primarily using the default scala/spark interpreter. It
> >> >>works
> >> >> great, except when I need to plot something. Is there a way I can
> >>take a
> >> >> scala object or spark data frame I¹ve created in a scala cell and
> >>pass
> >> >>it
> >> >> off to a pyspark cell for plotting?
> >> >>
> >> >> This documentation issue, might be related. I¹d be happy to try to
> >> >> document this once I know how :)
> >> >>
> >> >>
> >> >>
> >>
> >>
> https://urldefense.proofpoint.com/v2/url?u=https-3A__issues.apache.org_ji
> >>
> >>>>ra_browse_TOREE-2D286&d=DQIFaQ&c=nulvIAQnC0yOOjC0e0NVa8TOcyq9jNhjZ156R-
> >>>>JJ
> >>
> >>>>U10&r=CxpqDYMuQy-1uNI-UOyUbaX6BMPCZXH8d8evuCoP_OA&m=R6uBtgqaKfK_uE0gD6e
> >>>>Dj
> >>
> >>>>TZkYHrLjqtC0H66BkHvmVs&s=bRCGzWGCFGO54j-onzac_6v61jjY41QMDA1QQ_qLySQ&e=
> >> >>
> >> >> Thanks!
> >> >>
> >> >> Ian
> >> >>
> >> >>
> >>
> >>
>
>

Reply via email to