I just did that using the RC3 version of Toree for the 0.1.x branch. If you're on master, maybe it doesn't require _jvm_kernel. I just saw that was needed for our RC3.
On Wed, Nov 2, 2016 at 12:12 PM <[email protected]> wrote: > That is not working for me in the release I have 0.1.0… > > %%pyspark > print dir(kernel._jvm_kernel) > > > ['__call__', '__class__', '__delattr__', '__dict__', '__doc__', > '__format__', '__getattribute__', '__hash__', '__init__', '__module__', > '__new__', '__reduce__', '__reduce_ex__', '__repr__', '__setattr__', > '__sizeof__', '__str__', '__subclasshook__', '__weakref__', '_get_args', > 'command_header', 'container', 'converters', 'gateway_client', 'name', > 'pool', 'target_id’] > > > Any ideas? > > Thanks, > Ian > > > > > > > > On 11/2/16, 12:38 PM, "Chip Senkbeil" <[email protected]> wrote: > > >When running Scala as the default interpreter in a notebook: > > > >Cell 1: > >val x = 3 > >kernel.data.put("x", x) > > > >Cell2: > >%%pyspark > >x = kernel._jvm_kernel.data().get("x") > >kernel._jvm_kernel.data().put("x", x + 1) > > > >Cell3: > >println(kernel.data.get("x")) > > > >On Wed, Nov 2, 2016 at 11:27 AM <[email protected]> wrote: > > > >> Thanks Chip, now, I understand how to work with it from the JVM side. > >>Any > >> chance you have a snippet of how to get a value from the map in python? > >> > >> Ian > >> > >> > >> > >> > >> > >> > >> > >> > >> On 11/2/16, 11:39 AM, "Chip Senkbeil" <[email protected]> wrote: > >> > >> >While it isn't supported (we don't test its use in this case), you can > >> >store objects in a shared hashmap under the kernel object that is made > >> >available in each interpreter. The map is exposed as `kernel.data`, but > >> >the > >> >way you access and store data is different per language. > >> > > >> >The signature of the data map on the kernel is `val data: > >>java.util.Map[ > >> >String, Any]` and we use a concurrent hashmap, so it can handle being > >> >accessed from different threads. > >> > > >> >On Wed, Nov 2, 2016 at 10:28 AM <[email protected]> wrote: > >> > > >> >> Hi, > >> >> > >> >> I¹m working primarily using the default scala/spark interpreter. It > >> >>works > >> >> great, except when I need to plot something. Is there a way I can > >>take a > >> >> scala object or spark data frame I¹ve created in a scala cell and > >>pass > >> >>it > >> >> off to a pyspark cell for plotting? > >> >> > >> >> This documentation issue, might be related. I¹d be happy to try to > >> >> document this once I know how :) > >> >> > >> >> > >> >> > >> > >> > https://urldefense.proofpoint.com/v2/url?u=https-3A__issues.apache.org_ji > >> > >>>>ra_browse_TOREE-2D286&d=DQIFaQ&c=nulvIAQnC0yOOjC0e0NVa8TOcyq9jNhjZ156R- > >>>>JJ > >> > >>>>U10&r=CxpqDYMuQy-1uNI-UOyUbaX6BMPCZXH8d8evuCoP_OA&m=R6uBtgqaKfK_uE0gD6e > >>>>Dj > >> > >>>>TZkYHrLjqtC0H66BkHvmVs&s=bRCGzWGCFGO54j-onzac_6v61jjY41QMDA1QQ_qLySQ&e= > >> >> > >> >> Thanks! > >> >> > >> >> Ian > >> >> > >> >> > >> > >> > >
