yes that's a solid approach. For my personal julia - java integrations I also run the JVM in a separate process.
On Wednesday, April 15, 2015 at 9:30:28 PM UTC+2, wil...@gmail.com wrote: > > 1) simply wrap the Spark java API via JavaCall. This is the low level >> approach. BTW I've experimented with javaCall and found it was unstable & >> also lacking functionality (e.g. there's no way to shutdown the jvm or >> create a pool of JVM analogous to DB connections) so that might need some >> work before trying the Spark integration. >> > > Using JavaCall is not an option, especially when JVM became close-sourced, > see https://github.com/aviks/JavaCall.jl/issues/7. > > Python bindings are done through Py4J, which is RPC to JVM. If you look at > the sparkR <https://github.com/apache/spark/tree/master/R>, it is done in > a same way. sparkR uses a RPC interface to communicate with a Netty-based > Spark JVM backend that translates R calls into JVM calls, keeps > SparkContext on a JVM side, and ships serialized data to/from R. > > So it is just a matter of writing Julia RPC to JVM and wrapping necessary > Spark methods in a Julia friendly way. >