Trying again…anyone know how to make this work? > On Jul 9, 2018, at 3:45 PM, Mohit Jaggi <mohitja...@gmail.com> wrote: > > Folks, > I am writing some Scala/Java code and want it to be usable from pyspark. > > For example: > class MyStuff(addend: Int) { > def myMapFunction(x: Int) = x + addend > } > > I want to call it from pyspark as: > > df = ... > mystuff = sc._jvm.MyStuff(5) > df[‘x’].map(lambda x: mystuff.myMapFunction(x)) > > How can I do this? > > Mohit. > >
--------------------------------------------------------------------- To unsubscribe e-mail: dev-unsubscr...@spark.apache.org