Hi I believe this trick will work because I ran it through spark-submit using --jars option. Will give it a try though
Bala On 25-Jan-2016 9:57 pm, "Hyung Sung Shim" <hss...@nflabs.com> wrote: > Hello. > I also don't work the loadAndDist() method. Maybe it's deprecated. > I'll check and fix the documents. > > using spark-shell, you can run your application as following step. > 1. remove specified jar configuration in the spark-defaults.conf > 2. in the spark-home, bin/spark-shell* --jars "YOUR JAR COMMA SPERATED" * > > If you share your application code and environments informations(zeppelin > and spark version you're using, and zeppelin-env.sh etc..), i might help > you. > > Thanks. > > <https://gitlab.com/search?group_id=&project_id=769187&scope=issues&search=spark-shell#2-function-define> > > > > > 2016-01-25 19:08 GMT+09:00 Balachandar R.A. <balachandar...@gmail.com>: > >> Hello, >> >> I tried to use z.loadAndDist() but it says >> >> console>:17: error: value loadAndDist is not a member of >> org.apache.zeppelin.spark.dep.DependencyContext >> >> Any idea here what this method is for? >> >> >> regards >> Bala >> >> On 25 January 2016 at 15:34, Balachandar R.A. <balachandar...@gmail.com> >> wrote: >> >>> Hello, >>> >>> I have run the code in spark-shell successfully but the jar files were >>> all specified in the config files (spark-defaults.conf). However, I will >>> not be able to use z.load() in spark-shell. Isn't? I am sorry but I did not >>> pick up the idea of running using spark-shell. Wail suggestion is to create >>> a fatJar? I will give it a try but still how do i make sure this fatJar is >>> accessible to spark executors? ANyway, I will keep you posted on this >>> >>> regards >>> Bala >>> >>> On 25 January 2016 at 13:39, Hyung Sung Shim <hss...@nflabs.com> wrote: >>> >>>> Hello. >>>> I think Wail Alkowaileet's comment is possible. >>>> Balachandar, Could you try to run your application with spark-shell? >>>> >>>> >>>> 2016-01-25 15:45 GMT+09:00 Wail Alkowaileet <wael....@gmail.com>: >>>> >>>>> I used z.load in my case and it seems to be working just fine. >>>>> Can you try spark-shell with your jar file? and see what is the error? >>>>> >>>>> I assume the problem that your application requires third-party jars. >>>>> Therefore, you need to build your app with 'assembly'. >>>>> >>>>> >>>>> On Mon, Jan 25, 2016 at 9:39 AM, Balachandar R.A. < >>>>> balachandar...@gmail.com> wrote: >>>>> >>>>>> Hello Hyung, >>>>>> >>>>>> There is nothig I could make out from error log as it is plain >>>>>> straightforward that classNotFoundException >>>>>> >>>>>> On 25 January 2016 at 11:34, Hyung Sung Shim <hss...@nflabs.com> >>>>>> wrote: >>>>>> >>>>>>> It's weird..so Could you send the error log for details? >>>>>>> >>>>>>> 2016-01-25 15:00 GMT+09:00 Balachandar R.A. < >>>>>>> balachandar...@gmail.com>: >>>>>>> >>>>>>>> Hi Hyung, >>>>>>>> >>>>>>>> Thanks for the response. This I have tried but did not work. >>>>>>>> >>>>>>>> regards >>>>>>>> Bala >>>>>>>> >>>>>>>> On 25 January 2016 at 11:27, Hyung Sung Shim <hss...@nflabs.com> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Hello. Balachandar. >>>>>>>>> In case of third one that you've tried, It must be first executed >>>>>>>>> in the notebook. >>>>>>>>> Could you try restart the zeppelin and run first the "%dep >>>>>>>>> z.load()" paragraph? >>>>>>>>> >>>>>>>>> >>>>>>>>> 2016-01-25 14:39 GMT+09:00 Balachandar R.A. < >>>>>>>>> balachandar...@gmail.com>: >>>>>>>>> >>>>>>>>>> Hi >>>>>>>>>> >>>>>>>>>> Any help would be greatly appreciated :-) >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> ---------- Forwarded message ---------- >>>>>>>>>> From: Balachandar R.A. <balachandar...@gmail.com> >>>>>>>>>> Date: 21 January 2016 at 14:11 >>>>>>>>>> Subject: Providing third party jar files to spark >>>>>>>>>> To: users@zeppelin.incubator.apache.org >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Hello >>>>>>>>>> >>>>>>>>>> My spark based map tasks needs to access third party jar files. I >>>>>>>>>> found below options to submit third party jar files to spark >>>>>>>>>> interpreter >>>>>>>>>> >>>>>>>>>> 1. export SPARK_SUBMIT_OPTIONS=<all the jar files with comma >>>>>>>>>> seprated> in conf/zeppelin-env.sh >>>>>>>>>> >>>>>>>>>> 2. include the statement spark.jars <all the jar files with >>>>>>>>>> comma separated> in <spark>?conf/spark-defaults.conf >>>>>>>>>> >>>>>>>>>> 3. use the z.load("the location of jar file in the local >>>>>>>>>> filesystem") in zepelin notebook >>>>>>>>>> >>>>>>>>>> I could test the first two and they both works fine. The third >>>>>>>>>> one does not work. Here is the snippet i use >>>>>>>>>> >>>>>>>>>> %dep >>>>>>>>>> z.reset() >>>>>>>>>> >>>>>>>>>> z.load("file:///home/bala/Projects/pocv8.new/mapreduce/build/libs/mapreduce.jar") >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Further, the import of class belongs to the above jar file is >>>>>>>>>> working when I use the statement import com..... in zeppelin >>>>>>>>>> notebook. >>>>>>>>>> However, I get the class not found exception in the executor for the >>>>>>>>>> same >>>>>>>>>> class. >>>>>>>>>> >>>>>>>>>> Any clue here would help greatly >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> regards >>>>>>>>>> Bala >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>>> >>>>> -- >>>>> >>>>> *Regards,* >>>>> Wail Alkowaileet >>>>> >>>> >>>> >>> >> >