Hi Balachandar I think I just went through that very same problem and I
solved it with the help of *Moon Soo Lee:*

Here is the solution:
http://stackoverflow.com/questions/35005455/java-npe-when-loading-a-dependency-from-maven-from-within-zeppelin-on-aws-emr

On 25 January 2016 at 14:27, Hyung Sung Shim <hss...@nflabs.com> wrote:

> Hello.
> I also don't work the loadAndDist() method. Maybe it's deprecated.
> I'll check and fix the documents.
>
> using spark-shell, you can run your application as following step.
> 1. remove specified jar configuration in the spark-defaults.conf
> 2. in the spark-home, bin/spark-shell* --jars "YOUR JAR COMMA SPERATED" *
>
> If you share your application code and environments informations(zeppelin
> and spark version you're using, and zeppelin-env.sh etc..), i might help
> you.
>
> Thanks.
>
> <https://gitlab.com/search?group_id=&project_id=769187&scope=issues&search=spark-shell#2-function-define>
>
>
>
>
> 2016-01-25 19:08 GMT+09:00 Balachandar R.A. <balachandar...@gmail.com>:
>
>> Hello,
>>
>> I tried to use z.loadAndDist() but it says
>>
>> console>:17: error: value loadAndDist is not a member of
>> org.apache.zeppelin.spark.dep.DependencyContext
>>
>> Any idea here what this method is for?
>>
>>
>> regards
>> Bala
>>
>> On 25 January 2016 at 15:34, Balachandar R.A. <balachandar...@gmail.com>
>> wrote:
>>
>>> Hello,
>>>
>>> I have run the code in spark-shell successfully but the jar files were
>>> all specified in the config files (spark-defaults.conf). However, I will
>>> not be able to use z.load() in spark-shell. Isn't? I am sorry but I did not
>>> pick up the idea of running using spark-shell. Wail suggestion is to create
>>> a fatJar? I will give it a try but still how do i make sure this fatJar is
>>> accessible to spark executors? ANyway, I will keep you posted on this
>>>
>>> regards
>>> Bala
>>>
>>> On 25 January 2016 at 13:39, Hyung Sung Shim <hss...@nflabs.com> wrote:
>>>
>>>> Hello.
>>>> I think Wail Alkowaileet's comment is possible.
>>>> Balachandar, Could you try to run your application with spark-shell?
>>>>
>>>>
>>>> 2016-01-25 15:45 GMT+09:00 Wail Alkowaileet <wael....@gmail.com>:
>>>>
>>>>> I used z.load in my case and it seems to be working just fine.
>>>>> Can you try spark-shell with your jar file? and see what is the error?
>>>>>
>>>>> I assume the problem that your application requires third-party jars.
>>>>> Therefore, you need to build your app with 'assembly'.
>>>>>
>>>>>
>>>>> On Mon, Jan 25, 2016 at 9:39 AM, Balachandar R.A. <
>>>>> balachandar...@gmail.com> wrote:
>>>>>
>>>>>> Hello Hyung,
>>>>>>
>>>>>> There is nothig I could make out from error log as it is plain
>>>>>> straightforward that classNotFoundException
>>>>>>
>>>>>> On 25 January 2016 at 11:34, Hyung Sung Shim <hss...@nflabs.com>
>>>>>> wrote:
>>>>>>
>>>>>>> It's weird..so Could you send the error log for details?
>>>>>>>
>>>>>>> 2016-01-25 15:00 GMT+09:00 Balachandar R.A. <
>>>>>>> balachandar...@gmail.com>:
>>>>>>>
>>>>>>>> Hi Hyung,
>>>>>>>>
>>>>>>>> Thanks for the response. This I have tried but did not work.
>>>>>>>>
>>>>>>>> regards
>>>>>>>> Bala
>>>>>>>>
>>>>>>>> On 25 January 2016 at 11:27, Hyung Sung Shim <hss...@nflabs.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Hello. Balachandar.
>>>>>>>>> In case of third one that you've tried, It must be first executed
>>>>>>>>> in the notebook.
>>>>>>>>> Could you try restart the zeppelin and run first the "%dep
>>>>>>>>> z.load()" paragraph?
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> 2016-01-25 14:39 GMT+09:00 Balachandar R.A. <
>>>>>>>>> balachandar...@gmail.com>:
>>>>>>>>>
>>>>>>>>>> Hi
>>>>>>>>>>
>>>>>>>>>> Any help would be greatly appreciated :-)
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> ---------- Forwarded message ----------
>>>>>>>>>> From: Balachandar R.A. <balachandar...@gmail.com>
>>>>>>>>>> Date: 21 January 2016 at 14:11
>>>>>>>>>> Subject: Providing third party jar files to spark
>>>>>>>>>> To: users@zeppelin.incubator.apache.org
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Hello
>>>>>>>>>>
>>>>>>>>>> My spark based map tasks needs to access third party jar files. I
>>>>>>>>>> found below options to submit third party jar files to spark 
>>>>>>>>>> interpreter
>>>>>>>>>>
>>>>>>>>>> 1. export SPARK_SUBMIT_OPTIONS=<all the jar files with comma
>>>>>>>>>> seprated> in conf/zeppelin-env.sh
>>>>>>>>>>
>>>>>>>>>> 2. include the statement spark.jars  <all the jar files with
>>>>>>>>>> comma separated> in <spark>?conf/spark-defaults.conf
>>>>>>>>>>
>>>>>>>>>> 3. use the z.load("the location of jar file in the local
>>>>>>>>>> filesystem") in zepelin notebook
>>>>>>>>>>
>>>>>>>>>> I could test the first two and they both works fine. The third
>>>>>>>>>> one does not work. Here is the snippet i use
>>>>>>>>>>
>>>>>>>>>> %dep
>>>>>>>>>> z.reset()
>>>>>>>>>>
>>>>>>>>>> z.load("file:///home/bala/Projects/pocv8.new/mapreduce/build/libs/mapreduce.jar")
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Further, the import of class belongs to the above jar file is
>>>>>>>>>> working when I use the statement import com.....  in zeppelin 
>>>>>>>>>> notebook.
>>>>>>>>>> However, I get the class not found exception in the executor for the 
>>>>>>>>>> same
>>>>>>>>>> class.
>>>>>>>>>>
>>>>>>>>>> Any clue here would help greatly
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> regards
>>>>>>>>>> Bala
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>> --
>>>>>
>>>>> *Regards,*
>>>>> Wail Alkowaileet
>>>>>
>>>>
>>>>
>>>
>>
>


-- 
“Every time you stay out late; every time you sleep in; every time you miss
a workout; every time you don’t give 100% – You make it that much easier
for me to beat you.” - Unknown author

Reply via email to