Hi falmeida,

Thanks for the response but I do not want to use
SPARK-SUBMIT-OPTIONS

thanks and regards
Bala
On 26-Jan-2016 8:05 am, "Felipe Almeida" <falmeida1...@gmail.com> wrote:

> Hi Balachandar I think I just went through that very same problem and I
> solved it with the help of *Moon Soo Lee:*
>
> Here is the solution:
> http://stackoverflow.com/questions/35005455/java-npe-when-loading-a-dependency-from-maven-from-within-zeppelin-on-aws-emr
>
> On 25 January 2016 at 14:27, Hyung Sung Shim <hss...@nflabs.com> wrote:
>
>> Hello.
>> I also don't work the loadAndDist() method. Maybe it's deprecated.
>> I'll check and fix the documents.
>>
>> using spark-shell, you can run your application as following step.
>> 1. remove specified jar configuration in the spark-defaults.conf
>> 2. in the spark-home, bin/spark-shell* --jars "YOUR JAR COMMA SPERATED" *
>>
>> If you share your application code and environments informations(zeppelin
>> and spark version you're using, and zeppelin-env.sh etc..), i might help
>> you.
>>
>> Thanks.
>>
>> <https://gitlab.com/search?group_id=&project_id=769187&scope=issues&search=spark-shell#2-function-define>
>>
>>
>>
>>
>> 2016-01-25 19:08 GMT+09:00 Balachandar R.A. <balachandar...@gmail.com>:
>>
>>> Hello,
>>>
>>> I tried to use z.loadAndDist() but it says
>>>
>>> console>:17: error: value loadAndDist is not a member of
>>> org.apache.zeppelin.spark.dep.DependencyContext
>>>
>>> Any idea here what this method is for?
>>>
>>>
>>> regards
>>> Bala
>>>
>>> On 25 January 2016 at 15:34, Balachandar R.A. <balachandar...@gmail.com>
>>> wrote:
>>>
>>>> Hello,
>>>>
>>>> I have run the code in spark-shell successfully but the jar files were
>>>> all specified in the config files (spark-defaults.conf). However, I will
>>>> not be able to use z.load() in spark-shell. Isn't? I am sorry but I did not
>>>> pick up the idea of running using spark-shell. Wail suggestion is to create
>>>> a fatJar? I will give it a try but still how do i make sure this fatJar is
>>>> accessible to spark executors? ANyway, I will keep you posted on this
>>>>
>>>> regards
>>>> Bala
>>>>
>>>> On 25 January 2016 at 13:39, Hyung Sung Shim <hss...@nflabs.com> wrote:
>>>>
>>>>> Hello.
>>>>> I think Wail Alkowaileet's comment is possible.
>>>>> Balachandar, Could you try to run your application with spark-shell?
>>>>>
>>>>>
>>>>> 2016-01-25 15:45 GMT+09:00 Wail Alkowaileet <wael....@gmail.com>:
>>>>>
>>>>>> I used z.load in my case and it seems to be working just fine.
>>>>>> Can you try spark-shell with your jar file? and see what is the error?
>>>>>>
>>>>>> I assume the problem that your application requires third-party jars.
>>>>>> Therefore, you need to build your app with 'assembly'.
>>>>>>
>>>>>>
>>>>>> On Mon, Jan 25, 2016 at 9:39 AM, Balachandar R.A. <
>>>>>> balachandar...@gmail.com> wrote:
>>>>>>
>>>>>>> Hello Hyung,
>>>>>>>
>>>>>>> There is nothig I could make out from error log as it is plain
>>>>>>> straightforward that classNotFoundException
>>>>>>>
>>>>>>> On 25 January 2016 at 11:34, Hyung Sung Shim <hss...@nflabs.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> It's weird..so Could you send the error log for details?
>>>>>>>>
>>>>>>>> 2016-01-25 15:00 GMT+09:00 Balachandar R.A. <
>>>>>>>> balachandar...@gmail.com>:
>>>>>>>>
>>>>>>>>> Hi Hyung,
>>>>>>>>>
>>>>>>>>> Thanks for the response. This I have tried but did not work.
>>>>>>>>>
>>>>>>>>> regards
>>>>>>>>> Bala
>>>>>>>>>
>>>>>>>>> On 25 January 2016 at 11:27, Hyung Sung Shim <hss...@nflabs.com>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> Hello. Balachandar.
>>>>>>>>>> In case of third one that you've tried, It must be first executed
>>>>>>>>>> in the notebook.
>>>>>>>>>> Could you try restart the zeppelin and run first the "%dep
>>>>>>>>>> z.load()" paragraph?
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> 2016-01-25 14:39 GMT+09:00 Balachandar R.A. <
>>>>>>>>>> balachandar...@gmail.com>:
>>>>>>>>>>
>>>>>>>>>>> Hi
>>>>>>>>>>>
>>>>>>>>>>> Any help would be greatly appreciated :-)
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> ---------- Forwarded message ----------
>>>>>>>>>>> From: Balachandar R.A. <balachandar...@gmail.com>
>>>>>>>>>>> Date: 21 January 2016 at 14:11
>>>>>>>>>>> Subject: Providing third party jar files to spark
>>>>>>>>>>> To: users@zeppelin.incubator.apache.org
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Hello
>>>>>>>>>>>
>>>>>>>>>>> My spark based map tasks needs to access third party jar files.
>>>>>>>>>>> I found below options to submit third party jar files to spark 
>>>>>>>>>>> interpreter
>>>>>>>>>>>
>>>>>>>>>>> 1. export SPARK_SUBMIT_OPTIONS=<all the jar files with comma
>>>>>>>>>>> seprated> in conf/zeppelin-env.sh
>>>>>>>>>>>
>>>>>>>>>>> 2. include the statement spark.jars  <all the jar files with
>>>>>>>>>>> comma separated> in <spark>?conf/spark-defaults.conf
>>>>>>>>>>>
>>>>>>>>>>> 3. use the z.load("the location of jar file in the local
>>>>>>>>>>> filesystem") in zepelin notebook
>>>>>>>>>>>
>>>>>>>>>>> I could test the first two and they both works fine. The third
>>>>>>>>>>> one does not work. Here is the snippet i use
>>>>>>>>>>>
>>>>>>>>>>> %dep
>>>>>>>>>>> z.reset()
>>>>>>>>>>>
>>>>>>>>>>> z.load("file:///home/bala/Projects/pocv8.new/mapreduce/build/libs/mapreduce.jar")
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Further, the import of class belongs to the above jar file is
>>>>>>>>>>> working when I use the statement import com.....  in zeppelin 
>>>>>>>>>>> notebook.
>>>>>>>>>>> However, I get the class not found exception in the executor for 
>>>>>>>>>>> the same
>>>>>>>>>>> class.
>>>>>>>>>>>
>>>>>>>>>>> Any clue here would help greatly
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> regards
>>>>>>>>>>> Bala
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>>
>>>>>> *Regards,*
>>>>>> Wail Alkowaileet
>>>>>>
>>>>>
>>>>>
>>>>
>>>
>>
>
>
> --
> “Every time you stay out late; every time you sleep in; every time you
> miss a workout; every time you don’t give 100% – You make it that much
> easier for me to beat you.” - Unknown author
>

Reply via email to