Hello, I'm using the following version of Spark - 1.0.0+cdh5.1.0+41 (1.cdh5.1.0.p0.27).
I've tried to specify the libraries Spark uses using the following ways - 1) Adding it to spark context 2) Specifying the jar path in a) spark.executor.extraClassPath b) spark.executor.extraLibraryPath 3) Copying the libraries to spark/lib 4) Specifying the path in SPARK_CLASSPATH, even SPARK_LIBRARY_PATH 5) Passing as --jars argument but the spark application is not able to pick up the libraries although, I can see the message "SparkContext: Added JAR file". I get a NoClassDef found error. The only way I've been to able to make it work right now is by merging my application jar with all the library jars. What might be going on? I need it right now to specify hbase-protocol-0.98.1-cdh5.1.0.jar in SPARK_CLASSPATH as mentioned here https://issues.apache.org/jira/browse/HBASE-10877. I'm using spark-submit to submit the job Thanks Ashish