i have found no way around this. basically this makes SPARK_CLASSPATH
unusable. and the alternative for enabling lzo on a cluster is not
reasonable.

one has to set in spark-defaults.conf:
spark.executor.extraClassPath
/usr/lib/hadoop/lib/hadoop-lzo-cdh4-0.4.15-gplextras.jar
spark.executor.extraLibraryPath
/usr/lib/hadoop/lib/native/Linux-amd64-64:/usr/lib/hadoop/lib/native

and then for every job submitted one has to do using spark-submit (i get
errors if i leave this out):
--driver-class-path /usr/lib/hadoop/lib/hadoop-lzo-cdh4-0.4.15-gplextras.jar
--driver-library-path
/usr/lib/hadoop/lib/native/Linux-amd64-64:/usr/lib/hadoop/lib/native

i think the settings in spark-defaults.conf are fine, but expecting the
user to set the driver-class-path and driver-library-path like that is
cumbersome and requires knowledge of locations of certain files on the
cluster. thats not workable to me.

On Mon, Sep 15, 2014 at 11:16 AM, Koert Kuipers <ko...@tresata.com> wrote:

> in spark 1.1.0 i get this error:
>
> 2014-09-14 23:17:01 ERROR actor.OneForOneStrategy: Found both
> spark.driver.extraClassPath and SPARK_CLASSPATH. Use only the former.
>
> i checked my application. i do not set spark.driver.extraClassPath or
> SPARK_CLASSPATH.
>
> SPARK_CLASSPATH is set in spark-env.sh since the machine is a worker and
> this is how lzo is added to classpath. as a user i cannot modify this.
>
> looking at the logs i see the value of SPARK_CLASSPATH ends up in
> spark.driver.extraClassPath, but thats not done by me. i presume some logic
> in spark-submit does this.
>
>

Reply via email to