+user@hbase 2014-10-15 20:48 GMT+08:00 Fengyun RAO <raofeng...@gmail.com>:
> We use Spark 1.1, and HBase 0.98.1-cdh5.1.0, and need to read and write an > HBase table in Spark program. > > I notice there are: > spark.driver.extraClassPath > spark.executor.extraClassPathproperties to manage extra ClassPath, over > even an deprecated SPARK_CLASSPATH. > > The problem is what classpath or jars should we append? > I can simplely add the whole `hbase classpath`, which is huge, > but this leads to dependencies conflict, e.g. HBase uses guava-12 while > Spark uses guava-14. > >