HBASE-15282 Bump hbase-spark to use Spark 1.6.0
Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/f352f3c3 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/f352f3c3 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/f352f3c3 Branch: refs/heads/hbase-12439 Commit: f352f3c3717e1ebb129b8fb476cff6c8daa4ac06 Parents: d2ba875 Author: Jonathan M Hsieh <[email protected]> Authored: Thu Feb 18 17:31:42 2016 -0800 Committer: Jonathan M Hsieh <[email protected]> Committed: Thu Feb 18 17:31:42 2016 -0800 ---------------------------------------------------------------------- hbase-spark/pom.xml | 2 +- .../scala/org/apache/hadoop/hbase/spark/DefaultSourceSuite.scala | 4 ++-- 2 files changed, 3 insertions(+), 3 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hbase/blob/f352f3c3/hbase-spark/pom.xml ---------------------------------------------------------------------- diff --git a/hbase-spark/pom.xml b/hbase-spark/pom.xml index 251ea59..7c7590e 100644 --- a/hbase-spark/pom.xml +++ b/hbase-spark/pom.xml @@ -37,7 +37,7 @@ <name>Apache HBase - Spark</name> <properties> - <spark.version>1.3.0</spark.version> + <spark.version>1.6.0</spark.version> <scala.version>2.10.4</scala.version> <scala.binary.version>2.10</scala.binary.version> <surefire.skipSecondPart>true</surefire.skipSecondPart> http://git-wip-us.apache.org/repos/asf/hbase/blob/f352f3c3/hbase-spark/src/test/scala/org/apache/hadoop/hbase/spark/DefaultSourceSuite.scala ---------------------------------------------------------------------- diff --git a/hbase-spark/src/test/scala/org/apache/hadoop/hbase/spark/DefaultSourceSuite.scala b/hbase-spark/src/test/scala/org/apache/hadoop/hbase/spark/DefaultSourceSuite.scala index 30ddfc4..04dd9ba 100644 --- a/hbase-spark/src/test/scala/org/apache/hadoop/hbase/spark/DefaultSourceSuite.scala +++ b/hbase-spark/src/test/scala/org/apache/hadoop/hbase/spark/DefaultSourceSuite.scala @@ -20,7 +20,7 @@ package org.apache.hadoop.hbase.spark import org.apache.hadoop.hbase.client.{Put, ConnectionFactory} import org.apache.hadoop.hbase.spark.datasources.HBaseSparkConf import org.apache.hadoop.hbase.util.Bytes -import org.apache.hadoop.hbase.{TableNotFoundException, TableName, HBaseTestingUtility} +import org.apache.hadoop.hbase.{TableName, HBaseTestingUtility} import org.apache.spark.sql.{DataFrame, SQLContext} import org.apache.spark.{SparkConf, SparkContext, Logging} import org.scalatest.{BeforeAndAfterAll, BeforeAndAfterEach, FunSuite} @@ -514,7 +514,7 @@ BeforeAndAfterEach with BeforeAndAfterAll with Logging { test("Test table that doesn't exist") { - intercept[TableNotFoundException] { + intercept[Exception] { df = sqlContext.load("org.apache.hadoop.hbase.spark", Map("hbase.columns.mapping" -> "KEY_FIELD STRING :key, A_FIELD STRING c:a, B_FIELD STRING c:b,",
