Repository: bahir Updated Branches: refs/heads/master eb8d6f6ee -> c5263df23
[BAHIR-123] Upgrade to play-json 2.6.6 Fixed breaking API changes between play-json 2.5.x and 2.6.x in sql-cloudant by replacing deprecated methods. Closes #50 Project: http://git-wip-us.apache.org/repos/asf/bahir/repo Commit: http://git-wip-us.apache.org/repos/asf/bahir/commit/c5263df2 Tree: http://git-wip-us.apache.org/repos/asf/bahir/tree/c5263df2 Diff: http://git-wip-us.apache.org/repos/asf/bahir/diff/c5263df2 Branch: refs/heads/master Commit: c5263df233b53a603883c1a5c4aa6c652f0e7fab Parents: eb8d6f6 Author: Esteban Laver <emla...@us.ibm.com> Authored: Wed Oct 4 15:58:50 2017 -0400 Committer: Esteban Laver <emla...@us.ibm.com> Committed: Wed Oct 4 15:58:50 2017 -0400 ---------------------------------------------------------------------- pom.xml | 2 +- sql-cloudant/pom.xml | 11 ++++++++++- .../bahir/cloudant/ClientSparkFunSuite.scala | 2 +- .../bahir/cloudant/CloudantAllDocsDFSuite.scala | 15 +++++++++++---- .../bahir/cloudant/CloudantChangesDFSuite.scala | 19 +++++++++++++------ 5 files changed, 36 insertions(+), 13 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/bahir/blob/c5263df2/pom.xml ---------------------------------------------------------------------- diff --git a/pom.xml b/pom.xml index 37b764e..1ae76a2 100644 --- a/pom.xml +++ b/pom.xml @@ -296,7 +296,7 @@ <dependency> <groupId>com.typesafe.play</groupId> <artifactId>play-json_${scala.binary.version}</artifactId> - <version>2.5.9</version> + <version>2.6.6</version> </dependency> <dependency> http://git-wip-us.apache.org/repos/asf/bahir/blob/c5263df2/sql-cloudant/pom.xml ---------------------------------------------------------------------- diff --git a/sql-cloudant/pom.xml b/sql-cloudant/pom.xml index 45c0c42..312e2b0 100644 --- a/sql-cloudant/pom.xml +++ b/sql-cloudant/pom.xml @@ -38,7 +38,16 @@ <dependency> <groupId>com.typesafe.play</groupId> <artifactId>play-json_${scala.binary.version}</artifactId> - <!--version>2.4.8</version--> + </dependency> + <dependency> + <groupId>com.typesafe</groupId> + <artifactId>config</artifactId> + <version>1.3.1</version> + </dependency> + <dependency> + <groupId>org.joda</groupId> + <artifactId>joda-convert</artifactId> + <version>1.8.1</version> </dependency> <dependency> <groupId>com.fasterxml.jackson.core</groupId> http://git-wip-us.apache.org/repos/asf/bahir/blob/c5263df2/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/ClientSparkFunSuite.scala ---------------------------------------------------------------------- diff --git a/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/ClientSparkFunSuite.scala b/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/ClientSparkFunSuite.scala index 6bc5c2a..94ceadf 100644 --- a/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/ClientSparkFunSuite.scala +++ b/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/ClientSparkFunSuite.scala @@ -107,7 +107,7 @@ class ClientSparkFunSuite extends SparkFunSuite with BeforeAndAfter { def deleteTestDbs() { for (db: String <- TestUtils.testDatabasesList) { - client.deleteDB(db) + deleteTestDb(db) } } http://git-wip-us.apache.org/repos/asf/bahir/blob/c5263df2/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantAllDocsDFSuite.scala ---------------------------------------------------------------------- diff --git a/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantAllDocsDFSuite.scala b/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantAllDocsDFSuite.scala index a50b4a9..982bbf9 100644 --- a/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantAllDocsDFSuite.scala +++ b/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantAllDocsDFSuite.scala @@ -17,6 +17,8 @@ package org.apache.bahir.cloudant +import scala.util.Try + import org.apache.spark.sql.SparkSession class CloudantAllDocsDFSuite extends ClientSparkFunSuite { @@ -82,20 +84,25 @@ class CloudantAllDocsDFSuite extends ClientSparkFunSuite { val df = spark.read.format("org.apache.bahir.cloudant") .load("n_airportcodemapping") + val saveDfToDb = "airportcodemapping_df" + + // If 'airportcodemapping_df' exists, delete it. + Try { + client.deleteDB(saveDfToDb) + } + // Saving dataframe to Cloudant db // to create a Cloudant db during save set the option createDBOnSave=true val df2 = df.filter(df("_id") >= "CAA") .select("_id", "airportName") .write.format("org.apache.bahir.cloudant") .option("createDBOnSave", "true") - .save("airportcodemapping_df") + .save(saveDfToDb) val dfAirport = spark.read.format("org.apache.bahir.cloudant") - .load("airportcodemapping_df") + .load(saveDfToDb) assert(dfAirport.count() == 13) - - deleteTestDb("airportcodemapping_df") } // view option tests http://git-wip-us.apache.org/repos/asf/bahir/blob/c5263df2/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantChangesDFSuite.scala ---------------------------------------------------------------------- diff --git a/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantChangesDFSuite.scala b/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantChangesDFSuite.scala index 2f74c12..da51d9f 100644 --- a/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantChangesDFSuite.scala +++ b/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantChangesDFSuite.scala @@ -17,7 +17,9 @@ package org.apache.bahir.cloudant -import play.libs.Json +import scala.util.Try + +import play.api.libs.json._ import org.apache.spark.sql.SparkSession @@ -63,7 +65,7 @@ class CloudantChangesDFSuite extends ClientSparkFunSuite { // Find then delete a doc to verify it's not included when loading data val doc = db.find("003bd483-9f98-4203-afdd-c539a4f38d21") val json = try { Json.parse(doc) } finally { doc.close() } - db.remove(json.get("_id").textValue(), json.get("_rev").textValue()) + db.remove((json \ "_id").get.as[String], (json \ "_rev").get.as[String]) val df = spark.read.format("org.apache.bahir.cloudant").load("n_flight") // all docs in database minus the design doc and _deleted=true doc @@ -102,20 +104,25 @@ class CloudantChangesDFSuite extends ClientSparkFunSuite { val df = spark.read.format("org.apache.bahir.cloudant") .load("n_airportcodemapping") + val saveDfToDb = "airportcodemapping_df" + + // If 'airportcodemapping_df' exists, delete it. + Try { + client.deleteDB(saveDfToDb) + } + // Saving dataframe to Cloudant db // to create a Cloudant db during save set the option createDBOnSave=true val df2 = df.filter(df("_id") >= "CAA") .select("_id", "airportName") .write.format("org.apache.bahir.cloudant") .option("createDBOnSave", "true") - .save("airportcodemapping_df") + .save(saveDfToDb) val dfAirport = spark.read.format("org.apache.bahir.cloudant") - .load("airportcodemapping_df") + .load(saveDfToDb) assert(dfAirport.count() == 13) - - deleteTestDb("airportcodemapping_df") } // view option tests