This is an automated email from the ASF dual-hosted git repository. lresende pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/bahir.git
commit 843cd71a2c1b14e6babe9538cd0e9982f8bab836 Author: Luciano Resende <[email protected]> AuthorDate: Tue Jun 11 13:47:17 2019 +0200 Fix warnings around depecrated usage of shouldRunTest --- .../org/apache/bahir/cloudant/ClientSparkFunSuite.scala | 2 +- .../apache/bahir/cloudant/CloudantAllDocsDFSuite.scala | 15 ++++++++------- .../apache/bahir/cloudant/CloudantChangesDFSuite.scala | 17 +++++++++-------- .../org/apache/bahir/cloudant/CloudantOptionSuite.scala | 15 ++++++++------- .../apache/bahir/cloudant/CloudantSparkSQLSuite.scala | 4 ++-- .../spark/streaming/pubnub/PubNubStreamSuite.scala | 6 +++--- .../spark/streaming/pubsub/PubsubStreamSuite.scala | 2 +- .../spark/streaming/twitter/TwitterStreamSuite.scala | 2 +- 8 files changed, 33 insertions(+), 30 deletions(-) diff --git a/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/ClientSparkFunSuite.scala b/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/ClientSparkFunSuite.scala index ed14f17..6946803 100644 --- a/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/ClientSparkFunSuite.scala +++ b/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/ClientSparkFunSuite.scala @@ -40,7 +40,7 @@ class ClientSparkFunSuite extends ConditionalSparkFunSuite with BeforeAndAfter { var spark: SparkSession = _ override def beforeAll() { - runIf(TestUtils.shouldRunTest) { + runIf(() => TestUtils.shouldRunTest()) { tempDir.mkdirs() tempDir.deleteOnExit() setupClient() diff --git a/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantAllDocsDFSuite.scala b/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantAllDocsDFSuite.scala index c0389d8..e012336 100644 --- a/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantAllDocsDFSuite.scala +++ b/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantAllDocsDFSuite.scala @@ -35,7 +35,7 @@ class CloudantAllDocsDFSuite extends ClientSparkFunSuite { .getOrCreate() } - testIf("load and save data from Cloudant database", TestUtils.shouldRunTest) { + testIf("load and save data from Cloudant database", () => TestUtils.shouldRunTest()) { // Loading data from Cloudant db val df = spark.read.format("org.apache.bahir.cloudant").load("n_flight") // Caching df in memory to speed computations @@ -46,7 +46,7 @@ class CloudantAllDocsDFSuite extends ClientSparkFunSuite { assert(df.count() == 100) } - testIf("load and count data from Cloudant search index", TestUtils.shouldRunTest) { + testIf("load and count data from Cloudant search index", () => TestUtils.shouldRunTest()) { val df = spark.read.format("org.apache.bahir.cloudant") .option("index", "_design/view/_search/n_flights").load("n_flight") val total = df.filter(df("flightSegmentId") >"AA14") @@ -55,7 +55,7 @@ class CloudantAllDocsDFSuite extends ClientSparkFunSuite { assert(total == 89) } - testIf("load data and count rows in filtered dataframe", TestUtils.shouldRunTest) { + testIf("load data and count rows in filtered dataframe", () => TestUtils.shouldRunTest()) { // Loading data from Cloudant db val df = spark.read.format("org.apache.bahir.cloudant") .load("n_airportcodemapping") @@ -64,7 +64,7 @@ class CloudantAllDocsDFSuite extends ClientSparkFunSuite { } // save data to Cloudant test - testIf("save filtered dataframe to database", TestUtils.shouldRunTest) { + testIf("save filtered dataframe to database", () => TestUtils.shouldRunTest()) { val df = spark.read.format("org.apache.bahir.cloudant").load("n_flight") // Saving data frame with filter to Cloudant db @@ -81,7 +81,8 @@ class CloudantAllDocsDFSuite extends ClientSparkFunSuite { } // createDBOnSave option test - testIf("save dataframe to database using createDBOnSave=true option", TestUtils.shouldRunTest) { + testIf("save dataframe to database using createDBOnSave=true option", + () => TestUtils.shouldRunTest()) { val df = spark.read.format("org.apache.bahir.cloudant") .load("n_airportcodemapping") @@ -106,13 +107,13 @@ class CloudantAllDocsDFSuite extends ClientSparkFunSuite { } // view option tests - testIf("load and count data from view", TestUtils.shouldRunTest) { + testIf("load and count data from view", () => TestUtils.shouldRunTest()) { val df = spark.read.format("org.apache.bahir.cloudant") .option("view", "_design/view/_view/AA0").load("n_flight") assert(df.count() == 1) } - testIf("load data from view with MapReduce function", TestUtils.shouldRunTest) { + testIf("load data from view with MapReduce function", () => TestUtils.shouldRunTest()) { val df = spark.read.format("org.apache.bahir.cloudant") .option("view", "_design/view/_view/AAreduce?reduce=true") .load("n_flight") diff --git a/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantChangesDFSuite.scala b/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantChangesDFSuite.scala index dd5d508..fe92ef2 100644 --- a/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantChangesDFSuite.scala +++ b/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantChangesDFSuite.scala @@ -47,7 +47,7 @@ class CloudantChangesDFSuite extends ClientSparkFunSuite { assert(df.count() == 100) } - testIf("load and count data from Cloudant search index", TestUtils.shouldRunTest) { + testIf("load and count data from Cloudant search index", () => TestUtils.shouldRunTest()) { val df = spark.read.format("org.apache.bahir.cloudant") .option("index", "_design/view/_search/n_flights").load("n_flight") val total = df.filter(df("flightSegmentId") >"AA14") @@ -56,7 +56,7 @@ class CloudantChangesDFSuite extends ClientSparkFunSuite { assert(total == 89) } - testIf("load data and verify deleted doc is not in results", TestUtils.shouldRunTest) { + testIf("load data and verify deleted doc is not in results", () => TestUtils.shouldRunTest()) { val db = client.database("n_flight", false) // delete a saved doc to verify it's not included when loading data db.remove(deletedDoc.get("_id").getAsString, deletedDoc.get("_rev").getAsString) @@ -68,7 +68,7 @@ class CloudantChangesDFSuite extends ClientSparkFunSuite { assert(!df.columns.contains("_deleted")) } - testIf("load data and count rows in filtered dataframe", TestUtils.shouldRunTest) { + testIf("load data and count rows in filtered dataframe", () => TestUtils.shouldRunTest()) { // Loading data from Cloudant db val df = spark.read.format("org.apache.bahir.cloudant") .load("n_airportcodemapping") @@ -77,7 +77,7 @@ class CloudantChangesDFSuite extends ClientSparkFunSuite { } // save data to Cloudant test - testIf("save filtered dataframe to database", TestUtils.shouldRunTest) { + testIf("save filtered dataframe to database", () => TestUtils.shouldRunTest()) { val df = spark.read.format("org.apache.bahir.cloudant").load("n_flight") // Saving data frame with filter to Cloudant db @@ -94,7 +94,8 @@ class CloudantChangesDFSuite extends ClientSparkFunSuite { } // createDBOnSave option test - testIf("save dataframe to database using createDBOnSave=true option", TestUtils.shouldRunTest) { + testIf("save dataframe to database using createDBOnSave=true option", + () => TestUtils.shouldRunTest()) { val df = spark.read.format("org.apache.bahir.cloudant") .load("n_airportcodemapping") @@ -124,13 +125,13 @@ class CloudantChangesDFSuite extends ClientSparkFunSuite { } // view option tests - testIf("load and count data from view", TestUtils.shouldRunTest) { + testIf("load and count data from view", () => TestUtils.shouldRunTest()) { val df = spark.read.format("org.apache.bahir.cloudant") .option("view", "_design/view/_view/AA0").load("n_flight") assert(df.count() == 1) } - testIf("load data from view with MapReduce function", TestUtils.shouldRunTest) { + testIf("load data from view with MapReduce function", () => TestUtils.shouldRunTest()) { val df = spark.read.format("org.apache.bahir.cloudant") .option("view", "_design/view/_view/AAreduce?reduce=true") .load("n_flight") @@ -138,7 +139,7 @@ class CloudantChangesDFSuite extends ClientSparkFunSuite { } testIf("load data and verify total count of selector, filter, and view option", - TestUtils.shouldRunTest) { + () => TestUtils.shouldRunTest()) { val df = spark.read.format("org.apache.bahir.cloudant") .option("selector", "{\"flightSegmentId\": {\"$eq\": \"AA202\"}}") .load("n_flight") diff --git a/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantOptionSuite.scala b/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantOptionSuite.scala index f5df759..5fb5d53 100644 --- a/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantOptionSuite.scala +++ b/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantOptionSuite.scala @@ -29,7 +29,7 @@ class CloudantOptionSuite extends ClientSparkFunSuite with BeforeAndAfter { spark.close() } - testIf("invalid api receiver option throws an error message", TestUtils.shouldRunTest) { + testIf("invalid api receiver option throws an error message", () => TestUtils.shouldRunTest()) { spark = SparkSession.builder().config(conf) .config("cloudant.protocol", TestUtils.getProtocol) .config("cloudant.host", TestUtils.getHost) @@ -45,7 +45,7 @@ class CloudantOptionSuite extends ClientSparkFunSuite with BeforeAndAfter { s"is invalid. Please supply the valid option '_all_docs' or '_changes'.") } - testIf("empty username option throws an error message", TestUtils.shouldRunTest) { + testIf("empty username option throws an error message", () => TestUtils.shouldRunTest()) { spark = SparkSession.builder().config(conf) .config("cloudant.protocol", TestUtils.getProtocol) .config("cloudant.host", TestUtils.getHost) @@ -60,7 +60,7 @@ class CloudantOptionSuite extends ClientSparkFunSuite with BeforeAndAfter { s"is empty. Please supply the required value.") } - testIf("empty password option throws an error message", TestUtils.shouldRunTest) { + testIf("empty password option throws an error message", () => TestUtils.shouldRunTest()) { spark = SparkSession.builder().config(conf) .config("cloudant.protocol", TestUtils.getProtocol) .config("cloudant.host", TestUtils.getHost) @@ -75,7 +75,7 @@ class CloudantOptionSuite extends ClientSparkFunSuite with BeforeAndAfter { s"is empty. Please supply the required value.") } - testIf("empty databaseName throws an error message", TestUtils.shouldRunTest) { + testIf("empty databaseName throws an error message", () => TestUtils.shouldRunTest()) { spark = SparkSession.builder().config(conf) .config("cloudant.protocol", TestUtils.getProtocol) .config("cloudant.host", TestUtils.getHost) @@ -91,7 +91,7 @@ class CloudantOptionSuite extends ClientSparkFunSuite with BeforeAndAfter { } testIf("incorrect password throws an error message for changes receiver", - TestUtils.shouldRunTest) { + () => TestUtils.shouldRunTest()) { spark = SparkSession.builder().config(conf) .config("cloudant.protocol", TestUtils.getProtocol) .config("cloudant.host", TestUtils.getHost) @@ -108,7 +108,8 @@ class CloudantOptionSuite extends ClientSparkFunSuite with BeforeAndAfter { "\"reason\":\"Name or password is incorrect.\"}") } - testIf("string with valid value for cloudant.numberOfRetries option", TestUtils.shouldRunTest) { + testIf("string with valid value for cloudant.numberOfRetries option", + () => TestUtils.shouldRunTest()) { spark = SparkSession.builder().config(conf) .config("cloudant.protocol", TestUtils.getProtocol) .config("cloudant.host", TestUtils.getHost) @@ -122,7 +123,7 @@ class CloudantOptionSuite extends ClientSparkFunSuite with BeforeAndAfter { } testIf("invalid value for cloudant.numberOfRetries option throws an error message", - TestUtils.shouldRunTest) { + () => TestUtils.shouldRunTest()) { spark = SparkSession.builder().config(conf) .config("cloudant.protocol", TestUtils.getProtocol) .config("cloudant.host", TestUtils.getHost) diff --git a/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantSparkSQLSuite.scala b/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantSparkSQLSuite.scala index 6736da2..9100130 100644 --- a/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantSparkSQLSuite.scala +++ b/sql-cloudant/src/test/scala/org/apache/bahir/cloudant/CloudantSparkSQLSuite.scala @@ -39,7 +39,7 @@ class CloudantSparkSQLSuite extends ClientSparkFunSuite { } testIf("verify results from temp view of database n_airportcodemapping", - TestUtils.shouldRunTest) { + () => TestUtils.shouldRunTest()) { // create a temp table from Cloudant db and query it using sql syntax val sparkSql = spark.sql( s""" @@ -67,7 +67,7 @@ class CloudantSparkSQLSuite extends ClientSparkFunSuite { assert(df2count == airportData.count()) } - testIf("verify results from temp view of index in n_flight", TestUtils.shouldRunTest) { + testIf("verify results from temp view of index in n_flight", () => TestUtils.shouldRunTest()) { // create a temp table from Cloudant index and query it using sql syntax val sparkSql = spark.sql( s""" diff --git a/streaming-pubnub/src/test/scala/org/apache/spark/streaming/pubnub/PubNubStreamSuite.scala b/streaming-pubnub/src/test/scala/org/apache/spark/streaming/pubnub/PubNubStreamSuite.scala index 91011fb..4990b1f 100644 --- a/streaming-pubnub/src/test/scala/org/apache/spark/streaming/pubnub/PubNubStreamSuite.scala +++ b/streaming-pubnub/src/test/scala/org/apache/spark/streaming/pubnub/PubNubStreamSuite.scala @@ -65,7 +65,7 @@ class PubNubStreamSuite extends ConditionalSparkFunSuite with Eventually with Be } } - testIf("Stream receives messages", shouldRunTest) { + testIf("Stream receives messages", () => PubNubStreamSuite.this.shouldRunTest()) { val nbOfMsg = 5 var publishedMessages: List[JsonObject] = List() @volatile var receivedMessages: Set[SparkPubNubMessage] = Set() @@ -101,7 +101,7 @@ class PubNubStreamSuite extends ConditionalSparkFunSuite with Eventually with Be } } - testIf("Message filtering", shouldRunTest) { + testIf("Message filtering", () => PubNubStreamSuite.this.shouldRunTest()) { val config = new PNConfiguration() config.setSubscribeKey(subscribeKey) config.setPublishKey(publishKey) @@ -134,7 +134,7 @@ class PubNubStreamSuite extends ConditionalSparkFunSuite with Eventually with Be } } - testIf("Test time token", shouldRunTest) { + testIf("Test time token", () => PubNubStreamSuite.this.shouldRunTest()) { val config = new PNConfiguration() config.setSubscribeKey(subscribeKey) config.setPublishKey(publishKey) diff --git a/streaming-pubsub/src/test/scala/org/apache/spark/streaming/pubsub/PubsubStreamSuite.scala b/streaming-pubsub/src/test/scala/org/apache/spark/streaming/pubsub/PubsubStreamSuite.scala index b02f21d..8a67038 100644 --- a/streaming-pubsub/src/test/scala/org/apache/spark/streaming/pubsub/PubsubStreamSuite.scala +++ b/streaming-pubsub/src/test/scala/org/apache/spark/streaming/pubsub/PubsubStreamSuite.scala @@ -52,7 +52,7 @@ class PubsubStreamSuite extends ConditionalSparkFunSuite with Eventually with Be private var subForCreateFullName: String = _ override def beforeAll(): Unit = { - runIf(PubsubTestUtils.shouldRunTest) { + runIf(() => PubsubTestUtils.shouldRunTest()) { pubsubTestUtils = new PubsubTestUtils topicFullName = pubsubTestUtils.getFullTopicPath(topicName) subscriptionFullName = pubsubTestUtils.getFullSubscriptionPath(subscriptionName) diff --git a/streaming-twitter/src/test/scala/org/apache/spark/streaming/twitter/TwitterStreamSuite.scala b/streaming-twitter/src/test/scala/org/apache/spark/streaming/twitter/TwitterStreamSuite.scala index 7d7886d..77c6625 100644 --- a/streaming-twitter/src/test/scala/org/apache/spark/streaming/twitter/TwitterStreamSuite.scala +++ b/streaming-twitter/src/test/scala/org/apache/spark/streaming/twitter/TwitterStreamSuite.scala @@ -71,7 +71,7 @@ class TwitterStreamSuite extends ConditionalSparkFunSuite ssc, Some(authorization), Some(query), StorageLevel.MEMORY_AND_DISK_SER_2) } - testIf("messages received", shouldRunTest) { + testIf("messages received", () => TwitterStreamSuite.this.shouldRunTest()) { val userId = TwitterFactory.getSingleton.updateStatus( UUID.randomUUID().toString ).getUser.getId
