http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/internal/rdd/GeodeRegionRDD.scala ---------------------------------------------------------------------- diff --git a/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/internal/rdd/GeodeRegionRDD.scala b/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/internal/rdd/GeodeRegionRDD.scala index 6980c0f..a73be8f 100644 --- a/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/internal/rdd/GeodeRegionRDD.scala +++ b/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/internal/rdd/GeodeRegionRDD.scala @@ -14,14 +14,14 @@ * See the License for the specific language governing permissions and * limitations under the License. */ -package io.pivotal.geode.spark.connector.internal.rdd +package org.apache.geode.spark.connector.internal.rdd import scala.collection.Seq import scala.reflect.ClassTag import org.apache.spark.rdd.RDD import org.apache.spark.{TaskContext, Partition, SparkContext} -import io.pivotal.geode.spark.connector.{GeodeConnectionConf, PreferredPartitionerPropKey} -import io.pivotal.geode.spark.connector.internal.rdd.GeodeRDDPartitioner._ +import org.apache.geode.spark.connector.{GeodeConnectionConf, PreferredPartitionerPropKey} +import org.apache.geode.spark.connector.internal.rdd.GeodeRDDPartitioner._ /** * This class exposes Geode region as a RDD.
http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/javaapi/GeodeJavaRegionRDD.scala ---------------------------------------------------------------------- diff --git a/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/javaapi/GeodeJavaRegionRDD.scala b/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/javaapi/GeodeJavaRegionRDD.scala index f859173..9c1f923 100644 --- a/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/javaapi/GeodeJavaRegionRDD.scala +++ b/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/javaapi/GeodeJavaRegionRDD.scala @@ -14,9 +14,9 @@ * See the License for the specific language governing permissions and * limitations under the License. */ -package io.pivotal.geode.spark.connector.javaapi +package org.apache.geode.spark.connector.javaapi -import io.pivotal.geode.spark.connector.internal.rdd.GeodeRegionRDD +import org.apache.geode.spark.connector.internal.rdd.GeodeRegionRDD import org.apache.spark.api.java.JavaPairRDD class GeodeJavaRegionRDD[K, V](rdd: GeodeRegionRDD[K, V]) extends JavaPairRDD[K, V](rdd)(rdd.kClassTag, rdd.vClassTag) { http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/javaapi/JavaAPIHelper.scala ---------------------------------------------------------------------- diff --git a/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/javaapi/JavaAPIHelper.scala b/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/javaapi/JavaAPIHelper.scala index ffa6195..423b8bb 100644 --- a/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/javaapi/JavaAPIHelper.scala +++ b/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/javaapi/JavaAPIHelper.scala @@ -14,7 +14,7 @@ * See the License for the specific language governing permissions and * limitations under the License. */ -package io.pivotal.geode.spark.connector.javaapi +package org.apache.geode.spark.connector.javaapi import org.apache.spark.api.java.{JavaPairRDD, JavaRDD} import org.apache.spark.streaming.api.java.{JavaPairDStream, JavaDStream} http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/package.scala ---------------------------------------------------------------------- diff --git a/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/package.scala b/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/package.scala index 6f9a780..296695f 100644 --- a/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/package.scala +++ b/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/package.scala @@ -14,9 +14,9 @@ * See the License for the specific language governing permissions and * limitations under the License. */ -package io.pivotal.geode.spark +package org.apache.geode.spark -import io.pivotal.geode.spark.connector.internal.rdd.{ServerSplitsPartitioner, OnePartitionPartitioner} +import org.apache.geode.spark.connector.internal.rdd.{ServerSplitsPartitioner, OnePartitionPartitioner} import org.apache.spark.SparkContext import org.apache.spark.rdd.RDD import org.apache.spark.sql.SQLContext http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/streaming/GeodeDStreamFunctions.scala ---------------------------------------------------------------------- diff --git a/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/streaming/GeodeDStreamFunctions.scala b/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/streaming/GeodeDStreamFunctions.scala index 4d46429..7f06fba 100644 --- a/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/streaming/GeodeDStreamFunctions.scala +++ b/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/streaming/GeodeDStreamFunctions.scala @@ -14,17 +14,17 @@ * See the License for the specific language governing permissions and * limitations under the License. */ -package io.pivotal.geode.spark.connector.streaming +package org.apache.geode.spark.connector.streaming -import io.pivotal.geode.spark.connector.GeodeConnectionConf -import io.pivotal.geode.spark.connector.internal.rdd.{GeodePairRDDWriter, GeodeRDDWriter} +import org.apache.geode.spark.connector.GeodeConnectionConf +import org.apache.geode.spark.connector.internal.rdd.{GeodePairRDDWriter, GeodeRDDWriter} import org.apache.spark.Logging import org.apache.spark.api.java.function.PairFunction import org.apache.spark.streaming.dstream.DStream /** * Extra geode functions on DStream of non-pair elements through an implicit conversion. - * Import `io.pivotal.geode.spark.connector.streaming._` at the top of your program to + * Import `org.apache.geode.spark.connector.streaming._` at the top of your program to * use these functions. */ class GeodeDStreamFunctions[T](val dstream: DStream[T]) extends Serializable with Logging { @@ -63,7 +63,7 @@ class GeodeDStreamFunctions[T](val dstream: DStream[T]) extends Serializable wit /** * Extra geode functions on DStream of (key, value) pairs through an implicit conversion. - * Import `io.pivotal.geode.spark.connector.streaming._` at the top of your program to + * Import `org.apache.geode.spark.connector.streaming._` at the top of your program to * use these functions. */ class GeodePairDStreamFunctions[K, V](val dstream: DStream[(K,V)]) extends Serializable with Logging { http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/streaming/package.scala ---------------------------------------------------------------------- diff --git a/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/streaming/package.scala b/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/streaming/package.scala index 0d1f1eb..c4e0f27 100644 --- a/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/streaming/package.scala +++ b/geode-spark-connector/geode-spark-connector/src/main/scala/org/apache/geode/spark/connector/streaming/package.scala @@ -14,7 +14,7 @@ * See the License for the specific language governing permissions and * limitations under the License. */ -package io.pivotal.geode.spark.connector +package org.apache.geode.spark.connector import org.apache.spark.streaming.dstream.DStream http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/geode-spark-connector/src/test/java/org/apache/geode/spark/connector/JavaAPITest.java ---------------------------------------------------------------------- diff --git a/geode-spark-connector/geode-spark-connector/src/test/java/org/apache/geode/spark/connector/JavaAPITest.java b/geode-spark-connector/geode-spark-connector/src/test/java/org/apache/geode/spark/connector/JavaAPITest.java index 142907e..9af52da 100644 --- a/geode-spark-connector/geode-spark-connector/src/test/java/org/apache/geode/spark/connector/JavaAPITest.java +++ b/geode-spark-connector/geode-spark-connector/src/test/java/org/apache/geode/spark/connector/JavaAPITest.java @@ -14,9 +14,9 @@ * See the License for the specific language governing permissions and * limitations under the License. */ -package io.pivotal.geode.spark.connector; +package org.apache.geode.spark.connector; -import io.pivotal.geode.spark.connector.javaapi.*; +import org.apache.geode.spark.connector.javaapi.*; import org.apache.spark.SparkContext; import org.apache.spark.api.java.JavaPairRDD; import org.apache.spark.api.java.JavaRDD; @@ -38,7 +38,7 @@ import scala.collection.mutable.LinkedList; import scala.reflect.ClassTag; import static org.junit.Assert.*; -import static io.pivotal.geode.spark.connector.javaapi.GeodeJavaUtil.*; +import static org.apache.geode.spark.connector.javaapi.GeodeJavaUtil.*; import static org.mockito.Matchers.eq; import static org.mockito.Mockito.*; http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/geode-spark-connector/src/test/scala/org/apache/geode/spark/connector/GeodeFunctionDeployerTest.scala ---------------------------------------------------------------------- diff --git a/geode-spark-connector/geode-spark-connector/src/test/scala/org/apache/geode/spark/connector/GeodeFunctionDeployerTest.scala b/geode-spark-connector/geode-spark-connector/src/test/scala/org/apache/geode/spark/connector/GeodeFunctionDeployerTest.scala index 4e45dc2..77a5668 100644 --- a/geode-spark-connector/geode-spark-connector/src/test/scala/org/apache/geode/spark/connector/GeodeFunctionDeployerTest.scala +++ b/geode-spark-connector/geode-spark-connector/src/test/scala/org/apache/geode/spark/connector/GeodeFunctionDeployerTest.scala @@ -14,7 +14,7 @@ * See the License for the specific language governing permissions and * limitations under the License. */ -package io.pivotal.geode.spark.connector +package org.apache.geode.spark.connector import org.mockito.Mockito._ import org.scalatest.mock.MockitoSugar http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/geode-spark-connector/src/test/scala/org/apache/geode/spark/connector/internal/DefaultGeodeConnectionManagerTest.scala ---------------------------------------------------------------------- diff --git a/geode-spark-connector/geode-spark-connector/src/test/scala/org/apache/geode/spark/connector/internal/DefaultGeodeConnectionManagerTest.scala b/geode-spark-connector/geode-spark-connector/src/test/scala/org/apache/geode/spark/connector/internal/DefaultGeodeConnectionManagerTest.scala index 798912c..4863a61 100644 --- a/geode-spark-connector/geode-spark-connector/src/test/scala/org/apache/geode/spark/connector/internal/DefaultGeodeConnectionManagerTest.scala +++ b/geode-spark-connector/geode-spark-connector/src/test/scala/org/apache/geode/spark/connector/internal/DefaultGeodeConnectionManagerTest.scala @@ -14,9 +14,9 @@ * See the License for the specific language governing permissions and * limitations under the License. */ -package io.pivotal.geode.spark.connector.internal +package org.apache.geode.spark.connector.internal -import io.pivotal.geode.spark.connector.GeodeConnectionConf +import org.apache.geode.spark.connector.GeodeConnectionConf import org.mockito.Mockito._ import org.scalatest.mock.MockitoSugar import org.scalatest.{FunSuite, Matchers} http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/geode-spark-connector/src/test/scala/org/apache/geode/spark/connector/internal/gemfirefunctions/StructStreamingResultSenderAndCollectorTest.scala ---------------------------------------------------------------------- diff --git a/geode-spark-connector/geode-spark-connector/src/test/scala/org/apache/geode/spark/connector/internal/gemfirefunctions/StructStreamingResultSenderAndCollectorTest.scala b/geode-spark-connector/geode-spark-connector/src/test/scala/org/apache/geode/spark/connector/internal/gemfirefunctions/StructStreamingResultSenderAndCollectorTest.scala index c95f1dc..ca77fb0 100644 --- a/geode-spark-connector/geode-spark-connector/src/test/scala/org/apache/geode/spark/connector/internal/gemfirefunctions/StructStreamingResultSenderAndCollectorTest.scala +++ b/geode-spark-connector/geode-spark-connector/src/test/scala/org/apache/geode/spark/connector/internal/gemfirefunctions/StructStreamingResultSenderAndCollectorTest.scala @@ -14,7 +14,7 @@ * See the License for the specific language governing permissions and * limitations under the License. */ -package io.pivotal.geode.spark.connector.internal.geodefunctions +package org.apache.geode.spark.connector.internal.geodefunctions import org.apache.geode.DataSerializer import org.apache.geode.cache.execute.{ResultCollector, ResultSender} http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/geode-spark-connector/src/test/scala/org/apache/geode/spark/connector/internal/oql/QueryParserTest.scala ---------------------------------------------------------------------- diff --git a/geode-spark-connector/geode-spark-connector/src/test/scala/org/apache/geode/spark/connector/internal/oql/QueryParserTest.scala b/geode-spark-connector/geode-spark-connector/src/test/scala/org/apache/geode/spark/connector/internal/oql/QueryParserTest.scala index 54394e8..f5c32dc 100644 --- a/geode-spark-connector/geode-spark-connector/src/test/scala/org/apache/geode/spark/connector/internal/oql/QueryParserTest.scala +++ b/geode-spark-connector/geode-spark-connector/src/test/scala/org/apache/geode/spark/connector/internal/oql/QueryParserTest.scala @@ -14,7 +14,7 @@ * See the License for the specific language governing permissions and * limitations under the License. */ -package io.pivotal.geode.spark.connector.internal.oql +package org.apache.geode.spark.connector.internal.oql import org.scalatest.FunSuite @@ -55,8 +55,8 @@ class QueryParserTest extends FunSuite { assert(r == "List(/r1/r2)") } - test("IMPORT io.pivotal.geode IMPORT com.mypackage SELECT key,value FROM /root/sub.entries WHERE status = 'active' ORDER BY id desc") { - val r = QueryParser.parseOQL("IMPORT io.pivotal.geode IMPORT com.mypackage SELECT key,value FROM /root/sub.entries WHERE status = 'active' ORDER BY id desc").get + test("IMPORT org.apache.geode IMPORT com.mypackage SELECT key,value FROM /root/sub.entries WHERE status = 'active' ORDER BY id desc") { + val r = QueryParser.parseOQL("IMPORT org.apache.geode IMPORT com.mypackage SELECT key,value FROM /root/sub.entries WHERE status = 'active' ORDER BY id desc").get assert(r == "List(/root/sub.entries)") } http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/ConnectorImplicitsTest.scala ---------------------------------------------------------------------- diff --git a/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/ConnectorImplicitsTest.scala b/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/ConnectorImplicitsTest.scala index b0464cc..7b247af 100644 --- a/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/ConnectorImplicitsTest.scala +++ b/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/ConnectorImplicitsTest.scala @@ -14,9 +14,9 @@ * See the License for the specific language governing permissions and * limitations under the License. */ -package unittest.io.pivotal.geode.spark.connector +package unittest.org.apache.geode.spark.connector -import io.pivotal.geode.spark.connector._ +import org.apache.geode.spark.connector._ import org.apache.spark.SparkContext import org.apache.spark.sql.SQLContext import org.scalatest.FunSuite http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/GeodeConnectionConfTest.scala ---------------------------------------------------------------------- diff --git a/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/GeodeConnectionConfTest.scala b/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/GeodeConnectionConfTest.scala index a3076f4..2fad7a6 100644 --- a/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/GeodeConnectionConfTest.scala +++ b/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/GeodeConnectionConfTest.scala @@ -14,13 +14,13 @@ * See the License for the specific language governing permissions and * limitations under the License. */ -package unittest.io.pivotal.geode.spark.connector +package unittest.org.apache.geode.spark.connector import org.apache.spark.SparkConf import org.mockito.Mockito._ import org.scalatest.mock.MockitoSugar import org.scalatest.{Matchers, FunSuite} -import io.pivotal.geode.spark.connector._ +import org.apache.geode.spark.connector._ class GeodeConnectionConfTest extends FunSuite with Matchers with MockitoSugar { http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/GeodeDStreamFunctionsTest.scala ---------------------------------------------------------------------- diff --git a/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/GeodeDStreamFunctionsTest.scala b/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/GeodeDStreamFunctionsTest.scala index d671722..19ec4f0 100644 --- a/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/GeodeDStreamFunctionsTest.scala +++ b/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/GeodeDStreamFunctionsTest.scala @@ -14,10 +14,10 @@ * See the License for the specific language governing permissions and * limitations under the License. */ -package unittest.io.pivotal.geode.spark.connector +package unittest.org.apache.geode.spark.connector import org.apache.geode.cache.Region -import io.pivotal.geode.spark.connector.{GeodeConnection, GeodeConnectionConf} +import org.apache.geode.spark.connector.{GeodeConnection, GeodeConnectionConf} import org.apache.spark.rdd.RDD import org.apache.spark.streaming.dstream.DStream import org.mockito.Mockito._ @@ -30,7 +30,7 @@ import scala.reflect.ClassTag class GeodeDStreamFunctionsTest extends FunSuite with Matchers with MockitoSugar { test("test GeodePairDStreamFunctions Implicit") { - import io.pivotal.geode.spark.connector.streaming._ + import org.apache.geode.spark.connector.streaming._ val mockDStream = mock[DStream[(Int, String)]] // the implicit make the following line valid val pairDStream: GeodePairDStreamFunctions[Int, String] = mockDStream @@ -38,7 +38,7 @@ class GeodeDStreamFunctionsTest extends FunSuite with Matchers with MockitoSugar } test("test GeodeDStreamFunctions Implicit") { - import io.pivotal.geode.spark.connector.streaming._ + import org.apache.geode.spark.connector.streaming._ val mockDStream = mock[DStream[String]] // the implicit make the following line valid val dstream: GeodeDStreamFunctions[String] = mockDStream @@ -57,7 +57,7 @@ class GeodeDStreamFunctionsTest extends FunSuite with Matchers with MockitoSugar } test("test GeodePairDStreamFunctions.saveToGeode()") { - import io.pivotal.geode.spark.connector.streaming._ + import org.apache.geode.spark.connector.streaming._ val (regionPath, mockConnConf, mockConnection, mockRegion) = createMocks[String, String]("test") val mockDStream = mock[DStream[(String, String)]] mockDStream.saveToGeode(regionPath, mockConnConf) @@ -67,7 +67,7 @@ class GeodeDStreamFunctionsTest extends FunSuite with Matchers with MockitoSugar } test("test GeodeDStreamFunctions.saveToGeode()") { - import io.pivotal.geode.spark.connector.streaming._ + import org.apache.geode.spark.connector.streaming._ val (regionPath, mockConnConf, mockConnection, mockRegion) = createMocks[String, Int]("test") val mockDStream = mock[DStream[String]] mockDStream.saveToGeode[String, Int](regionPath, (s: String) => (s, s.length), mockConnConf) http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/GeodeRDDFunctionsTest.scala ---------------------------------------------------------------------- diff --git a/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/GeodeRDDFunctionsTest.scala b/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/GeodeRDDFunctionsTest.scala index 5259198..a6fc91d 100644 --- a/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/GeodeRDDFunctionsTest.scala +++ b/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/GeodeRDDFunctionsTest.scala @@ -14,11 +14,11 @@ * See the License for the specific language governing permissions and * limitations under the License. */ -package unittest.io.pivotal.geode.spark.connector +package unittest.org.apache.geode.spark.connector import org.apache.geode.cache.Region -import io.pivotal.geode.spark.connector._ -import io.pivotal.geode.spark.connector.internal.rdd.{GeodeRDDWriter, GeodePairRDDWriter} +import org.apache.geode.spark.connector._ +import org.apache.geode.spark.connector.internal.rdd.{GeodeRDDWriter, GeodePairRDDWriter} import org.apache.spark.{TaskContext, SparkContext} import org.apache.spark.rdd.RDD import org.mockito.Mockito._ @@ -31,7 +31,7 @@ import org.mockito.Matchers.{eq => mockEq, any => mockAny} class GeodeRDDFunctionsTest extends FunSuite with Matchers with MockitoSugar { test("test PairRDDFunction Implicit") { - import io.pivotal.geode.spark.connector._ + import org.apache.geode.spark.connector._ val mockRDD = mock[RDD[(Int, String)]] // the implicit make the following line valid val pairRDD: GeodePairRDDFunctions[Int, String] = mockRDD @@ -39,7 +39,7 @@ class GeodeRDDFunctionsTest extends FunSuite with Matchers with MockitoSugar { } test("test RDDFunction Implicit") { - import io.pivotal.geode.spark.connector._ + import org.apache.geode.spark.connector._ val mockRDD = mock[RDD[String]] // the implicit make the following line valid val nonPairRDD: GeodeRDDFunctions[String] = mockRDD @@ -85,7 +85,7 @@ class GeodeRDDFunctionsTest extends FunSuite with Matchers with MockitoSugar { } def verifyPairRDDFunction(useOpConf: Boolean): Unit = { - import io.pivotal.geode.spark.connector._ + import org.apache.geode.spark.connector._ val (regionPath, mockConnConf, mockConnection, mockRegion) = createMocks[String, String]("test") val mockRDD = mock[RDD[(String, String)]] val mockSparkContext = mock[SparkContext] @@ -115,7 +115,7 @@ class GeodeRDDFunctionsTest extends FunSuite with Matchers with MockitoSugar { } def verifyRDDFunction(useOpConf: Boolean): Unit = { - import io.pivotal.geode.spark.connector._ + import org.apache.geode.spark.connector._ val (regionPath, mockConnConf, mockConnection, mockRegion) = createMocks[Int, String]("test") val mockRDD = mock[RDD[(String)]] val mockSparkContext = mock[SparkContext] http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/LocatorHelperTest.scala ---------------------------------------------------------------------- diff --git a/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/LocatorHelperTest.scala b/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/LocatorHelperTest.scala index c775784..5719996 100644 --- a/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/LocatorHelperTest.scala +++ b/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/LocatorHelperTest.scala @@ -14,11 +14,11 @@ * See the License for the specific language governing permissions and * limitations under the License. */ -package unittest.io.pivotal.geode.spark.connector +package unittest.org.apache.geode.spark.connector import java.net.InetAddress -import io.pivotal.geode.spark.connector.internal.LocatorHelper +import org.apache.geode.spark.connector.internal.LocatorHelper import org.scalatest.FunSuite class LocatorHelperTest extends FunSuite { http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/rdd/GeodeRDDPartitionerTest.scala ---------------------------------------------------------------------- diff --git a/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/rdd/GeodeRDDPartitionerTest.scala b/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/rdd/GeodeRDDPartitionerTest.scala index 2f92c1d..9a42866 100644 --- a/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/rdd/GeodeRDDPartitionerTest.scala +++ b/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/rdd/GeodeRDDPartitionerTest.scala @@ -14,13 +14,13 @@ * See the License for the specific language governing permissions and * limitations under the License. */ -package unittest.io.pivotal.geode.spark.connector.rdd +package unittest.org.apache.geode.spark.connector.rdd import org.apache.geode.distributed.internal.ServerLocation -import io.pivotal.geode.spark.connector.internal.RegionMetadata -import io.pivotal.geode.spark.connector.internal.rdd.GeodeRDDPartitioner._ -import io.pivotal.geode.spark.connector.GeodeConnection -import io.pivotal.geode.spark.connector.internal.rdd._ +import org.apache.geode.spark.connector.internal.RegionMetadata +import org.apache.geode.spark.connector.internal.rdd.GeodeRDDPartitioner._ +import org.apache.geode.spark.connector.GeodeConnection +import org.apache.geode.spark.connector.internal.rdd._ import org.apache.spark.Partition import org.mockito.Mockito._ import org.scalatest.mock.MockitoSugar @@ -55,7 +55,7 @@ class GeodeRDDPartitionerTest extends FunSuite with Matchers with MockitoSugar { } test("GeodeRDDPartitioner.apply method") { - import io.pivotal.geode.spark.connector.internal.rdd.GeodeRDDPartitioner._ + import org.apache.geode.spark.connector.internal.rdd.GeodeRDDPartitioner._ for ((name, partitioner) <- partitioners) assert(GeodeRDDPartitioner(name) == partitioner) assert(GeodeRDDPartitioner("dummy") == GeodeRDDPartitioner.defaultPartitionedRegionPartitioner) assert(GeodeRDDPartitioner() == GeodeRDDPartitioner.defaultPartitionedRegionPartitioner) @@ -158,7 +158,7 @@ class GeodeRDDPartitionerTest extends FunSuite with Matchers with MockitoSugar { } test("ServerSplitsPartitioner.partitions(): partitioned region w/ some data ") { - import io.pivotal.geode.spark.connector.NumberPartitionsPerServerPropKey + import org.apache.geode.spark.connector.NumberPartitionsPerServerPropKey val regionPath = "test" val mockConnection = mock[GeodeConnection] val map: Map[(String, Int), Set[Int]] = Map( http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/rdd/GeodeRegionRDDTest.scala ---------------------------------------------------------------------- diff --git a/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/rdd/GeodeRegionRDDTest.scala b/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/rdd/GeodeRegionRDDTest.scala index 63eddf9..8a5c67e 100644 --- a/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/rdd/GeodeRegionRDDTest.scala +++ b/geode-spark-connector/geode-spark-connector/src/test/scala/unittest/org/apache/geode/spark/connector/rdd/GeodeRegionRDDTest.scala @@ -14,12 +14,12 @@ * See the License for the specific language governing permissions and * limitations under the License. */ -package unittest.io.pivotal.geode.spark.connector.rdd +package unittest.org.apache.geode.spark.connector.rdd import org.apache.geode.cache.Region -import io.pivotal.geode.spark.connector.internal.RegionMetadata -import io.pivotal.geode.spark.connector.internal.rdd.{GeodeRDDPartition, GeodeRegionRDD} -import io.pivotal.geode.spark.connector.{GeodeConnectionConf, GeodeConnection} +import org.apache.geode.spark.connector.internal.RegionMetadata +import org.apache.geode.spark.connector.internal.rdd.{GeodeRDDPartition, GeodeRegionRDD} +import org.apache.geode.spark.connector.{GeodeConnectionConf, GeodeConnection} import org.apache.spark.{TaskContext, Partition, SparkContext} import org.mockito.Mockito._ import org.mockito.Matchers.{eq => mockEq, any => mockAny} @@ -78,7 +78,7 @@ class GeodeRegionRDDTest extends FunSuite with Matchers with MockitoSugar { test("getPartitions with replicated region and preferred OnePartitionPartitioner") { // since it's replicated region, so OnePartitionPartitioner will be used, i.e., override preferred partitioner - import io.pivotal.geode.spark.connector.{PreferredPartitionerPropKey, OnePartitionPartitionerName} + import org.apache.geode.spark.connector.{PreferredPartitionerPropKey, OnePartitionPartitionerName} val (regionPath, mockRegion, mockConnConf, mockConnection) = createMocks[String, String]("test") when(mockConnection.getRegionMetadata[String, String](regionPath)).thenReturn(Some(new RegionMetadata(regionPath, false, 0, null))) implicit val mockConnConf2 = mockConnConf http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/geode-spark-demos/basic-demos/src/main/java/demo/OQLJavaDemo.java ---------------------------------------------------------------------- diff --git a/geode-spark-connector/geode-spark-demos/basic-demos/src/main/java/demo/OQLJavaDemo.java b/geode-spark-connector/geode-spark-demos/basic-demos/src/main/java/demo/OQLJavaDemo.java index adcf072..9107796 100644 --- a/geode-spark-connector/geode-spark-demos/basic-demos/src/main/java/demo/OQLJavaDemo.java +++ b/geode-spark-connector/geode-spark-demos/basic-demos/src/main/java/demo/OQLJavaDemo.java @@ -20,7 +20,7 @@ import org.apache.spark.SparkConf; import org.apache.spark.api.java.JavaSparkContext; import org.apache.spark.sql.DataFrame; import org.apache.spark.sql.SQLContext; -import static io.pivotal.geode.spark.connector.javaapi.GeodeJavaUtil.*; +import static org.apache.geode.spark.connector.javaapi.GeodeJavaUtil.*; /** http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/geode-spark-demos/basic-demos/src/main/java/demo/PairRDDSaveJavaDemo.java ---------------------------------------------------------------------- diff --git a/geode-spark-connector/geode-spark-demos/basic-demos/src/main/java/demo/PairRDDSaveJavaDemo.java b/geode-spark-connector/geode-spark-demos/basic-demos/src/main/java/demo/PairRDDSaveJavaDemo.java index 52d2a99..d1edd41 100644 --- a/geode-spark-connector/geode-spark-demos/basic-demos/src/main/java/demo/PairRDDSaveJavaDemo.java +++ b/geode-spark-connector/geode-spark-demos/basic-demos/src/main/java/demo/PairRDDSaveJavaDemo.java @@ -16,7 +16,7 @@ */ package demo; -import io.pivotal.geode.spark.connector.GeodeConnectionConf; +import org.apache.geode.spark.connector.GeodeConnectionConf; import org.apache.spark.SparkConf; import org.apache.spark.api.java.JavaPairRDD; import org.apache.spark.api.java.JavaRDD; @@ -24,7 +24,7 @@ import org.apache.spark.api.java.JavaSparkContext; import scala.Tuple2; import java.util.*; -import static io.pivotal.geode.spark.connector.javaapi.GeodeJavaUtil.*; +import static org.apache.geode.spark.connector.javaapi.GeodeJavaUtil.*; /** * This Spark application demonstrates how to save a RDD to Geode using Geode Spark http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/geode-spark-demos/basic-demos/src/main/java/demo/RDDSaveJavaDemo.java ---------------------------------------------------------------------- diff --git a/geode-spark-connector/geode-spark-demos/basic-demos/src/main/java/demo/RDDSaveJavaDemo.java b/geode-spark-connector/geode-spark-demos/basic-demos/src/main/java/demo/RDDSaveJavaDemo.java index 1125de5..22c01f4 100644 --- a/geode-spark-connector/geode-spark-demos/basic-demos/src/main/java/demo/RDDSaveJavaDemo.java +++ b/geode-spark-connector/geode-spark-demos/basic-demos/src/main/java/demo/RDDSaveJavaDemo.java @@ -16,7 +16,7 @@ */ package demo; -import io.pivotal.geode.spark.connector.GeodeConnectionConf; +import org.apache.geode.spark.connector.GeodeConnectionConf; import org.apache.spark.SparkConf; import org.apache.spark.api.java.JavaRDD; import org.apache.spark.api.java.JavaSparkContext; @@ -26,7 +26,7 @@ import scala.Tuple2; import java.util.ArrayList; import java.util.List; -import static io.pivotal.geode.spark.connector.javaapi.GeodeJavaUtil.*; +import static org.apache.geode.spark.connector.javaapi.GeodeJavaUtil.*; /** * This Spark application demonstrates how to save a RDD to Geode using Geode Spark http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/geode-spark-demos/basic-demos/src/main/java/demo/RegionToRDDJavaDemo.java ---------------------------------------------------------------------- diff --git a/geode-spark-connector/geode-spark-demos/basic-demos/src/main/java/demo/RegionToRDDJavaDemo.java b/geode-spark-connector/geode-spark-demos/basic-demos/src/main/java/demo/RegionToRDDJavaDemo.java index 1ce8ceb..41a07f5 100644 --- a/geode-spark-connector/geode-spark-demos/basic-demos/src/main/java/demo/RegionToRDDJavaDemo.java +++ b/geode-spark-connector/geode-spark-demos/basic-demos/src/main/java/demo/RegionToRDDJavaDemo.java @@ -19,7 +19,7 @@ package demo; import org.apache.spark.SparkConf; import org.apache.spark.api.java.JavaPairRDD; import org.apache.spark.api.java.JavaSparkContext; -import static io.pivotal.geode.spark.connector.javaapi.GeodeJavaUtil.*; +import static org.apache.geode.spark.connector.javaapi.GeodeJavaUtil.*; /** * This Spark application demonstrates how to expose a region in Geode as a RDD using Geode http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/geode-spark-demos/basic-demos/src/main/scala/demo/NetworkWordCount.scala ---------------------------------------------------------------------- diff --git a/geode-spark-connector/geode-spark-demos/basic-demos/src/main/scala/demo/NetworkWordCount.scala b/geode-spark-connector/geode-spark-demos/basic-demos/src/main/scala/demo/NetworkWordCount.scala index 810b380..063ea69 100644 --- a/geode-spark-connector/geode-spark-demos/basic-demos/src/main/scala/demo/NetworkWordCount.scala +++ b/geode-spark-connector/geode-spark-demos/basic-demos/src/main/scala/demo/NetworkWordCount.scala @@ -18,8 +18,8 @@ package demo import org.apache.spark.SparkConf import org.apache.spark.streaming.{Seconds, StreamingContext} -import io.pivotal.geode.spark.connector.GeodeLocatorPropKey -import io.pivotal.geode.spark.connector.streaming._ +import org.apache.geode.spark.connector.GeodeLocatorPropKey +import org.apache.geode.spark.connector.streaming._ /** * Counts words in UTF8 encoded, '\n' delimited text received from the network every second. http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/c32fccba/geode-spark-connector/project/Settings.scala ---------------------------------------------------------------------- diff --git a/geode-spark-connector/project/Settings.scala b/geode-spark-connector/project/Settings.scala index 9aefa9b..8cfe38f 100644 --- a/geode-spark-connector/project/Settings.scala +++ b/geode-spark-connector/project/Settings.scala @@ -24,7 +24,7 @@ object Settings extends Build { organization := "io.pivotal", version := "0.5.0", scalaVersion := "2.10.4", - organization := "io.pivotal.geode.spark", + organization := "org.apache.geode.spark", organizationHomepage := Some(url("http://www.pivotal.io/")) )