Repository: spark
Updated Branches:
  refs/heads/branch-1.0 fdebb6952 -> 74718285a


SPARK-1488. Resolve scalac feature warnings during build

For your consideration: scalac currently notes a number of feature warnings 
during compilation:

```
[warn] there were 65 feature warning(s); re-run with -feature for details
```

Warnings are like:

```
[warn] 
/Users/srowen/Documents/spark/core/src/main/scala/org/apache/spark/SparkContext.scala:1261:
 implicit conversion method rddToPairRDDFunctions should be enabled
[warn] by making the implicit value scala.language.implicitConversions visible.
[warn] This can be achieved by adding the import clause 'import 
scala.language.implicitConversions'
[warn] or by setting the compiler option -language:implicitConversions.
[warn] See the Scala docs for value scala.language.implicitConversions for a 
discussion
[warn] why the feature should be explicitly enabled.
[warn]   implicit def rddToPairRDDFunctions[K: ClassTag, V: ClassTag](rdd: 
RDD[(K, V)]) =
[warn]                ^
```

scalac is suggesting that it's just best practice to explicitly enable certain 
language features by importing them where used.

This PR simply adds the imports it suggests (and squashes one other Java 
warning along the way). This leaves just deprecation warnings in the build.

Author: Sean Owen <so...@cloudera.com>

Closes #404 from srowen/SPARK-1488 and squashes the following commits:

8598980 [Sean Owen] Quiet scalac warnings about language features by explicitly 
importing language features.
39bc831 [Sean Owen] Enable -feature in scalac to emit language feature warnings
(cherry picked from commit 0247b5c5467ca1b0d03ba929a78fa4d805582d84)

Signed-off-by: Patrick Wendell <pwend...@gmail.com>


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/74718285
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/74718285
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/74718285

Branch: refs/heads/branch-1.0
Commit: 74718285adc9ddce434473fcae32fe32a6c95c74
Parents: fdebb69
Author: Sean Owen <so...@cloudera.com>
Authored: Mon Apr 14 19:50:00 2014 -0700
Committer: Patrick Wendell <pwend...@gmail.com>
Committed: Mon Apr 14 19:55:48 2014 -0700

----------------------------------------------------------------------
 bagel/src/test/scala/org/apache/spark/bagel/BagelSuite.scala      | 2 ++
 core/src/main/scala/org/apache/spark/SparkContext.scala           | 1 +
 core/src/main/scala/org/apache/spark/api/java/JavaDoubleRDD.scala | 1 +
 core/src/main/scala/org/apache/spark/api/java/JavaPairRDD.scala   | 1 +
 core/src/main/scala/org/apache/spark/api/java/JavaRDD.scala       | 1 +
 .../main/scala/org/apache/spark/api/java/JavaSparkContext.scala   | 1 +
 .../main/scala/org/apache/spark/deploy/FaultToleranceTest.scala   | 1 +
 core/src/main/scala/org/apache/spark/deploy/master/Master.scala   | 1 +
 core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala   | 1 +
 .../main/scala/org/apache/spark/network/ConnectionManager.scala   | 1 +
 core/src/main/scala/org/apache/spark/rdd/CoGroupedRDD.scala       | 1 +
 core/src/main/scala/org/apache/spark/rdd/CoalescedRDD.scala       | 1 +
 .../main/scala/org/apache/spark/scheduler/DAGSchedulerEvent.scala | 1 +
 core/src/main/scala/org/apache/spark/scheduler/ResultTask.scala   | 1 +
 .../main/scala/org/apache/spark/scheduler/ShuffleMapTask.scala    | 1 +
 .../main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala | 1 +
 core/src/main/scala/org/apache/spark/ui/JettyUtils.scala          | 1 +
 .../scala/org/apache/spark/util/TimeStampedWeakValueHashMap.scala | 1 +
 core/src/main/scala/org/apache/spark/util/Vector.scala            | 1 +
 core/src/test/scala/org/apache/spark/ConnectionManagerSuite.scala | 2 +-
 core/src/test/scala/org/apache/spark/ContextCleanerSuite.scala    | 2 ++
 core/src/test/scala/org/apache/spark/DriverSuite.scala            | 2 ++
 core/src/test/scala/org/apache/spark/PipedRDDSuite.scala          | 3 +--
 .../test/scala/org/apache/spark/rdd/AsyncRDDActionsSuite.scala    | 1 +
 .../test/scala/org/apache/spark/scheduler/DAGSchedulerSuite.scala | 1 +
 .../test/scala/org/apache/spark/storage/BlockManagerSuite.scala   | 3 +++
 .../scala/org/apache/spark/storage/DiskBlockManagerSuite.scala    | 1 +
 core/src/test/scala/org/apache/spark/ui/UISuite.scala             | 1 +
 .../scala/org/apache/spark/util/random/XORShiftRandomSuite.scala  | 2 ++
 .../org/apache/spark/streaming/examples/ZeroMQWordCount.scala     | 2 ++
 graphx/src/main/scala/org/apache/spark/graphx/Graph.scala         | 1 +
 .../scala/org/apache/spark/graphx/impl/MessageToPartition.scala   | 1 +
 .../src/main/scala/org/apache/spark/graphx/impl/Serializers.scala | 3 ++-
 .../main/scala/org/apache/spark/graphx/util/BytecodeUtils.scala   | 1 +
 .../src/main/scala/org/apache/spark/mllib/rdd/RDDFunctions.scala  | 1 +
 .../main/scala/org/apache/spark/mllib/util/MFDataGenerator.scala  | 1 +
 .../test/java/org/apache/spark/mllib/linalg/JavaVectorsSuite.java | 1 +
 pom.xml                                                           | 1 +
 project/SparkBuild.scala                                          | 2 +-
 .../src/main/scala/org/apache/spark/sql/catalyst/SqlParser.scala  | 1 +
 .../main/scala/org/apache/spark/streaming/StreamingContext.scala  | 1 +
 .../scala/org/apache/spark/streaming/api/java/JavaDStream.scala   | 1 +
 .../org/apache/spark/streaming/api/java/JavaDStreamLike.scala     | 1 +
 .../org/apache/spark/streaming/api/java/JavaPairDStream.scala     | 1 +
 .../org/apache/spark/streaming/receivers/ActorReceiver.scala      | 1 +
 .../apache/spark/streaming/scheduler/NetworkInputTracker.scala    | 1 +
 .../scala/org/apache/spark/streaming/StreamingContextSuite.scala  | 2 ++
 streaming/src/test/scala/org/apache/spark/streaming/UISuite.scala | 1 +
 .../scala/org/apache/spark/tools/JavaAPICompletenessChecker.scala | 1 +
 49 files changed, 58 insertions(+), 5 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/bagel/src/test/scala/org/apache/spark/bagel/BagelSuite.scala
----------------------------------------------------------------------
diff --git a/bagel/src/test/scala/org/apache/spark/bagel/BagelSuite.scala 
b/bagel/src/test/scala/org/apache/spark/bagel/BagelSuite.scala
index 69144e3..8e0f82d 100644
--- a/bagel/src/test/scala/org/apache/spark/bagel/BagelSuite.scala
+++ b/bagel/src/test/scala/org/apache/spark/bagel/BagelSuite.scala
@@ -24,6 +24,8 @@ import org.scalatest.time.SpanSugar._
 import org.apache.spark._
 import org.apache.spark.storage.StorageLevel
 
+import scala.language.postfixOps
+
 class TestVertex(val active: Boolean, val age: Int) extends Vertex with 
Serializable
 class TestMessage(val targetId: String) extends Message[String] with 
Serializable
 

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/main/scala/org/apache/spark/SparkContext.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/SparkContext.scala 
b/core/src/main/scala/org/apache/spark/SparkContext.scala
index 456070f..3ddc0d5 100644
--- a/core/src/main/scala/org/apache/spark/SparkContext.scala
+++ b/core/src/main/scala/org/apache/spark/SparkContext.scala
@@ -25,6 +25,7 @@ import java.util.UUID.randomUUID
 import scala.collection.{Map, Set}
 import scala.collection.generic.Growable
 import scala.collection.mutable.{ArrayBuffer, HashMap}
+import scala.language.implicitConversions
 import scala.reflect.{ClassTag, classTag}
 import org.apache.hadoop.conf.Configuration
 import org.apache.hadoop.fs.Path

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/main/scala/org/apache/spark/api/java/JavaDoubleRDD.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/api/java/JavaDoubleRDD.scala 
b/core/src/main/scala/org/apache/spark/api/java/JavaDoubleRDD.scala
index 537f410..4330cef 100644
--- a/core/src/main/scala/org/apache/spark/api/java/JavaDoubleRDD.scala
+++ b/core/src/main/scala/org/apache/spark/api/java/JavaDoubleRDD.scala
@@ -19,6 +19,7 @@ package org.apache.spark.api.java
 
 import java.lang.{Double => JDouble}
 
+import scala.language.implicitConversions
 import scala.reflect.ClassTag
 
 import org.apache.spark.Partitioner

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/main/scala/org/apache/spark/api/java/JavaPairRDD.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/api/java/JavaPairRDD.scala 
b/core/src/main/scala/org/apache/spark/api/java/JavaPairRDD.scala
index a41c7db..e5b2c8a 100644
--- a/core/src/main/scala/org/apache/spark/api/java/JavaPairRDD.scala
+++ b/core/src/main/scala/org/apache/spark/api/java/JavaPairRDD.scala
@@ -21,6 +21,7 @@ import java.util.{Comparator, List => JList}
 import java.lang.{Iterable => JIterable}
 
 import scala.collection.JavaConversions._
+import scala.language.implicitConversions
 import scala.reflect.ClassTag
 
 import com.google.common.base.Optional

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/main/scala/org/apache/spark/api/java/JavaRDD.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/api/java/JavaRDD.scala 
b/core/src/main/scala/org/apache/spark/api/java/JavaRDD.scala
index 01d9357..327c155 100644
--- a/core/src/main/scala/org/apache/spark/api/java/JavaRDD.scala
+++ b/core/src/main/scala/org/apache/spark/api/java/JavaRDD.scala
@@ -17,6 +17,7 @@
 
 package org.apache.spark.api.java
 
+import scala.language.implicitConversions
 import scala.reflect.ClassTag
 
 import org.apache.spark._

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/main/scala/org/apache/spark/api/java/JavaSparkContext.scala
----------------------------------------------------------------------
diff --git 
a/core/src/main/scala/org/apache/spark/api/java/JavaSparkContext.scala 
b/core/src/main/scala/org/apache/spark/api/java/JavaSparkContext.scala
index 7fbefe1..e6a3f06 100644
--- a/core/src/main/scala/org/apache/spark/api/java/JavaSparkContext.scala
+++ b/core/src/main/scala/org/apache/spark/api/java/JavaSparkContext.scala
@@ -22,6 +22,7 @@ import java.util.{Map => JMap}
 
 import scala.collection.JavaConversions
 import scala.collection.JavaConversions._
+import scala.language.implicitConversions
 import scala.reflect.ClassTag
 
 import com.google.common.base.Optional

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/main/scala/org/apache/spark/deploy/FaultToleranceTest.scala
----------------------------------------------------------------------
diff --git 
a/core/src/main/scala/org/apache/spark/deploy/FaultToleranceTest.scala 
b/core/src/main/scala/org/apache/spark/deploy/FaultToleranceTest.scala
index f4eb160..47dbcd8 100644
--- a/core/src/main/scala/org/apache/spark/deploy/FaultToleranceTest.scala
+++ b/core/src/main/scala/org/apache/spark/deploy/FaultToleranceTest.scala
@@ -25,6 +25,7 @@ import scala.collection.mutable.ListBuffer
 import scala.concurrent.{Await, future, promise}
 import scala.concurrent.ExecutionContext.Implicits.global
 import scala.concurrent.duration._
+import scala.language.postfixOps
 import scala.sys.process._
 
 import org.json4s._

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/main/scala/org/apache/spark/deploy/master/Master.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/deploy/master/Master.scala 
b/core/src/main/scala/org/apache/spark/deploy/master/Master.scala
index 6c58e74..81f990b 100644
--- a/core/src/main/scala/org/apache/spark/deploy/master/Master.scala
+++ b/core/src/main/scala/org/apache/spark/deploy/master/Master.scala
@@ -23,6 +23,7 @@ import java.util.Date
 import scala.collection.mutable.{ArrayBuffer, HashMap, HashSet}
 import scala.concurrent.Await
 import scala.concurrent.duration._
+import scala.language.postfixOps
 import scala.util.Random
 
 import akka.actor._

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala 
b/core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala
index 52c164c..dd0a136 100755
--- a/core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala
+++ b/core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala
@@ -23,6 +23,7 @@ import java.util.Date
 
 import scala.collection.mutable.HashMap
 import scala.concurrent.duration._
+import scala.language.postfixOps
 
 import akka.actor._
 import akka.remote.{DisassociatedEvent, RemotingLifecycleEvent}

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/main/scala/org/apache/spark/network/ConnectionManager.scala
----------------------------------------------------------------------
diff --git 
a/core/src/main/scala/org/apache/spark/network/ConnectionManager.scala 
b/core/src/main/scala/org/apache/spark/network/ConnectionManager.scala
index cfee41c..dcbbc18 100644
--- a/core/src/main/scala/org/apache/spark/network/ConnectionManager.scala
+++ b/core/src/main/scala/org/apache/spark/network/ConnectionManager.scala
@@ -33,6 +33,7 @@ import scala.collection.mutable.SynchronizedQueue
 
 import scala.concurrent.{Await, ExecutionContext, Future, Promise}
 import scala.concurrent.duration._
+import scala.language.postfixOps
 
 import org.apache.spark._
 import org.apache.spark.util.{SystemClock, Utils}

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/main/scala/org/apache/spark/rdd/CoGroupedRDD.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/rdd/CoGroupedRDD.scala 
b/core/src/main/scala/org/apache/spark/rdd/CoGroupedRDD.scala
index c6e7955..9ff7689 100644
--- a/core/src/main/scala/org/apache/spark/rdd/CoGroupedRDD.scala
+++ b/core/src/main/scala/org/apache/spark/rdd/CoGroupedRDD.scala
@@ -20,6 +20,7 @@ package org.apache.spark.rdd
 import java.io.{IOException, ObjectOutputStream}
 
 import scala.collection.mutable.ArrayBuffer
+import scala.language.existentials
 
 import org.apache.spark.{InterruptibleIterator, Partition, Partitioner, 
SparkEnv, TaskContext}
 import org.apache.spark.{Dependency, OneToOneDependency, ShuffleDependency}

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/main/scala/org/apache/spark/rdd/CoalescedRDD.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/rdd/CoalescedRDD.scala 
b/core/src/main/scala/org/apache/spark/rdd/CoalescedRDD.scala
index 44401a6..c45b759 100644
--- a/core/src/main/scala/org/apache/spark/rdd/CoalescedRDD.scala
+++ b/core/src/main/scala/org/apache/spark/rdd/CoalescedRDD.scala
@@ -21,6 +21,7 @@ import java.io.{IOException, ObjectOutputStream}
 
 import scala.collection.mutable
 import scala.collection.mutable.ArrayBuffer
+import scala.language.existentials
 import scala.reflect.ClassTag
 
 import org.apache.spark._

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/main/scala/org/apache/spark/scheduler/DAGSchedulerEvent.scala
----------------------------------------------------------------------
diff --git 
a/core/src/main/scala/org/apache/spark/scheduler/DAGSchedulerEvent.scala 
b/core/src/main/scala/org/apache/spark/scheduler/DAGSchedulerEvent.scala
index 7367c08..0800c56 100644
--- a/core/src/main/scala/org/apache/spark/scheduler/DAGSchedulerEvent.scala
+++ b/core/src/main/scala/org/apache/spark/scheduler/DAGSchedulerEvent.scala
@@ -20,6 +20,7 @@ package org.apache.spark.scheduler
 import java.util.Properties
 
 import scala.collection.mutable.Map
+import scala.language.existentials
 
 import org.apache.spark._
 import org.apache.spark.executor.TaskMetrics

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/main/scala/org/apache/spark/scheduler/ResultTask.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/scheduler/ResultTask.scala 
b/core/src/main/scala/org/apache/spark/scheduler/ResultTask.scala
index 0b38130..0e8d551 100644
--- a/core/src/main/scala/org/apache/spark/scheduler/ResultTask.scala
+++ b/core/src/main/scala/org/apache/spark/scheduler/ResultTask.scala
@@ -21,6 +21,7 @@ import java.io._
 import java.util.zip.{GZIPInputStream, GZIPOutputStream}
 
 import scala.collection.mutable.HashMap
+import scala.language.existentials
 
 import org.apache.spark._
 import org.apache.spark.rdd.{RDD, RDDCheckpointData}

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/main/scala/org/apache/spark/scheduler/ShuffleMapTask.scala
----------------------------------------------------------------------
diff --git 
a/core/src/main/scala/org/apache/spark/scheduler/ShuffleMapTask.scala 
b/core/src/main/scala/org/apache/spark/scheduler/ShuffleMapTask.scala
index 23f3b3e..02b62de 100644
--- a/core/src/main/scala/org/apache/spark/scheduler/ShuffleMapTask.scala
+++ b/core/src/main/scala/org/apache/spark/scheduler/ShuffleMapTask.scala
@@ -21,6 +21,7 @@ import java.io._
 import java.util.zip.{GZIPInputStream, GZIPOutputStream}
 
 import scala.collection.mutable.HashMap
+import scala.language.existentials
 
 import org.apache.spark._
 import org.apache.spark.executor.ShuffleWriteMetrics

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala
----------------------------------------------------------------------
diff --git 
a/core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala 
b/core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala
index acd152d..a3439b5 100644
--- a/core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala
+++ b/core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala
@@ -25,6 +25,7 @@ import scala.concurrent.duration._
 import scala.collection.mutable.ArrayBuffer
 import scala.collection.mutable.HashMap
 import scala.collection.mutable.HashSet
+import scala.language.postfixOps
 import scala.util.Random
 
 import org.apache.spark._

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/main/scala/org/apache/spark/ui/JettyUtils.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/ui/JettyUtils.scala 
b/core/src/main/scala/org/apache/spark/ui/JettyUtils.scala
index 3ae147a..750f5a5 100644
--- a/core/src/main/scala/org/apache/spark/ui/JettyUtils.scala
+++ b/core/src/main/scala/org/apache/spark/ui/JettyUtils.scala
@@ -22,6 +22,7 @@ import javax.servlet.DispatcherType
 import javax.servlet.http.{HttpServlet, HttpServletRequest, 
HttpServletResponse}
 
 import scala.annotation.tailrec
+import scala.language.implicitConversions
 import scala.util.{Failure, Success, Try}
 import scala.xml.Node
 

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/main/scala/org/apache/spark/util/TimeStampedWeakValueHashMap.scala
----------------------------------------------------------------------
diff --git 
a/core/src/main/scala/org/apache/spark/util/TimeStampedWeakValueHashMap.scala 
b/core/src/main/scala/org/apache/spark/util/TimeStampedWeakValueHashMap.scala
index b65017d..f5be585 100644
--- 
a/core/src/main/scala/org/apache/spark/util/TimeStampedWeakValueHashMap.scala
+++ 
b/core/src/main/scala/org/apache/spark/util/TimeStampedWeakValueHashMap.scala
@@ -21,6 +21,7 @@ import java.lang.ref.WeakReference
 import java.util.concurrent.atomic.AtomicInteger
 
 import scala.collection.mutable
+import scala.language.implicitConversions
 
 import org.apache.spark.Logging
 

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/main/scala/org/apache/spark/util/Vector.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/util/Vector.scala 
b/core/src/main/scala/org/apache/spark/util/Vector.scala
index 1a647fa..c6cab82 100644
--- a/core/src/main/scala/org/apache/spark/util/Vector.scala
+++ b/core/src/main/scala/org/apache/spark/util/Vector.scala
@@ -17,6 +17,7 @@
 
 package org.apache.spark.util
 
+import scala.language.implicitConversions
 import scala.util.Random
 
 import org.apache.spark.util.random.XORShiftRandom

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/test/scala/org/apache/spark/ConnectionManagerSuite.scala
----------------------------------------------------------------------
diff --git a/core/src/test/scala/org/apache/spark/ConnectionManagerSuite.scala 
b/core/src/test/scala/org/apache/spark/ConnectionManagerSuite.scala
index 80f7ec0..df6b260 100644
--- a/core/src/test/scala/org/apache/spark/ConnectionManagerSuite.scala
+++ b/core/src/test/scala/org/apache/spark/ConnectionManagerSuite.scala
@@ -25,7 +25,7 @@ import org.apache.spark.network.{ConnectionManager, Message, 
ConnectionManagerId
 import scala.concurrent.Await
 import scala.concurrent.TimeoutException
 import scala.concurrent.duration._
-
+import scala.language.postfixOps
 
 /**
   * Test the ConnectionManager with various security settings.

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/test/scala/org/apache/spark/ContextCleanerSuite.scala
----------------------------------------------------------------------
diff --git a/core/src/test/scala/org/apache/spark/ContextCleanerSuite.scala 
b/core/src/test/scala/org/apache/spark/ContextCleanerSuite.scala
index e50981c..5a83100 100644
--- a/core/src/test/scala/org/apache/spark/ContextCleanerSuite.scala
+++ b/core/src/test/scala/org/apache/spark/ContextCleanerSuite.scala
@@ -20,6 +20,8 @@ package org.apache.spark
 import java.lang.ref.WeakReference
 
 import scala.collection.mutable.{HashSet, SynchronizedSet}
+import scala.language.existentials
+import scala.language.postfixOps
 import scala.util.Random
 
 import org.scalatest.{BeforeAndAfter, FunSuite}

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/test/scala/org/apache/spark/DriverSuite.scala
----------------------------------------------------------------------
diff --git a/core/src/test/scala/org/apache/spark/DriverSuite.scala 
b/core/src/test/scala/org/apache/spark/DriverSuite.scala
index 7f59bdc..de4bd90 100644
--- a/core/src/test/scala/org/apache/spark/DriverSuite.scala
+++ b/core/src/test/scala/org/apache/spark/DriverSuite.scala
@@ -29,6 +29,8 @@ import org.scalatest.time.SpanSugar._
 
 import org.apache.spark.util.Utils
 
+import scala.language.postfixOps
+
 class DriverSuite extends FunSuite with Timeouts {
 
   test("driver should exit after finishing") {

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/test/scala/org/apache/spark/PipedRDDSuite.scala
----------------------------------------------------------------------
diff --git a/core/src/test/scala/org/apache/spark/PipedRDDSuite.scala 
b/core/src/test/scala/org/apache/spark/PipedRDDSuite.scala
index 867b28c..dfe0575 100644
--- a/core/src/test/scala/org/apache/spark/PipedRDDSuite.scala
+++ b/core/src/test/scala/org/apache/spark/PipedRDDSuite.scala
@@ -19,8 +19,6 @@ package org.apache.spark
 
 import java.io.File
 
-import com.google.common.io.Files
-
 import org.scalatest.FunSuite
 
 import org.apache.spark.rdd.{HadoopRDD, PipedRDD, HadoopPartition}
@@ -28,6 +26,7 @@ import org.apache.hadoop.mapred.{JobConf, TextInputFormat, 
FileSplit}
 import org.apache.hadoop.fs.Path
 
 import scala.collection.Map
+import scala.language.postfixOps
 import scala.sys.process._
 import scala.util.Try
 

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/test/scala/org/apache/spark/rdd/AsyncRDDActionsSuite.scala
----------------------------------------------------------------------
diff --git 
a/core/src/test/scala/org/apache/spark/rdd/AsyncRDDActionsSuite.scala 
b/core/src/test/scala/org/apache/spark/rdd/AsyncRDDActionsSuite.scala
index 3b833f2..2819765 100644
--- a/core/src/test/scala/org/apache/spark/rdd/AsyncRDDActionsSuite.scala
+++ b/core/src/test/scala/org/apache/spark/rdd/AsyncRDDActionsSuite.scala
@@ -22,6 +22,7 @@ import java.util.concurrent.Semaphore
 import scala.concurrent.{Await, TimeoutException}
 import scala.concurrent.duration.Duration
 import scala.concurrent.ExecutionContext.Implicits.global
+import scala.language.postfixOps
 
 import org.scalatest.{BeforeAndAfterAll, FunSuite}
 import org.scalatest.concurrent.Timeouts

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/test/scala/org/apache/spark/scheduler/DAGSchedulerSuite.scala
----------------------------------------------------------------------
diff --git 
a/core/src/test/scala/org/apache/spark/scheduler/DAGSchedulerSuite.scala 
b/core/src/test/scala/org/apache/spark/scheduler/DAGSchedulerSuite.scala
index db4df1d..35a7ac9 100644
--- a/core/src/test/scala/org/apache/spark/scheduler/DAGSchedulerSuite.scala
+++ b/core/src/test/scala/org/apache/spark/scheduler/DAGSchedulerSuite.scala
@@ -19,6 +19,7 @@ package org.apache.spark.scheduler
 
 import scala.Tuple2
 import scala.collection.mutable.{HashSet, HashMap, Map}
+import scala.language.reflectiveCalls
 
 import org.scalatest.{BeforeAndAfter, FunSuite}
 

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/test/scala/org/apache/spark/storage/BlockManagerSuite.scala
----------------------------------------------------------------------
diff --git 
a/core/src/test/scala/org/apache/spark/storage/BlockManagerSuite.scala 
b/core/src/test/scala/org/apache/spark/storage/BlockManagerSuite.scala
index e10ec7d..907428d 100644
--- a/core/src/test/scala/org/apache/spark/storage/BlockManagerSuite.scala
+++ b/core/src/test/scala/org/apache/spark/storage/BlockManagerSuite.scala
@@ -33,6 +33,9 @@ import org.apache.spark.scheduler.LiveListenerBus
 import org.apache.spark.serializer.{JavaSerializer, KryoSerializer}
 import org.apache.spark.util.{AkkaUtils, ByteBufferInputStream, SizeEstimator, 
Utils}
 
+import scala.language.implicitConversions
+import scala.language.postfixOps
+
 class BlockManagerSuite extends FunSuite with BeforeAndAfter with 
PrivateMethodTester {
   private val conf = new SparkConf(false)
   var store: BlockManager = null

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/test/scala/org/apache/spark/storage/DiskBlockManagerSuite.scala
----------------------------------------------------------------------
diff --git 
a/core/src/test/scala/org/apache/spark/storage/DiskBlockManagerSuite.scala 
b/core/src/test/scala/org/apache/spark/storage/DiskBlockManagerSuite.scala
index 808ddfd..9b29e2a 100644
--- a/core/src/test/scala/org/apache/spark/storage/DiskBlockManagerSuite.scala
+++ b/core/src/test/scala/org/apache/spark/storage/DiskBlockManagerSuite.scala
@@ -20,6 +20,7 @@ package org.apache.spark.storage
 import java.io.{File, FileWriter}
 
 import scala.collection.mutable
+import scala.language.reflectiveCalls
 
 import com.google.common.io.Files
 import org.scalatest.{BeforeAndAfterEach, FunSuite}

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/test/scala/org/apache/spark/ui/UISuite.scala
----------------------------------------------------------------------
diff --git a/core/src/test/scala/org/apache/spark/ui/UISuite.scala 
b/core/src/test/scala/org/apache/spark/ui/UISuite.scala
index b85c483..ed02b0b 100644
--- a/core/src/test/scala/org/apache/spark/ui/UISuite.scala
+++ b/core/src/test/scala/org/apache/spark/ui/UISuite.scala
@@ -21,6 +21,7 @@ import java.net.ServerSocket
 import javax.servlet.http.HttpServletRequest
 
 import scala.io.Source
+import scala.language.postfixOps
 import scala.util.{Failure, Success, Try}
 
 import org.eclipse.jetty.server.Server

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/core/src/test/scala/org/apache/spark/util/random/XORShiftRandomSuite.scala
----------------------------------------------------------------------
diff --git 
a/core/src/test/scala/org/apache/spark/util/random/XORShiftRandomSuite.scala 
b/core/src/test/scala/org/apache/spark/util/random/XORShiftRandomSuite.scala
index 39199a1..0865c63 100644
--- a/core/src/test/scala/org/apache/spark/util/random/XORShiftRandomSuite.scala
+++ b/core/src/test/scala/org/apache/spark/util/random/XORShiftRandomSuite.scala
@@ -22,6 +22,8 @@ import org.scalatest.matchers.ShouldMatchers
 
 import org.apache.spark.util.Utils.times
 
+import scala.language.reflectiveCalls
+
 class XORShiftRandomSuite extends FunSuite with ShouldMatchers {
 
   def fixture = new {

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/examples/src/main/scala/org/apache/spark/streaming/examples/ZeroMQWordCount.scala
----------------------------------------------------------------------
diff --git 
a/examples/src/main/scala/org/apache/spark/streaming/examples/ZeroMQWordCount.scala
 
b/examples/src/main/scala/org/apache/spark/streaming/examples/ZeroMQWordCount.scala
index 445d202..6f88db1 100644
--- 
a/examples/src/main/scala/org/apache/spark/streaming/examples/ZeroMQWordCount.scala
+++ 
b/examples/src/main/scala/org/apache/spark/streaming/examples/ZeroMQWordCount.scala
@@ -27,6 +27,8 @@ import org.apache.spark.streaming.{Seconds, StreamingContext}
 import org.apache.spark.streaming.StreamingContext._
 import org.apache.spark.streaming.zeromq._
 
+import scala.language.implicitConversions
+
 /**
  * A simple publisher for demonstration purposes, repeatedly publishes random 
Messages
  * every one second.

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/graphx/src/main/scala/org/apache/spark/graphx/Graph.scala
----------------------------------------------------------------------
diff --git a/graphx/src/main/scala/org/apache/spark/graphx/Graph.scala 
b/graphx/src/main/scala/org/apache/spark/graphx/Graph.scala
index ef05623..4534969 100644
--- a/graphx/src/main/scala/org/apache/spark/graphx/Graph.scala
+++ b/graphx/src/main/scala/org/apache/spark/graphx/Graph.scala
@@ -17,6 +17,7 @@
 
 package org.apache.spark.graphx
 
+import scala.language.implicitConversions
 import scala.reflect.ClassTag
 
 import org.apache.spark.graphx.impl._

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/graphx/src/main/scala/org/apache/spark/graphx/impl/MessageToPartition.scala
----------------------------------------------------------------------
diff --git 
a/graphx/src/main/scala/org/apache/spark/graphx/impl/MessageToPartition.scala 
b/graphx/src/main/scala/org/apache/spark/graphx/impl/MessageToPartition.scala
index 9d4f375..c45ba3d 100644
--- 
a/graphx/src/main/scala/org/apache/spark/graphx/impl/MessageToPartition.scala
+++ 
b/graphx/src/main/scala/org/apache/spark/graphx/impl/MessageToPartition.scala
@@ -17,6 +17,7 @@
 
 package org.apache.spark.graphx.impl
 
+import scala.language.implicitConversions
 import scala.reflect.{classTag, ClassTag}
 
 import org.apache.spark.Partitioner

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/graphx/src/main/scala/org/apache/spark/graphx/impl/Serializers.scala
----------------------------------------------------------------------
diff --git 
a/graphx/src/main/scala/org/apache/spark/graphx/impl/Serializers.scala 
b/graphx/src/main/scala/org/apache/spark/graphx/impl/Serializers.scala
index 2f2c524..2f0531e 100644
--- a/graphx/src/main/scala/org/apache/spark/graphx/impl/Serializers.scala
+++ b/graphx/src/main/scala/org/apache/spark/graphx/impl/Serializers.scala
@@ -20,10 +20,11 @@ package org.apache.spark.graphx.impl
 import java.io.{EOFException, InputStream, OutputStream}
 import java.nio.ByteBuffer
 
-import org.apache.spark.SparkConf
 import org.apache.spark.graphx._
 import org.apache.spark.serializer._
 
+import scala.language.existentials
+
 private[graphx]
 class VertexIdMsgSerializer extends Serializer with Serializable {
   override def newInstance(): SerializerInstance = new 
ShuffleSerializerInstance {

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/graphx/src/main/scala/org/apache/spark/graphx/util/BytecodeUtils.scala
----------------------------------------------------------------------
diff --git 
a/graphx/src/main/scala/org/apache/spark/graphx/util/BytecodeUtils.scala 
b/graphx/src/main/scala/org/apache/spark/graphx/util/BytecodeUtils.scala
index 087b115..be6b904 100644
--- a/graphx/src/main/scala/org/apache/spark/graphx/util/BytecodeUtils.scala
+++ b/graphx/src/main/scala/org/apache/spark/graphx/util/BytecodeUtils.scala
@@ -20,6 +20,7 @@ package org.apache.spark.graphx.util
 import java.io.{ByteArrayInputStream, ByteArrayOutputStream}
 
 import scala.collection.mutable.HashSet
+import scala.language.existentials
 
 import org.apache.spark.util.Utils
 

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/mllib/src/main/scala/org/apache/spark/mllib/rdd/RDDFunctions.scala
----------------------------------------------------------------------
diff --git a/mllib/src/main/scala/org/apache/spark/mllib/rdd/RDDFunctions.scala 
b/mllib/src/main/scala/org/apache/spark/mllib/rdd/RDDFunctions.scala
index 873de87..365b5e7 100644
--- a/mllib/src/main/scala/org/apache/spark/mllib/rdd/RDDFunctions.scala
+++ b/mllib/src/main/scala/org/apache/spark/mllib/rdd/RDDFunctions.scala
@@ -17,6 +17,7 @@
 
 package org.apache.spark.mllib.rdd
 
+import scala.language.implicitConversions
 import scala.reflect.ClassTag
 
 import org.apache.spark.rdd.RDD

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/mllib/src/main/scala/org/apache/spark/mllib/util/MFDataGenerator.scala
----------------------------------------------------------------------
diff --git 
a/mllib/src/main/scala/org/apache/spark/mllib/util/MFDataGenerator.scala 
b/mllib/src/main/scala/org/apache/spark/mllib/util/MFDataGenerator.scala
index 3f413fa..b76fbe8 100644
--- a/mllib/src/main/scala/org/apache/spark/mllib/util/MFDataGenerator.scala
+++ b/mllib/src/main/scala/org/apache/spark/mllib/util/MFDataGenerator.scala
@@ -17,6 +17,7 @@
 
 package org.apache.spark.mllib.util
 
+import scala.language.postfixOps
 import scala.util.Random
 
 import org.jblas.DoubleMatrix

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/mllib/src/test/java/org/apache/spark/mllib/linalg/JavaVectorsSuite.java
----------------------------------------------------------------------
diff --git 
a/mllib/src/test/java/org/apache/spark/mllib/linalg/JavaVectorsSuite.java 
b/mllib/src/test/java/org/apache/spark/mllib/linalg/JavaVectorsSuite.java
index c6d8425..1421067 100644
--- a/mllib/src/test/java/org/apache/spark/mllib/linalg/JavaVectorsSuite.java
+++ b/mllib/src/test/java/org/apache/spark/mllib/linalg/JavaVectorsSuite.java
@@ -36,6 +36,7 @@ public class JavaVectorsSuite implements Serializable {
 
   @Test
   public void sparseArrayConstruction() {
+    @SuppressWarnings("unchecked")
     Vector v = Vectors.sparse(3, Lists.<Tuple2<Integer, Double>>newArrayList(
         new Tuple2<Integer, Double>(0, 2.0),
         new Tuple2<Integer, Double>(2, 3.0)));

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/pom.xml
----------------------------------------------------------------------
diff --git a/pom.xml b/pom.xml
index 5f66cbe..0eacedf 100644
--- a/pom.xml
+++ b/pom.xml
@@ -648,6 +648,7 @@
             <args>
               <arg>-unchecked</arg>
               <arg>-deprecation</arg>
+              <arg>-feature</arg>
             </args>
             <jvmArgs>
               <jvmArg>-Xms1024m</jvmArg>

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/project/SparkBuild.scala
----------------------------------------------------------------------
diff --git a/project/SparkBuild.scala b/project/SparkBuild.scala
index a6058bb..aac07b9 100644
--- a/project/SparkBuild.scala
+++ b/project/SparkBuild.scala
@@ -162,7 +162,7 @@ object SparkBuild extends Build {
     organization       := "org.apache.spark",
     version            := SPARK_VERSION,
     scalaVersion       := "2.10.4",
-    scalacOptions := Seq("-Xmax-classfile-name", "120", "-unchecked", 
"-deprecation",
+    scalacOptions := Seq("-Xmax-classfile-name", "120", "-unchecked", 
"-deprecation", "-feature",
       "-target:" + SCALAC_JVM_VERSION),
     javacOptions := Seq("-target", JAVAC_JVM_VERSION, "-source", 
JAVAC_JVM_VERSION),
     unmanagedJars in Compile <<= baseDirectory map { base => (base / "lib" ** 
"*.jar").classpath },

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/SqlParser.scala
----------------------------------------------------------------------
diff --git 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/SqlParser.scala 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/SqlParser.scala
index 5b6aea8..13a19d0 100644
--- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/SqlParser.scala
+++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/SqlParser.scala
@@ -17,6 +17,7 @@
 
 package org.apache.spark.sql.catalyst
 
+import scala.language.implicitConversions
 import scala.util.parsing.combinator.lexical.StdLexical
 import scala.util.parsing.combinator.syntactical.StandardTokenParsers
 import scala.util.parsing.input.CharArrayReader.EofCh

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/streaming/src/main/scala/org/apache/spark/streaming/StreamingContext.scala
----------------------------------------------------------------------
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/StreamingContext.scala 
b/streaming/src/main/scala/org/apache/spark/streaming/StreamingContext.scala
index ff5d0aa..e9a4f7b 100644
--- a/streaming/src/main/scala/org/apache/spark/streaming/StreamingContext.scala
+++ b/streaming/src/main/scala/org/apache/spark/streaming/StreamingContext.scala
@@ -22,6 +22,7 @@ import java.util.concurrent.atomic.AtomicInteger
 
 import scala.collection.Map
 import scala.collection.mutable.Queue
+import scala.language.implicitConversions
 import scala.reflect.ClassTag
 
 import akka.actor.{Props, SupervisorStrategy}

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaDStream.scala
----------------------------------------------------------------------
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaDStream.scala
 
b/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaDStream.scala
index 721d502..13e2bac 100644
--- 
a/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaDStream.scala
+++ 
b/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaDStream.scala
@@ -23,6 +23,7 @@ import org.apache.spark.api.java.JavaRDD
 import org.apache.spark.storage.StorageLevel
 import org.apache.spark.rdd.RDD
 
+import scala.language.implicitConversions
 import scala.reflect.ClassTag
 import org.apache.spark.streaming.dstream.DStream
 

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaDStreamLike.scala
----------------------------------------------------------------------
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaDStreamLike.scala
 
b/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaDStreamLike.scala
index bb2f492..a6184de 100644
--- 
a/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaDStreamLike.scala
+++ 
b/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaDStreamLike.scala
@@ -22,6 +22,7 @@ import java.lang.{Long => JLong}
 import java.util.{List => JList}
 
 import scala.collection.JavaConversions._
+import scala.language.implicitConversions
 import scala.reflect.ClassTag
 
 import org.apache.spark.api.java.{JavaPairRDD, JavaRDD, JavaRDDLike}

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaPairDStream.scala
----------------------------------------------------------------------
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaPairDStream.scala
 
b/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaPairDStream.scala
index 2ac943d..cb8e8f0 100644
--- 
a/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaPairDStream.scala
+++ 
b/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaPairDStream.scala
@@ -21,6 +21,7 @@ import java.lang.{Long => JLong, Iterable => JIterable}
 import java.util.{List => JList}
 
 import scala.collection.JavaConversions._
+import scala.language.implicitConversions
 import scala.reflect.ClassTag
 
 import com.google.common.base.Optional

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/streaming/src/main/scala/org/apache/spark/streaming/receivers/ActorReceiver.scala
----------------------------------------------------------------------
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/receivers/ActorReceiver.scala
 
b/streaming/src/main/scala/org/apache/spark/streaming/receivers/ActorReceiver.scala
index f5984d0..da0d364 100644
--- 
a/streaming/src/main/scala/org/apache/spark/streaming/receivers/ActorReceiver.scala
+++ 
b/streaming/src/main/scala/org/apache/spark/streaming/receivers/ActorReceiver.scala
@@ -23,6 +23,7 @@ import akka.actor.{ PossiblyHarmful, OneForOneStrategy }
 import akka.actor.SupervisorStrategy._
 
 import scala.concurrent.duration._
+import scala.language.postfixOps
 import scala.reflect.ClassTag
 
 import org.apache.spark.storage.{StorageLevel, StreamBlockId}

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/streaming/src/main/scala/org/apache/spark/streaming/scheduler/NetworkInputTracker.scala
----------------------------------------------------------------------
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/scheduler/NetworkInputTracker.scala
 
b/streaming/src/main/scala/org/apache/spark/streaming/scheduler/NetworkInputTracker.scala
index a1e6f51..438e72a 100644
--- 
a/streaming/src/main/scala/org/apache/spark/streaming/scheduler/NetworkInputTracker.scala
+++ 
b/streaming/src/main/scala/org/apache/spark/streaming/scheduler/NetworkInputTracker.scala
@@ -18,6 +18,7 @@
 package org.apache.spark.streaming.scheduler
 
 import scala.collection.mutable.{HashMap, SynchronizedMap, SynchronizedQueue}
+import scala.language.existentials
 
 import akka.actor._
 

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/streaming/src/test/scala/org/apache/spark/streaming/StreamingContextSuite.scala
----------------------------------------------------------------------
diff --git 
a/streaming/src/test/scala/org/apache/spark/streaming/StreamingContextSuite.scala
 
b/streaming/src/test/scala/org/apache/spark/streaming/StreamingContextSuite.scala
index efd0d22..ad5367a 100644
--- 
a/streaming/src/test/scala/org/apache/spark/streaming/StreamingContextSuite.scala
+++ 
b/streaming/src/test/scala/org/apache/spark/streaming/StreamingContextSuite.scala
@@ -28,6 +28,8 @@ import org.scalatest.concurrent.Timeouts
 import org.scalatest.exceptions.TestFailedDueToTimeoutException
 import org.scalatest.time.SpanSugar._
 
+import scala.language.postfixOps
+
 class StreamingContextSuite extends FunSuite with BeforeAndAfter with Timeouts 
with Logging {
 
   val master = "local[2]"

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/streaming/src/test/scala/org/apache/spark/streaming/UISuite.scala
----------------------------------------------------------------------
diff --git a/streaming/src/test/scala/org/apache/spark/streaming/UISuite.scala 
b/streaming/src/test/scala/org/apache/spark/streaming/UISuite.scala
index 35538ec..031e93a 100644
--- a/streaming/src/test/scala/org/apache/spark/streaming/UISuite.scala
+++ b/streaming/src/test/scala/org/apache/spark/streaming/UISuite.scala
@@ -18,6 +18,7 @@
 package org.apache.spark.streaming
 
 import scala.io.Source
+import scala.language.postfixOps
 
 import org.scalatest.FunSuite
 import org.scalatest.concurrent.Eventually._

http://git-wip-us.apache.org/repos/asf/spark/blob/74718285/tools/src/main/scala/org/apache/spark/tools/JavaAPICompletenessChecker.scala
----------------------------------------------------------------------
diff --git 
a/tools/src/main/scala/org/apache/spark/tools/JavaAPICompletenessChecker.scala 
b/tools/src/main/scala/org/apache/spark/tools/JavaAPICompletenessChecker.scala
index 4205224..8d0f099 100644
--- 
a/tools/src/main/scala/org/apache/spark/tools/JavaAPICompletenessChecker.scala
+++ 
b/tools/src/main/scala/org/apache/spark/tools/JavaAPICompletenessChecker.scala
@@ -20,6 +20,7 @@ package org.apache.spark.tools
 import java.lang.reflect.Method
 
 import scala.collection.mutable.ArrayBuffer
+import scala.language.existentials
 
 import org.apache.spark._
 import org.apache.spark.api.java._

Reply via email to