Repository: spark Updated Branches: refs/heads/branch-1.0 775020f00 -> b0ded1f6f
Handle the vals that never used In XORShiftRandom.scala, use val "million" instead of constant "1e6.toInt". Delete vals that never used in other files. Author: WangTao <[email protected]> Closes #565 from WangTaoTheTonic/master and squashes the following commits: 17cacfc [WangTao] Handle the unused assignment, method parameters and symbol inspected by Intellij IDEA 37b4090 [WangTao] Handle the vals that never used (cherry picked from commit 7025dda8fa84b57d6f12bc770df2fa10eef21d88) Signed-off-by: Reynold Xin <[email protected]> Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/b0ded1f6 Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/b0ded1f6 Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/b0ded1f6 Branch: refs/heads/branch-1.0 Commit: b0ded1f6f23f326b6b7a8d3b118cd03a5476655a Parents: 775020f Author: WangTao <[email protected]> Authored: Tue Apr 29 22:07:20 2014 -0700 Committer: Reynold Xin <[email protected]> Committed: Tue Apr 29 22:07:27 2014 -0700 ---------------------------------------------------------------------- core/src/main/scala/org/apache/spark/network/SecurityMessage.scala | 1 - .../main/scala/org/apache/spark/partial/GroupedMeanEvaluator.scala | 1 - .../main/scala/org/apache/spark/storage/BlockFetcherIterator.scala | 2 -- .../main/scala/org/apache/spark/util/random/XORShiftRandom.scala | 2 +- examples/src/main/java/org/apache/spark/examples/JavaLogQuery.java | 1 - examples/src/main/java/org/apache/spark/examples/JavaTC.java | 2 +- .../src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java | 1 - 7 files changed, 2 insertions(+), 8 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/spark/blob/b0ded1f6/core/src/main/scala/org/apache/spark/network/SecurityMessage.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/network/SecurityMessage.scala b/core/src/main/scala/org/apache/spark/network/SecurityMessage.scala index a1dfc40..9af9e2e 100644 --- a/core/src/main/scala/org/apache/spark/network/SecurityMessage.scala +++ b/core/src/main/scala/org/apache/spark/network/SecurityMessage.scala @@ -106,7 +106,6 @@ private[spark] class SecurityMessage() extends Logging { * @return BufferMessage */ def toBufferMessage: BufferMessage = { - val startTime = System.currentTimeMillis val buffers = new ArrayBuffer[ByteBuffer]() // 4 bytes for the length of the connectionId http://git-wip-us.apache.org/repos/asf/spark/blob/b0ded1f6/core/src/main/scala/org/apache/spark/partial/GroupedMeanEvaluator.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/partial/GroupedMeanEvaluator.scala b/core/src/main/scala/org/apache/spark/partial/GroupedMeanEvaluator.scala index b511189..af26c3d 100644 --- a/core/src/main/scala/org/apache/spark/partial/GroupedMeanEvaluator.scala +++ b/core/src/main/scala/org/apache/spark/partial/GroupedMeanEvaluator.scala @@ -61,7 +61,6 @@ private[spark] class GroupedMeanEvaluator[T](totalOutputs: Int, confidence: Doub } else if (outputsMerged == 0) { new HashMap[T, BoundedDouble] } else { - val p = outputsMerged.toDouble / totalOutputs val studentTCacher = new StudentTCacher(confidence) val result = new JHashMap[T, BoundedDouble](sums.size) val iter = sums.entrySet.iterator() http://git-wip-us.apache.org/repos/asf/spark/blob/b0ded1f6/core/src/main/scala/org/apache/spark/storage/BlockFetcherIterator.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/storage/BlockFetcherIterator.scala b/core/src/main/scala/org/apache/spark/storage/BlockFetcherIterator.scala index a02dd94..408a797 100644 --- a/core/src/main/scala/org/apache/spark/storage/BlockFetcherIterator.scala +++ b/core/src/main/scala/org/apache/spark/storage/BlockFetcherIterator.scala @@ -118,11 +118,9 @@ object BlockFetcherIterator { }) bytesInFlight += req.size val sizeMap = req.blocks.toMap // so we can look up the size of each blockID - val fetchStart = System.currentTimeMillis() val future = connectionManager.sendMessageReliably(cmId, blockMessageArray.toBufferMessage) future.onSuccess { case Some(message) => { - val fetchDone = System.currentTimeMillis() val bufferMessage = message.asInstanceOf[BufferMessage] val blockMessageArray = BlockMessageArray.fromBufferMessage(bufferMessage) for (blockMessage <- blockMessageArray) { http://git-wip-us.apache.org/repos/asf/spark/blob/b0ded1f6/core/src/main/scala/org/apache/spark/util/random/XORShiftRandom.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/util/random/XORShiftRandom.scala b/core/src/main/scala/org/apache/spark/util/random/XORShiftRandom.scala index 7f22038..55b5713 100644 --- a/core/src/main/scala/org/apache/spark/util/random/XORShiftRandom.scala +++ b/core/src/main/scala/org/apache/spark/util/random/XORShiftRandom.scala @@ -91,7 +91,7 @@ private[spark] object XORShiftRandom { val xorRand = new XORShiftRandom(seed) // this is just to warm up the JIT - we're not timing anything - timeIt(1e6.toInt) { + timeIt(million) { javaRand.nextInt() xorRand.nextInt() } http://git-wip-us.apache.org/repos/asf/spark/blob/b0ded1f6/examples/src/main/java/org/apache/spark/examples/JavaLogQuery.java ---------------------------------------------------------------------- diff --git a/examples/src/main/java/org/apache/spark/examples/JavaLogQuery.java b/examples/src/main/java/org/apache/spark/examples/JavaLogQuery.java index 2a4278d..3f7a879 100644 --- a/examples/src/main/java/org/apache/spark/examples/JavaLogQuery.java +++ b/examples/src/main/java/org/apache/spark/examples/JavaLogQuery.java @@ -75,7 +75,6 @@ public final class JavaLogQuery { public static Tuple3<String, String, String> extractKey(String line) { Matcher m = apacheLogRegex.matcher(line); - List<String> key = Collections.emptyList(); if (m.find()) { String ip = m.group(1); String user = m.group(3); http://git-wip-us.apache.org/repos/asf/spark/blob/b0ded1f6/examples/src/main/java/org/apache/spark/examples/JavaTC.java ---------------------------------------------------------------------- diff --git a/examples/src/main/java/org/apache/spark/examples/JavaTC.java b/examples/src/main/java/org/apache/spark/examples/JavaTC.java index 1d77694..d66b9ba 100644 --- a/examples/src/main/java/org/apache/spark/examples/JavaTC.java +++ b/examples/src/main/java/org/apache/spark/examples/JavaTC.java @@ -85,7 +85,7 @@ public final class JavaTC { } }); - long oldCount = 0; + long oldCount; long nextCount = tc.count(); do { oldCount = nextCount; http://git-wip-us.apache.org/repos/asf/spark/blob/b0ded1f6/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java ---------------------------------------------------------------------- diff --git a/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java b/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java index b5b438e..d62a72f 100644 --- a/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java +++ b/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java @@ -23,7 +23,6 @@ import java.util.List; import org.apache.spark.api.java.JavaRDD; import org.apache.spark.api.java.JavaSparkContext; import org.apache.spark.api.java.function.Function; -import org.apache.spark.api.java.function.VoidFunction; import org.apache.spark.sql.api.java.JavaSQLContext; import org.apache.spark.sql.api.java.JavaSchemaRDD;
