Repository: spark Updated Branches: refs/heads/master 04b01bb10 -> cb09e93c1
Reformat multi-line closure argument. Author: William Benton <wi...@redhat.com> Closes #1419 from willb/reformat-2486 and squashes the following commits: 2676231 [William Benton] Reformat multi-line closure argument. Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/cb09e93c Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/cb09e93c Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/cb09e93c Branch: refs/heads/master Commit: cb09e93c1d7ef9c8f0a1abe4e659783c74993a4e Parents: 04b01bb Author: William Benton <wi...@redhat.com> Authored: Tue Jul 15 09:13:39 2014 -0700 Committer: Aaron Davidson <aa...@databricks.com> Committed: Tue Jul 15 09:13:39 2014 -0700 ---------------------------------------------------------------------- core/src/main/scala/org/apache/spark/util/Utils.scala | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/spark/blob/cb09e93c/core/src/main/scala/org/apache/spark/util/Utils.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/util/Utils.scala b/core/src/main/scala/org/apache/spark/util/Utils.scala index d72c97b..10c33d6 100644 --- a/core/src/main/scala/org/apache/spark/util/Utils.scala +++ b/core/src/main/scala/org/apache/spark/util/Utils.scala @@ -809,11 +809,12 @@ private[spark] object Utils extends Logging { */ def getCallSite: CallSite = { val trace = Thread.currentThread.getStackTrace() - .filterNot((ste:StackTraceElement) => + .filterNot { ste:StackTraceElement => // When running under some profilers, the current stack trace might contain some bogus // frames. This is intended to ensure that we don't crash in these situations by // ignoring any frames that we can't examine. - (ste == null || ste.getMethodName == null || ste.getMethodName.contains("getStackTrace"))) + (ste == null || ste.getMethodName == null || ste.getMethodName.contains("getStackTrace")) + } // Keep crawling up the stack trace until we find the first function not inside of the spark // package. We track the last (shallowest) contiguous Spark method. This might be an RDD