git commit: [SPARK-959] Updated SBT from 0.13.1 to 0.13.2

2014-04-16 Thread pwendell
Repository: spark Updated Branches: refs/heads/master 273c2fd08 - 6a10d8016 [SPARK-959] Updated SBT from 0.13.1 to 0.13.2 JIRA issue: [SPARK-959](https://spark-project.atlassian.net/browse/SPARK-959) SBT 0.13.2 has been officially released. This version updated Ivy 2.0 to Ivy 2.3, which

git commit: [SPARK-959] Updated SBT from 0.13.1 to 0.13.2

2014-04-16 Thread pwendell
Repository: spark Updated Branches: refs/heads/branch-1.0 e5130d978 - 1ea9a21f4 [SPARK-959] Updated SBT from 0.13.1 to 0.13.2 JIRA issue: [SPARK-959](https://spark-project.atlassian.net/browse/SPARK-959) SBT 0.13.2 has been officially released. This version updated Ivy 2.0 to Ivy 2.3, which

git commit: Loads test tables when running sbt hive/console without HIVE_DEV_HOME

2014-04-16 Thread pwendell
Repository: spark Updated Branches: refs/heads/branch-1.0 5fe18a74f - 9e908ab2e Loads test tables when running sbt hive/console without HIVE_DEV_HOME When running Hive tests, the working directory is `$SPARK_HOME/sql/hive`, while when running `sbt hive/console`, it becomes `$SPARK_HOME`, and

git commit: Loads test tables when running sbt hive/console without HIVE_DEV_HOME

2014-04-16 Thread pwendell
Repository: spark Updated Branches: refs/heads/master c0273d806 - fec462c15 Loads test tables when running sbt hive/console without HIVE_DEV_HOME When running Hive tests, the working directory is `$SPARK_HOME/sql/hive`, while when running `sbt hive/console`, it becomes `$SPARK_HOME`, and

git commit: update spark.default.parallelism

2014-04-16 Thread pwendell
Repository: spark Updated Branches: refs/heads/branch-1.0 9e908ab2e - e4f5577e2 update spark.default.parallelism actually, the value 8 is only valid in mesos fine-grained mode : code override def defaultParallelism() = sc.conf.getInt(spark.default.parallelism, 8) /code while in

git commit: SPARK-1469: Scheduler mode should accept lower-case definitions and have...

2014-04-16 Thread pwendell
Repository: spark Updated Branches: refs/heads/master 82349fbd2 - e269c24db SPARK-1469: Scheduler mode should accept lower-case definitions and have... ... nicer error messages There are two improvements to Scheduler Mode: 1. Made the built in ones case insensitive (fair/FAIR, fifo/FIFO).

git commit: SPARK-1469: Scheduler mode should accept lower-case definitions and have...

2014-04-16 Thread pwendell
Repository: spark Updated Branches: refs/heads/branch-1.0 4479ecd08 - b75301f1f SPARK-1469: Scheduler mode should accept lower-case definitions and have... ... nicer error messages There are two improvements to Scheduler Mode: 1. Made the built in ones case insensitive (fair/FAIR,

git commit: SPARK-1465: Spark compilation is broken with the latest hadoop-2.4.0 release

2014-04-16 Thread tgraves
Repository: spark Updated Branches: refs/heads/master e269c24db - 725925cf2 SPARK-1465: Spark compilation is broken with the latest hadoop-2.4.0 release YARN-1824 changes the APIs (addToEnvironment, setEnvFromInputString) in Apps, which causes the spark build to break if built against a

git commit: SPARK-1465: Spark compilation is broken with the latest hadoop-2.4.0 release

2014-04-16 Thread tgraves
Repository: spark Updated Branches: refs/heads/branch-1.0 b75301f1f - d8fc4a4b6 SPARK-1465: Spark compilation is broken with the latest hadoop-2.4.0 release YARN-1824 changes the APIs (addToEnvironment, setEnvFromInputString) in Apps, which causes the spark build to break if built against a

git commit: Rebuild routing table after Graph.reverse

2014-04-16 Thread rxin
Repository: spark Updated Branches: refs/heads/master 987760ec0 - 235a47ce1 Rebuild routing table after Graph.reverse GraphImpl.reverse used to reverse edges in each partition of the edge RDD but preserve the routing table and replicated vertex view, since reversing should not affect

git commit: SPARK-1329: Create pid2vid with correct number of partitions

2014-04-16 Thread rxin
Repository: spark Updated Branches: refs/heads/branch-1.0 602b9ea65 - b4ea3d972 SPARK-1329: Create pid2vid with correct number of partitions Each vertex partition is co-located with a pid2vid array created in RoutingTable.scala. This array maps edge partition IDs to the list of vertices in

git commit: remove unnecessary brace and semicolon in 'putBlockInfo.synchronize' block

2014-04-16 Thread rxin
Repository: spark Updated Branches: refs/heads/master 17d323455 - 016a87764 remove unnecessary brace and semicolon in 'putBlockInfo.synchronize' block delete semicolon Author: Chen Chao crazy...@gmail.com Closes #411 from CrazyJvm/patch-5 and squashes the following commits: 72333a3 [Chen

git commit: remove unnecessary brace and semicolon in 'putBlockInfo.synchronize' block

2014-04-16 Thread rxin
Repository: spark Updated Branches: refs/heads/branch-1.0 b4ea3d972 - e43e31ded remove unnecessary brace and semicolon in 'putBlockInfo.synchronize' block delete semicolon Author: Chen Chao crazy...@gmail.com Closes #411 from CrazyJvm/patch-5 and squashes the following commits: 72333a3

git commit: Fixing a race condition in event listener unit test

2014-04-16 Thread rxin
Repository: spark Updated Branches: refs/heads/master 016a87764 - 38877ccf3 Fixing a race condition in event listener unit test Author: Kan Zhang kzh...@apache.org Closes #401 from kanzhang/fix-1475 and squashes the following commits: c6058bd [Kan Zhang] Fixing a race condition in event

git commit: misleading task number of groupByKey

2014-04-16 Thread rxin
Repository: spark Updated Branches: refs/heads/branch-1.0 f0abf5f08 - 51c41da51 misleading task number of groupByKey By default, this uses only 8 parallel tasks to do the grouping. is a big misleading. Please refer to https://github.com/apache/spark/pull/389 detail is as following code :

git commit: misleading task number of groupByKey

2014-04-16 Thread rxin
Repository: spark Updated Branches: refs/heads/master 38877ccf3 - 9c40b9ead misleading task number of groupByKey By default, this uses only 8 parallel tasks to do the grouping. is a big misleading. Please refer to https://github.com/apache/spark/pull/389 detail is as following code : def

git commit: Update ReducedWindowedDStream.scala

2014-04-16 Thread rxin
Repository: spark Updated Branches: refs/heads/branch-1.0 51c41da51 - 822353dc5 Update ReducedWindowedDStream.scala change _slideDuration to _windowDuration Author: baishuo(白硕) vc_j...@hotmail.com Closes #425 from baishuo/master and squashes the following commits: 6f09ea1

git commit: Include stack trace for exceptions thrown by user code.

2014-04-16 Thread rxin
Repository: spark Updated Branches: refs/heads/branch-1.0 822353dc5 - aef8a4a51 Include stack trace for exceptions thrown by user code. It is very confusing when your code throws an exception, but the only stack trace show is in the DAGScheduler. This is a simple patch to include the stack

git commit: SPARK-1462: Examples of ML algorithms are using deprecated APIs

2014-04-16 Thread matei
Repository: spark Updated Branches: refs/heads/master d4916a8ee - 6ad4c5498 SPARK-1462: Examples of ML algorithms are using deprecated APIs This will also fix SPARK-1464: Update MLLib Examples to Use Breeze. Author: Sandeep sand...@techaddict.me Closes #416 from techaddict/1462 and squashes

git commit: SPARK-1462: Examples of ML algorithms are using deprecated APIs

2014-04-16 Thread matei
Repository: spark Updated Branches: refs/heads/branch-1.0 aef8a4a51 - 13fb4c782 SPARK-1462: Examples of ML algorithms are using deprecated APIs This will also fix SPARK-1464: Update MLLib Examples to Use Breeze. Author: Sandeep sand...@techaddict.me Closes #416 from techaddict/1462 and

git commit: [python alternative] pyspark require Python2, failing if system default is Py3 from shell.py

2014-04-16 Thread rxin
Repository: spark Updated Branches: refs/heads/master 6ad4c5498 - bb76eae1b [python alternative] pyspark require Python2, failing if system default is Py3 from shell.py Python alternative for https://github.com/apache/spark/pull/392; managed from shell.py Author: AbhishekKr

git commit: [python alternative] pyspark require Python2, failing if system default is Py3 from shell.py

2014-04-16 Thread rxin
Repository: spark Updated Branches: refs/heads/branch-1.0 13fb4c782 - b3ad707c4 [python alternative] pyspark require Python2, failing if system default is Py3 from shell.py Python alternative for https://github.com/apache/spark/pull/392; managed from shell.py Author: AbhishekKr