+1 Tested on Mac OS X, and verified that sort-based shuffle bug is fixed.
Matei > On Nov 14, 2014, at 10:45 AM, Andrew Or <and...@databricks.com> wrote: > > Hi all, since the vote ends on a Sunday, please let me know if you would > like to extend the deadline to allow more time for testing. > > 2014-11-13 12:10 GMT-08:00 Sean Owen <so...@cloudera.com>: > >> Ah right. This is because I'm running Java 8. This was fixed in >> SPARK-3329 ( >> https://github.com/apache/spark/commit/2b7ab814f9bde65ebc57ebd04386e56c97f06f4a#diff-7bfd8d7c8cbb02aa0023e4c3497ee832 >> ). >> Consider back-porting it if other reasons arise, but this is specific >> to tests and to Java 8. >> >> On Thu, Nov 13, 2014 at 8:01 PM, Andrew Or <and...@databricks.com> wrote: >>> Yeah, this seems to be somewhat environment specific too. The same test >> has >>> been passing here for a while: >>> >> https://amplab.cs.berkeley.edu/jenkins/job/Spark-1.1-Maven-pre-YARN/hadoop.version=1.0.4,label=centos/lastBuild/consoleFull >>> >>> 2014-11-13 11:26 GMT-08:00 Michael Armbrust <mich...@databricks.com>: >>> >>>> Hey Sean, >>>> >>>> Thanks for pointing this out. Looks like a bad test where we should be >>>> doing Set comparison instead of Array. >>>> >>>> Michael >>>> >>>> On Thu, Nov 13, 2014 at 2:05 AM, Sean Owen <so...@cloudera.com> wrote: >>>>> >>>>> LICENSE and NOTICE are fine. Signature and checksum is fine. I >>>>> unzipped and built the plain source distribution, which built. >>>>> >>>>> However I am seeing a consistent test failure with "mvn -DskipTests >>>>> clean package; mvn test". In the Hive module: >>>>> >>>>> - SET commands semantics for a HiveContext *** FAILED *** >>>>> Expected Array("spark.sql.key.usedfortestonly=test.val.0", >>>>> >>>>> >> "spark.sql.key.usedfortestonlyspark.sql.key.usedfortestonly=test.val.0test.val.0"), >>>>> but got >>>>> >> Array("spark.sql.key.usedfortestonlyspark.sql.key.usedfortestonly=test.val.0test.val.0", >>>>> "spark.sql.key.usedfortestonly=test.val.0") (HiveQuerySuite.scala:544) >>>>> >>>>> Anyone else seeing this? >>>>> >>>>> >>>>> On Thu, Nov 13, 2014 at 8:18 AM, Krishna Sankar <ksanka...@gmail.com> >>>>> wrote: >>>>>> +1 >>>>>> 1. Compiled OSX 10.10 (Yosemite) mvn -Pyarn -Phadoop-2.4 >>>>>> -Dhadoop.version=2.4.0 -DskipTests clean package 10:49 min >>>>>> 2. Tested pyspark, mlib >>>>>> 2.1. statistics OK >>>>>> 2.2. Linear/Ridge/Laso Regression OK >>>>>> 2.3. Decision Tree, Naive Bayes OK >>>>>> 2.4. KMeans OK >>>>>> 2.5. rdd operations OK >>>>>> 2.6. recommendation OK >>>>>> 2.7. Good work ! In 1.1.0, there was an error and my program used to >>>>>> hang >>>>>> (over memory allocation) consistently running validation using >>>>>> itertools, >>>>>> compute optimum rank, lambda,numofiterations/rmse; data - movielens >>>>>> medium >>>>>> dataset (1 million records) . It works well in 1.1.1 ! >>>>>> Cheers >>>>>> <k/> >>>>>> P.S: Missed Reply all, first time >>>>>> >>>>>> On Wed, Nov 12, 2014 at 8:35 PM, Andrew Or <and...@databricks.com> >>>>>> wrote: >>>>>> >>>>>>> I will start the vote with a +1 >>>>>>> >>>>>>> 2014-11-12 20:34 GMT-08:00 Andrew Or <and...@databricks.com>: >>>>>>> >>>>>>>> Please vote on releasing the following candidate as Apache Spark >>>>>>>> version >>>>>>> 1 >>>>>>>> .1.1. >>>>>>>> >>>>>>>> This release fixes a number of bugs in Spark 1.1.0. Some of the >>>>>>>> notable >>>>>>>> ones are >>>>>>>> - [SPARK-3426] Sort-based shuffle compression settings are >>>>>>>> incompatible >>>>>>>> - [SPARK-3948] Stream corruption issues in sort-based shuffle >>>>>>>> - [SPARK-4107] Incorrect handling of Channel.read() led to data >>>>>>> truncation >>>>>>>> The full list is at http://s.apache.org/z9h and in the >> CHANGES.txt >>>>>>>> attached. >>>>>>>> >>>>>>>> The tag to be voted on is v1.1.1-rc1 (commit 72a4fdbe): >>>>>>>> http://s.apache.org/cZC >>>>>>>> >>>>>>>> The release files, including signatures, digests, etc can be found >>>>>>>> at: >>>>>>>> http://people.apache.org/~andrewor14/spark-1.1.1-rc1/ >>>>>>>> >>>>>>>> Release artifacts are signed with the following key: >>>>>>>> https://people.apache.org/keys/committer/andrewor14.asc >>>>>>>> >>>>>>>> The staging repository for this release can be found at: >>>>>>>> >>>>>>>> >> https://repository.apache.org/content/repositories/orgapachespark-1034/ >>>>>>>> >>>>>>>> The documentation corresponding to this release can be found at: >>>>>>>> http://people.apache.org/~andrewor14/spark-1.1.1-rc1-docs/ >>>>>>>> >>>>>>>> Please vote on releasing this package as Apache Spark 1.1.1! >>>>>>>> >>>>>>>> The vote is open until Sunday, November 16, at 04:30 UTC and >> passes >>>>>>>> if >>>>>>>> a majority of at least 3 +1 PMC votes are cast. >>>>>>>> [ ] +1 Release this package as Apache Spark 1.1.1 >>>>>>>> [ ] -1 Do not release this package because ... >>>>>>>> >>>>>>>> To learn more about Apache Spark, please see >>>>>>>> http://spark.apache.org/ >>>>>>>> >>>>>>>> Cheers, >>>>>>>> Andrew >>>>>>>> >>>>>>> >>>>> >>>>> --------------------------------------------------------------------- >>>>> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org >>>>> For additional commands, e-mail: dev-h...@spark.apache.org >>>>> >>>> >>> >> --------------------------------------------------------------------- To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org For additional commands, e-mail: dev-h...@spark.apache.org