I took the always fun task of testing it on Windows, and unfortunately, I found some small problems with the prebuilt packages due to recent changes to the launch scripts: bin/spark-class2.cmd looks in ./jars instead of ./lib for the assembly JAR, and bin/run-example2.cmd doesn’t quite match the master-setting behavior of the Unix based one. I’ll send a pull request to fix them soon.
Matei On May 17, 2014, at 11:32 AM, Sandy Ryza <sandy.r...@cloudera.com> wrote: > +1 > > Reran my tests from rc5: > > * Built the release from source. > * Compiled Java and Scala apps that interact with HDFS against it. > * Ran them in local mode. > * Ran them against a pseudo-distributed YARN cluster in both yarn-client > mode and yarn-cluster mode. > > > On Sat, May 17, 2014 at 10:08 AM, Andrew Or <and...@databricks.com> wrote: > >> +1 >> >> >> 2014-05-17 8:53 GMT-07:00 Mark Hamstra <m...@clearstorydata.com>: >> >>> +1 >>> >>> >>> On Sat, May 17, 2014 at 12:58 AM, Patrick Wendell <pwend...@gmail.com >>>> wrote: >>> >>>> I'll start the voting with a +1. >>>> >>>> On Sat, May 17, 2014 at 12:58 AM, Patrick Wendell <pwend...@gmail.com> >>>> wrote: >>>>> Please vote on releasing the following candidate as Apache Spark >>> version >>>> 1.0.0! >>>>> This has one bug fix and one minor feature on top of rc8: >>>>> SPARK-1864: https://github.com/apache/spark/pull/808 >>>>> SPARK-1808: https://github.com/apache/spark/pull/799 >>>>> >>>>> The tag to be voted on is v1.0.0-rc9 (commit 920f947): >>>>> >>>> >>> >> https://git-wip-us.apache.org/repos/asf?p=spark.git;a=commit;h=920f947eb5a22a679c0c3186cf69ee75f6041c75 >>>>> >>>>> The release files, including signatures, digests, etc. can be found >> at: >>>>> http://people.apache.org/~pwendell/spark-1.0.0-rc9/ >>>>> >>>>> Release artifacts are signed with the following key: >>>>> https://people.apache.org/keys/committer/pwendell.asc >>>>> >>>>> The staging repository for this release can be found at: >>>>> >>> https://repository.apache.org/content/repositories/orgapachespark-1017/ >>>>> >>>>> The documentation corresponding to this release can be found at: >>>>> http://people.apache.org/~pwendell/spark-1.0.0-rc9-docs/ >>>>> >>>>> Please vote on releasing this package as Apache Spark 1.0.0! >>>>> >>>>> The vote is open until Tuesday, May 20, at 08:56 UTC and passes if >>>>> amajority of at least 3 +1 PMC votes are cast. >>>>> >>>>> [ ] +1 Release this package as Apache Spark 1.0.0 >>>>> [ ] -1 Do not release this package because ... >>>>> >>>>> To learn more about Apache Spark, please see >>>>> http://spark.apache.org/ >>>>> >>>>> == API Changes == >>>>> We welcome users to compile Spark applications against 1.0. There are >>>>> a few API changes in this release. Here are links to the associated >>>>> upgrade guides - user facing changes have been kept as small as >>>>> possible. >>>>> >>>>> changes to ML vector specification: >>>>> >>>> >>> >> http://people.apache.org/~pwendell/spark-1.0.0-rc8-docs/mllib-guide.html#from-09-to-10 >>>>> >>>>> changes to the Java API: >>>>> >>>> >>> >> http://people.apache.org/~pwendell/spark-1.0.0-rc8-docs/java-programming-guide.html#upgrading-from-pre-10-versions-of-spark >>>>> >>>>> changes to the streaming API: >>>>> >>>> >>> >> http://people.apache.org/~pwendell/spark-1.0.0-rc8-docs/streaming-programming-guide.html#migration-guide-from-091-or-below-to-1x >>>>> >>>>> changes to the GraphX API: >>>>> >>>> >>> >> http://people.apache.org/~pwendell/spark-1.0.0-rc8-docs/graphx-programming-guide.html#upgrade-guide-from-spark-091 >>>>> >>>>> coGroup and related functions now return Iterable[T] instead of >> Seq[T] >>>>> ==> Call toSeq on the result to restore the old behavior >>>>> >>>>> SparkContext.jarOfClass returns Option[String] instead of Seq[String] >>>>> ==> Call toSeq on the result to restore old behavior >>>> >>> >>