+1. I also tested on Windows just in case, with jars referring other jars and python files referring other python files. Path resolution still works.
2014-12-02 10:16 GMT-08:00 Jeremy Freeman <freeman.jer...@gmail.com>: > +1 (non-binding) > > Installed version pre-built for Hadoop on a private HPC > ran PySpark shell w/ iPython > loaded data using custom Hadoop input formats > ran MLlib routines in PySpark > ran custom workflows in PySpark > browsed the web UI > > Noticeable improvements in stability and performance during large shuffles > (as well as the elimination of frequent but unpredictable “FileNotFound / > too many open files” errors). > > We initially hit errors during large collects that ran fine in 1.1, but > setting the new spark.driver.maxResultSize to 0 preserved the old behavior. > Definitely worth highlighting this setting in the release notes, as the new > default may be too small for some users and workloads. > > — Jeremy > > ------------------------- > jeremyfreeman.net > @thefreemanlab > > On Dec 2, 2014, at 3:22 AM, Denny Lee <denny.g....@gmail.com> wrote: > > > +1 (non-binding) > > > > Verified on OSX 10.10.2, built from source, > > spark-shell / spark-submit jobs > > ran various simple Spark / Scala queries > > ran various SparkSQL queries (including HiveContext) > > ran ThriftServer service and connected via beeline > > ran SparkSVD > > > > > > On Mon Dec 01 2014 at 11:09:26 PM Patrick Wendell <pwend...@gmail.com> > > wrote: > > > >> Hey All, > >> > >> Just an update. Josh, Andrew, and others are working to reproduce > >> SPARK-4498 and fix it. Other than that issue no serious regressions > >> have been reported so far. If we are able to get a fix in for that > >> soon, we'll likely cut another RC with the patch. > >> > >> Continued testing of RC1 is definitely appreciated! > >> > >> I'll leave this vote open to allow folks to continue posting comments. > >> It's fine to still give "+1" from your own testing... i.e. you can > >> assume at this point SPARK-4498 will be fixed before releasing. > >> > >> - Patrick > >> > >> On Mon, Dec 1, 2014 at 3:30 PM, Matei Zaharia <matei.zaha...@gmail.com> > >> wrote: > >>> +0.9 from me. Tested it on Mac and Windows (someone has to do it) and > >> while things work, I noticed a few recent scripts don't have Windows > >> equivalents, namely https://issues.apache.org/jira/browse/SPARK-4683 > and > >> https://issues.apache.org/jira/browse/SPARK-4684. The first one at > least > >> would be good to fix if we do another RC. Not blocking the release but > >> useful to fix in docs is > https://issues.apache.org/jira/browse/SPARK-4685. > >>> > >>> Matei > >>> > >>> > >>>> On Dec 1, 2014, at 11:18 AM, Josh Rosen <rosenvi...@gmail.com> wrote: > >>>> > >>>> Hi everyone, > >>>> > >>>> There's an open bug report related to Spark standalone which could be > a > >> potential release-blocker (pending investigation / a bug fix): > >> https://issues.apache.org/jira/browse/SPARK-4498. This issue seems > >> non-deterministc and only affects long-running Spark standalone > >> deployments, so it may be hard to reproduce. I'm going to work on a > patch > >> to add additional logging in order to help with debugging. > >>>> > >>>> I just wanted to give an early head's up about this issue and to get > >> more eyes on it in case anyone else has run into it or wants to help > with > >> debugging. > >>>> > >>>> - Josh > >>>> > >>>> On November 28, 2014 at 9:18:09 PM, Patrick Wendell ( > pwend...@gmail.com) > >> wrote: > >>>> > >>>> Please vote on releasing the following candidate as Apache Spark > >> version 1.2.0! > >>>> > >>>> The tag to be voted on is v1.2.0-rc1 (commit 1056e9ec1): > >>>> https://git-wip-us.apache.org/repos/asf?p=spark.git;a=commit;h= > >> 1056e9ec13203d0c51564265e94d77a054498fdb > >>>> > >>>> The release files, including signatures, digests, etc. can be found > at: > >>>> http://people.apache.org/~pwendell/spark-1.2.0-rc1/ > >>>> > >>>> Release artifacts are signed with the following key: > >>>> https://people.apache.org/keys/committer/pwendell.asc > >>>> > >>>> The staging repository for this release can be found at: > >>>> > https://repository.apache.org/content/repositories/orgapachespark-1048/ > >>>> > >>>> The documentation corresponding to this release can be found at: > >>>> http://people.apache.org/~pwendell/spark-1.2.0-rc1-docs/ > >>>> > >>>> Please vote on releasing this package as Apache Spark 1.2.0! > >>>> > >>>> The vote is open until Tuesday, December 02, at 05:15 UTC and passes > >>>> if a majority of at least 3 +1 PMC votes are cast. > >>>> > >>>> [ ] +1 Release this package as Apache Spark 1.1.0 > >>>> [ ] -1 Do not release this package because ... > >>>> > >>>> To learn more about Apache Spark, please see > >>>> http://spark.apache.org/ > >>>> > >>>> == What justifies a -1 vote for this release? == > >>>> This vote is happening very late into the QA period compared with > >>>> previous votes, so -1 votes should only occur for significant > >>>> regressions from 1.0.2. Bugs already present in 1.1.X, minor > >>>> regressions, or bugs related to new features will not block this > >>>> release. > >>>> > >>>> == What default changes should I be aware of? == > >>>> 1. The default value of "spark.shuffle.blockTransferService" has been > >>>> changed to "netty" > >>>> --> Old behavior can be restored by switching to "nio" > >>>> > >>>> 2. The default value of "spark.shuffle.manager" has been changed to > >> "sort". > >>>> --> Old behavior can be restored by setting "spark.shuffle.manager" to > >> "hash". > >>>> > >>>> == Other notes == > >>>> Because this vote is occurring over a weekend, I will likely extend > >>>> the vote if this RC survives until the end of the vote period. > >>>> > >>>> - Patrick > >>>> > >>>> --------------------------------------------------------------------- > >>>> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org > >>>> For additional commands, e-mail: dev-h...@spark.apache.org > >>>> > >>> > >>> > >>> --------------------------------------------------------------------- > >>> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org > >>> For additional commands, e-mail: dev-h...@spark.apache.org > >>> > >> > >> --------------------------------------------------------------------- > >> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org > >> For additional commands, e-mail: dev-h...@spark.apache.org > >> > >> > >