+1 Signatures and hashes check out. I checked that the Kinesis
assembly artifacts are not present.

I compiled and tested on Java 8 / Ubuntu 16 with -Pyarn -Phive
-Phive-thriftserver -Phadoop-2.7 -Psparkr and only saw one test
problem. This test never completed. If nobody else sees it, +1,
assuming it's a bad test or env issue.

- should clone and clean line object in ClosureCleaner *** FAILED ***
  isContain was true Interpreter output contained 'Exception':
  Welcome to
        ____              __
       / __/__  ___ _____/ /__
      _\ \/ _ \/ _ `/ __/  '_/
     /___/ .__/\_,_/_/ /_/\_\   version 2.0.1
        /_/

  Using Scala version 2.11.8 (OpenJDK 64-Bit Server VM, Java 1.8.0_91)
  Type in expressions to have them evaluated.
  Type :help for more information.

  scala> // Entering paste mode (ctrl-D to finish)


  // Exiting paste mode, now interpreting.

  org.apache.spark.SparkException: Job 0 cancelled because
SparkContext was shut down
    at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$cleanUpAfterSchedulerStop$1.apply(DAGScheduler.scala:818)
...


On Fri, Sep 23, 2016 at 7:01 AM, Reynold Xin <r...@databricks.com> wrote:
> Please vote on releasing the following candidate as Apache Spark version
> 2.0.1. The vote is open until Sunday, Sep 25, 2016 at 23:59 PDT and passes
> if a majority of at least 3+1 PMC votes are cast.
>
> [ ] +1 Release this package as Apache Spark 2.0.1
> [ ] -1 Do not release this package because ...
>
>
> The tag to be voted on is v2.0.1-rc2
> (04141ad49806a48afccc236b699827997142bd57)
>
> This release candidate resolves 284 issues:
> https://s.apache.org/spark-2.0.1-jira
>
> The release files, including signatures, digests, etc. can be found at:
> http://people.apache.org/~pwendell/spark-releases/spark-2.0.1-rc2-bin/
>
> Release artifacts are signed with the following key:
> https://people.apache.org/keys/committer/pwendell.asc
>
> The staging repository for this release can be found at:
> https://repository.apache.org/content/repositories/orgapachespark-1199
>
> The documentation corresponding to this release can be found at:
> http://people.apache.org/~pwendell/spark-releases/spark-2.0.1-rc2-docs/
>
>
> Q: How can I help test this release?
> A: If you are a Spark user, you can help us test this release by taking an
> existing Spark workload and running on this release candidate, then
> reporting any regressions from 2.0.0.
>
> Q: What justifies a -1 vote for this release?
> A: This is a maintenance release in the 2.0.x series.  Bugs already present
> in 2.0.0, missing features, or bugs related to new features will not
> necessarily block this release.
>
> Q: What happened to 2.0.1 RC1?
> A: There was an issue with RC1 R documentation during release candidate
> preparation. As a result, rc1 was canceled before a vote was called.
>

---------------------------------------------------------------------
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org

Reply via email to