I'm waiting for SPARK-20814 <https://issues.apache.org/jira/browse/SPARK-20814> at Marcelo's request and I'd also like to include SPARK-20844 <https://issues.apache.org/jira/browse/SPARK-20844>. I think we should be able to cut another RC midweek.
On Fri, May 19, 2017 at 11:53 AM, Nick Pentreath <nick.pentre...@gmail.com> wrote: > All the outstanding ML QA doc and user guide items are done for 2.2 so > from that side we should be good to cut another RC :) > > > On Thu, 18 May 2017 at 00:18 Russell Spitzer <russell.spit...@gmail.com> > wrote: > >> Seeing an issue with the DataScanExec and some of our integration tests >> for the SCC. Running dataframe read and writes from the shell seems fine >> but the Redaction code seems to get a "None" when doing >> SparkSession.getActiveSession.get in our integration tests. I'm not sure >> why but i'll dig into this later if I get a chance. >> >> Example Failed Test >> https://github.com/datastax/spark-cassandra-connector/ >> blob/v2.0.1/spark-cassandra-connector/src/it/scala/com/ >> datastax/spark/connector/sql/CassandraSQLSpec.scala#L311 >> >> ```[info] org.apache.spark.SparkException: Job aborted due to stage >> failure: Task serialization failed: java.util.NoSuchElementException: >> None.get >> [info] java.util.NoSuchElementException: None.get >> [info] at scala.None$.get(Option.scala:347) >> [info] at scala.None$.get(Option.scala:345) >> [info] at org.apache.spark.sql.execution.DataSourceScanExec$class.org >> $apache$spark$sql$execution$DataSourceScanExec$$ >> redact(DataSourceScanExec.scala:70) >> [info] at org.apache.spark.sql.execution.DataSourceScanExec$$ >> anonfun$4.apply(DataSourceScanExec.scala:54) >> [info] at org.apache.spark.sql.execution.DataSourceScanExec$$ >> anonfun$4.apply(DataSourceScanExec.scala:52) >> ``` >> >> Again this only seems to repo in our IT suite so i'm not sure if this is >> a real issue. >> >> >> On Tue, May 16, 2017 at 1:40 PM Joseph Bradley <jos...@databricks.com> >> wrote: >> >>> All of the ML/Graph/SparkR QA blocker JIRAs have been resolved. Thanks >>> everyone who helped out on those! >>> >>> We still have open ML/Graph/SparkR JIRAs targeted at 2.2, but they are >>> essentially all for documentation. >>> >>> Joseph >>> >>> On Thu, May 11, 2017 at 3:08 PM, Marcelo Vanzin <van...@cloudera.com> >>> wrote: >>> >>>> Since you'll be creating a new RC, I'd wait until SPARK-20666 is >>>> fixed, since the change that caused it is in branch-2.2. Probably a >>>> good idea to raise it to blocker at this point. >>>> >>>> On Thu, May 11, 2017 at 2:59 PM, Michael Armbrust >>>> <mich...@databricks.com> wrote: >>>> > I'm going to -1 given the outstanding issues and lack of +1s. I'll >>>> create >>>> > another RC once ML has had time to take care of the more critical >>>> problems. >>>> > In the meantime please keep testing this release! >>>> > >>>> > On Tue, May 9, 2017 at 2:00 AM, Kazuaki Ishizaki <ishiz...@jp.ibm.com >>>> > >>>> > wrote: >>>> >> >>>> >> +1 (non-binding) >>>> >> >>>> >> I tested it on Ubuntu 16.04 and OpenJDK8 on ppc64le. All of the >>>> tests for >>>> >> core have passed. >>>> >> >>>> >> $ java -version >>>> >> openjdk version "1.8.0_111" >>>> >> OpenJDK Runtime Environment (build >>>> >> 1.8.0_111-8u111-b14-2ubuntu0.16.04.2-b14) >>>> >> OpenJDK 64-Bit Server VM (build 25.111-b14, mixed mode) >>>> >> $ build/mvn -DskipTests -Phive -Phive-thriftserver -Pyarn >>>> -Phadoop-2.7 >>>> >> package install >>>> >> $ build/mvn -Phive -Phive-thriftserver -Pyarn -Phadoop-2.7 test -pl >>>> core >>>> >> ... >>>> >> Run completed in 15 minutes, 12 seconds. >>>> >> Total number of tests run: 1940 >>>> >> Suites: completed 206, aborted 0 >>>> >> Tests: succeeded 1940, failed 0, canceled 4, ignored 8, pending 0 >>>> >> All tests passed. >>>> >> [INFO] >>>> >> ------------------------------------------------------------ >>>> ------------ >>>> >> [INFO] BUILD SUCCESS >>>> >> [INFO] >>>> >> ------------------------------------------------------------ >>>> ------------ >>>> >> [INFO] Total time: 16:51 min >>>> >> [INFO] Finished at: 2017-05-09T17:51:04+09:00 >>>> >> [INFO] Final Memory: 53M/514M >>>> >> [INFO] >>>> >> ------------------------------------------------------------ >>>> ------------ >>>> >> [WARNING] The requested profile "hive" could not be activated >>>> because it >>>> >> does not exist. >>>> >> >>>> >> >>>> >> Kazuaki Ishizaki, >>>> >> >>>> >> >>>> >> >>>> >> From: Michael Armbrust <mich...@databricks.com> >>>> >> To: "dev@spark.apache.org" <dev@spark.apache.org> >>>> >> Date: 2017/05/05 02:08 >>>> >> Subject: [VOTE] Apache Spark 2.2.0 (RC2) >>>> >> ________________________________ >>>> >> >>>> >> >>>> >> >>>> >> Please vote on releasing the following candidate as Apache Spark >>>> version >>>> >> 2.2.0. The vote is open until Tues, May 9th, 2017 at 12:00 PST and >>>> passes if >>>> >> a majority of at least 3 +1 PMC votes are cast. >>>> >> >>>> >> [ ] +1 Release this package as Apache Spark 2.2.0 >>>> >> [ ] -1 Do not release this package because ... >>>> >> >>>> >> >>>> >> To learn more about Apache Spark, please see >>>> http://spark.apache.org/ >>>> >> >>>> >> The tag to be voted on is v2.2.0-rc2 >>>> >> (1d4017b44d5e6ad156abeaae6371747f111dd1f9) >>>> >> >>>> >> List of JIRA tickets resolved can be found with this filter. >>>> >> >>>> >> The release files, including signatures, digests, etc. can be found >>>> at: >>>> >> http://home.apache.org/~pwendell/spark-releases/spark-2.2.0-rc2-bin/ >>>> >> >>>> >> Release artifacts are signed with the following key: >>>> >> https://people.apache.org/keys/committer/pwendell.asc >>>> >> >>>> >> The staging repository for this release can be found at: >>>> >> https://repository.apache.org/content/repositories/ >>>> orgapachespark-1236/ >>>> >> >>>> >> The documentation corresponding to this release can be found at: >>>> >> http://people.apache.org/~pwendell/spark-releases/spark- >>>> 2.2.0-rc2-docs/ >>>> >> >>>> >> >>>> >> FAQ >>>> >> >>>> >> How can I help test this release? >>>> >> >>>> >> If you are a Spark user, you can help us test this release by taking >>>> an >>>> >> existing Spark workload and running on this release candidate, then >>>> >> reporting any regressions. >>>> >> >>>> >> What should happen to JIRA tickets still targeting 2.2.0? >>>> >> >>>> >> Committers should look at those and triage. Extremely important bug >>>> fixes, >>>> >> documentation, and API tweaks that impact compatibility should be >>>> worked on >>>> >> immediately. Everything else please retarget to 2.3.0 or 2.2.1. >>>> >> >>>> >> But my bug isn't fixed!??! >>>> >> >>>> >> In order to make timely releases, we will typically not hold the >>>> release >>>> >> unless the bug in question is a regression from 2.1.1. >>>> >> >>>> > >>>> >>>> >>>> >>>> -- >>>> Marcelo >>>> >>>> --------------------------------------------------------------------- >>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org >>>> >>>> >>> >>> >>> -- >>> >>> Joseph Bradley >>> >>> Software Engineer - Machine Learning >>> >>> Databricks, Inc. >>> >>> [image: http://databricks.com] <http://databricks.com/> >>> >>