Hey Michael, There is a discussion on TIMESTAMP semantics going on the thread "SQL TIMESTAMP semantics vs. SPARK-18350" which might impact Spark 2.2. Should we make a decision there before voting on the next RC for Spark 2.2?
Thanks, Kostas On Tue, May 30, 2017 at 12:09 PM, Michael Armbrust <mich...@databricks.com> wrote: > Last call, anything else important in-flight for 2.2? > > On Thu, May 25, 2017 at 10:56 AM, Michael Allman <mich...@videoamp.com> > wrote: > >> PR is here: https://github.com/apache/spark/pull/18112 >> >> >> On May 25, 2017, at 10:28 AM, Michael Allman <mich...@videoamp.com> >> wrote: >> >> Michael, >> >> If you haven't started cutting the new RC, I'm working on a documentation >> PR right now I'm hoping we can get into Spark 2.2 as a migration note, even >> if it's just a mention: https://issues.apache.org/jira/browse/SPARK-20888 >> . >> >> Michael >> >> >> On May 22, 2017, at 11:39 AM, Michael Armbrust <mich...@databricks.com> >> wrote: >> >> I'm waiting for SPARK-20814 >> <https://issues.apache.org/jira/browse/SPARK-20814> at Marcelo's >> request and I'd also like to include SPARK-20844 >> <https://issues.apache.org/jira/browse/SPARK-20844>. I think we should >> be able to cut another RC midweek. >> >> On Fri, May 19, 2017 at 11:53 AM, Nick Pentreath < >> nick.pentre...@gmail.com> wrote: >> >>> All the outstanding ML QA doc and user guide items are done for 2.2 so >>> from that side we should be good to cut another RC :) >>> >>> >>> On Thu, 18 May 2017 at 00:18 Russell Spitzer <russell.spit...@gmail.com> >>> wrote: >>> >>>> Seeing an issue with the DataScanExec and some of our integration tests >>>> for the SCC. Running dataframe read and writes from the shell seems fine >>>> but the Redaction code seems to get a "None" when doing >>>> SparkSession.getActiveSession.get in our integration tests. I'm not >>>> sure why but i'll dig into this later if I get a chance. >>>> >>>> Example Failed Test >>>> https://github.com/datastax/spark-cassandra-connector/blob/v >>>> 2.0.1/spark-cassandra-connector/src/it/scala/com/datastax/sp >>>> ark/connector/sql/CassandraSQLSpec.scala#L311 >>>> >>>> ```[info] org.apache.spark.SparkException: Job aborted due to stage >>>> failure: Task serialization failed: java.util.NoSuchElementException: >>>> None.get >>>> [info] java.util.NoSuchElementException: None.get >>>> [info] at scala.None$.get(Option.scala:347) >>>> [info] at scala.None$.get(Option.scala:345) >>>> [info] at org.apache.spark.sql.execution.DataSourceScanExec$class.org$ >>>> apache$spark$sql$execution$DataSourceScanExec$$redact(DataSo >>>> urceScanExec.scala:70) >>>> [info] at org.apache.spark.sql.execution.DataSourceScanExec$$anonfun$4 >>>> .apply(DataSourceScanExec.scala:54) >>>> [info] at org.apache.spark.sql.execution.DataSourceScanExec$$anonfun$4 >>>> .apply(DataSourceScanExec.scala:52) >>>> ``` >>>> >>>> Again this only seems to repo in our IT suite so i'm not sure if this >>>> is a real issue. >>>> >>>> >>>> On Tue, May 16, 2017 at 1:40 PM Joseph Bradley <jos...@databricks.com> >>>> wrote: >>>> >>>>> All of the ML/Graph/SparkR QA blocker JIRAs have been resolved. >>>>> Thanks everyone who helped out on those! >>>>> >>>>> We still have open ML/Graph/SparkR JIRAs targeted at 2.2, but they are >>>>> essentially all for documentation. >>>>> >>>>> Joseph >>>>> >>>>> On Thu, May 11, 2017 at 3:08 PM, Marcelo Vanzin <van...@cloudera.com> >>>>> wrote: >>>>> >>>>>> Since you'll be creating a new RC, I'd wait until SPARK-20666 is >>>>>> fixed, since the change that caused it is in branch-2.2. Probably a >>>>>> good idea to raise it to blocker at this point. >>>>>> >>>>>> On Thu, May 11, 2017 at 2:59 PM, Michael Armbrust >>>>>> <mich...@databricks.com> wrote: >>>>>> > I'm going to -1 given the outstanding issues and lack of +1s. I'll >>>>>> create >>>>>> > another RC once ML has had time to take care of the more critical >>>>>> problems. >>>>>> > In the meantime please keep testing this release! >>>>>> > >>>>>> > On Tue, May 9, 2017 at 2:00 AM, Kazuaki Ishizaki < >>>>>> ishiz...@jp.ibm.com> >>>>>> > wrote: >>>>>> >> >>>>>> >> +1 (non-binding) >>>>>> >> >>>>>> >> I tested it on Ubuntu 16.04 and OpenJDK8 on ppc64le. All of the >>>>>> tests for >>>>>> >> core have passed. >>>>>> >> >>>>>> >> $ java -version >>>>>> >> openjdk version "1.8.0_111" >>>>>> >> OpenJDK Runtime Environment (build >>>>>> >> 1.8.0_111-8u111-b14-2ubuntu0.16.04.2-b14) >>>>>> >> OpenJDK 64-Bit Server VM (build 25.111-b14, mixed mode) >>>>>> >> $ build/mvn -DskipTests -Phive -Phive-thriftserver -Pyarn >>>>>> -Phadoop-2.7 >>>>>> >> package install >>>>>> >> $ build/mvn -Phive -Phive-thriftserver -Pyarn -Phadoop-2.7 test >>>>>> -pl core >>>>>> >> ... >>>>>> >> Run completed in 15 minutes, 12 seconds. >>>>>> >> Total number of tests run: 1940 >>>>>> >> Suites: completed 206, aborted 0 >>>>>> >> Tests: succeeded 1940, failed 0, canceled 4, ignored 8, pending 0 >>>>>> >> All tests passed. >>>>>> >> [INFO] >>>>>> >> ------------------------------------------------------------ >>>>>> ------------ >>>>>> >> [INFO] BUILD SUCCESS >>>>>> >> [INFO] >>>>>> >> ------------------------------------------------------------ >>>>>> ------------ >>>>>> >> [INFO] Total time: 16:51 min >>>>>> >> [INFO] Finished at: 2017-05-09T17:51:04+09:00 >>>>>> >> [INFO] Final Memory: 53M/514M >>>>>> >> [INFO] >>>>>> >> ------------------------------------------------------------ >>>>>> ------------ >>>>>> >> [WARNING] The requested profile "hive" could not be activated >>>>>> because it >>>>>> >> does not exist. >>>>>> >> >>>>>> >> >>>>>> >> Kazuaki Ishizaki, >>>>>> >> >>>>>> >> >>>>>> >> >>>>>> >> From: Michael Armbrust <mich...@databricks.com> >>>>>> >> To: "dev@spark.apache.org" <dev@spark.apache.org> >>>>>> >> Date: 2017/05/05 02:08 >>>>>> >> Subject: [VOTE] Apache Spark 2.2.0 (RC2) >>>>>> >> ________________________________ >>>>>> >> >>>>>> >> >>>>>> >> >>>>>> >> Please vote on releasing the following candidate as Apache Spark >>>>>> version >>>>>> >> 2.2.0. The vote is open until Tues, May 9th, 2017 at 12:00 PST and >>>>>> passes if >>>>>> >> a majority of at least 3 +1 PMC votes are cast. >>>>>> >> >>>>>> >> [ ] +1 Release this package as Apache Spark 2.2.0 >>>>>> >> [ ] -1 Do not release this package because ... >>>>>> >> >>>>>> >> >>>>>> >> To learn more about Apache Spark, please see >>>>>> http://spark.apache.org/ >>>>>> >> >>>>>> >> The tag to be voted on is v2.2.0-rc2 >>>>>> >> (1d4017b44d5e6ad156abeaae6371747f111dd1f9) >>>>>> >> >>>>>> >> List of JIRA tickets resolved can be found with this filter. >>>>>> >> >>>>>> >> The release files, including signatures, digests, etc. can be >>>>>> found at: >>>>>> >> http://home.apache.org/~pwendell/spark-releases/spark-2.2.0- >>>>>> rc2-bin/ >>>>>> >> >>>>>> >> Release artifacts are signed with the following key: >>>>>> >> https://people.apache.org/keys/committer/pwendell.asc >>>>>> >> >>>>>> >> The staging repository for this release can be found at: >>>>>> >> https://repository.apache.org/content/repositories/orgapache >>>>>> spark-1236/ >>>>>> >> >>>>>> >> The documentation corresponding to this release can be found at: >>>>>> >> http://people.apache.org/~pwendell/spark-releases/spark-2.2. >>>>>> 0-rc2-docs/ >>>>>> >> >>>>>> >> >>>>>> >> FAQ >>>>>> >> >>>>>> >> How can I help test this release? >>>>>> >> >>>>>> >> If you are a Spark user, you can help us test this release by >>>>>> taking an >>>>>> >> existing Spark workload and running on this release candidate, then >>>>>> >> reporting any regressions. >>>>>> >> >>>>>> >> What should happen to JIRA tickets still targeting 2.2.0? >>>>>> >> >>>>>> >> Committers should look at those and triage. Extremely important >>>>>> bug fixes, >>>>>> >> documentation, and API tweaks that impact compatibility should be >>>>>> worked on >>>>>> >> immediately. Everything else please retarget to 2.3.0 or 2.2.1. >>>>>> >> >>>>>> >> But my bug isn't fixed!??! >>>>>> >> >>>>>> >> In order to make timely releases, we will typically not hold the >>>>>> release >>>>>> >> unless the bug in question is a regression from 2.1.1. >>>>>> >> >>>>>> > >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> Marcelo >>>>>> >>>>>> --------------------------------------------------------------------- >>>>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org >>>>>> >>>>>> >>>>> >>>>> >>>>> -- >>>>> Joseph Bradley >>>>> Software Engineer - Machine Learning >>>>> Databricks, Inc. >>>>> [image: http://databricks.com] <http://databricks.com/> >>>>> >>>> >> >> >> >