This vote fails. Following shortly with RC3 On Thu, Jun 1, 2017 at 8:28 PM, Reynold Xin <r...@databricks.com> wrote:
> Again (I've probably said this more than 10 times already in different > threads), SPARK-18350 has no impact on whether the timestamp type is with > timezone or without timezone. It simply allows a session specific timezone > setting rather than having Spark always rely on the machine timezone. > > On Wed, May 31, 2017 at 11:58 AM, Kostas Sakellis <kos...@cloudera.com> > wrote: > >> Hey Michael, >> >> There is a discussion on TIMESTAMP semantics going on the thread "SQL >> TIMESTAMP semantics vs. SPARK-18350" which might impact Spark 2.2. Should >> we make a decision there before voting on the next RC for Spark 2.2? >> >> Thanks, >> Kostas >> >> On Tue, May 30, 2017 at 12:09 PM, Michael Armbrust < >> mich...@databricks.com> wrote: >> >>> Last call, anything else important in-flight for 2.2? >>> >>> On Thu, May 25, 2017 at 10:56 AM, Michael Allman <mich...@videoamp.com> >>> wrote: >>> >>>> PR is here: https://github.com/apache/spark/pull/18112 >>>> >>>> >>>> On May 25, 2017, at 10:28 AM, Michael Allman <mich...@videoamp.com> >>>> wrote: >>>> >>>> Michael, >>>> >>>> If you haven't started cutting the new RC, I'm working on a >>>> documentation PR right now I'm hoping we can get into Spark 2.2 as a >>>> migration note, even if it's just a mention: https://issues.apache >>>> .org/jira/browse/SPARK-20888. >>>> >>>> Michael >>>> >>>> >>>> On May 22, 2017, at 11:39 AM, Michael Armbrust <mich...@databricks.com> >>>> wrote: >>>> >>>> I'm waiting for SPARK-20814 >>>> <https://issues.apache.org/jira/browse/SPARK-20814> at Marcelo's >>>> request and I'd also like to include SPARK-20844 >>>> <https://issues.apache.org/jira/browse/SPARK-20844>. I think we >>>> should be able to cut another RC midweek. >>>> >>>> On Fri, May 19, 2017 at 11:53 AM, Nick Pentreath < >>>> nick.pentre...@gmail.com> wrote: >>>> >>>>> All the outstanding ML QA doc and user guide items are done for 2.2 so >>>>> from that side we should be good to cut another RC :) >>>>> >>>>> >>>>> On Thu, 18 May 2017 at 00:18 Russell Spitzer < >>>>> russell.spit...@gmail.com> wrote: >>>>> >>>>>> Seeing an issue with the DataScanExec and some of our integration >>>>>> tests for the SCC. Running dataframe read and writes from the shell seems >>>>>> fine but the Redaction code seems to get a "None" when doing >>>>>> SparkSession.getActiveSession.get in our integration tests. I'm not >>>>>> sure why but i'll dig into this later if I get a chance. >>>>>> >>>>>> Example Failed Test >>>>>> https://github.com/datastax/spark-cassandra-connector/blob/v >>>>>> 2.0.1/spark-cassandra-connector/src/it/scala/com/datastax/sp >>>>>> ark/connector/sql/CassandraSQLSpec.scala#L311 >>>>>> >>>>>> ```[info] org.apache.spark.SparkException: Job aborted due to >>>>>> stage failure: Task serialization failed: >>>>>> java.util.NoSuchElementException: >>>>>> None.get >>>>>> [info] java.util.NoSuchElementException: None.get >>>>>> [info] at scala.None$.get(Option.scala:347) >>>>>> [info] at scala.None$.get(Option.scala:345) >>>>>> [info] at org.apache.spark.sql.execution.DataSourceScanExec$class.org >>>>>> $apache$spark$sql$execution$DataSourceScanExec$$redact(DataSo >>>>>> urceScanExec.scala:70) >>>>>> [info] at org.apache.spark.sql.execution >>>>>> .DataSourceScanExec$$anonfun$4.apply(DataSourceScanExec.scala:54) >>>>>> [info] at org.apache.spark.sql.execution >>>>>> .DataSourceScanExec$$anonfun$4.apply(DataSourceScanExec.scala:52) >>>>>> ``` >>>>>> >>>>>> Again this only seems to repo in our IT suite so i'm not sure if this >>>>>> is a real issue. >>>>>> >>>>>> >>>>>> On Tue, May 16, 2017 at 1:40 PM Joseph Bradley <jos...@databricks.com> >>>>>> wrote: >>>>>> >>>>>>> All of the ML/Graph/SparkR QA blocker JIRAs have been resolved. >>>>>>> Thanks everyone who helped out on those! >>>>>>> >>>>>>> We still have open ML/Graph/SparkR JIRAs targeted at 2.2, but they >>>>>>> are essentially all for documentation. >>>>>>> >>>>>>> Joseph >>>>>>> >>>>>>> On Thu, May 11, 2017 at 3:08 PM, Marcelo Vanzin <van...@cloudera.com >>>>>>> > wrote: >>>>>>> >>>>>>>> Since you'll be creating a new RC, I'd wait until SPARK-20666 is >>>>>>>> fixed, since the change that caused it is in branch-2.2. Probably a >>>>>>>> good idea to raise it to blocker at this point. >>>>>>>> >>>>>>>> On Thu, May 11, 2017 at 2:59 PM, Michael Armbrust >>>>>>>> <mich...@databricks.com> wrote: >>>>>>>> > I'm going to -1 given the outstanding issues and lack of +1s. >>>>>>>> I'll create >>>>>>>> > another RC once ML has had time to take care of the more critical >>>>>>>> problems. >>>>>>>> > In the meantime please keep testing this release! >>>>>>>> > >>>>>>>> > On Tue, May 9, 2017 at 2:00 AM, Kazuaki Ishizaki < >>>>>>>> ishiz...@jp.ibm.com> >>>>>>>> > wrote: >>>>>>>> >> >>>>>>>> >> +1 (non-binding) >>>>>>>> >> >>>>>>>> >> I tested it on Ubuntu 16.04 and OpenJDK8 on ppc64le. All of the >>>>>>>> tests for >>>>>>>> >> core have passed. >>>>>>>> >> >>>>>>>> >> $ java -version >>>>>>>> >> openjdk version "1.8.0_111" >>>>>>>> >> OpenJDK Runtime Environment (build >>>>>>>> >> 1.8.0_111-8u111-b14-2ubuntu0.16.04.2-b14) >>>>>>>> >> OpenJDK 64-Bit Server VM (build 25.111-b14, mixed mode) >>>>>>>> >> $ build/mvn -DskipTests -Phive -Phive-thriftserver -Pyarn >>>>>>>> -Phadoop-2.7 >>>>>>>> >> package install >>>>>>>> >> $ build/mvn -Phive -Phive-thriftserver -Pyarn -Phadoop-2.7 test >>>>>>>> -pl core >>>>>>>> >> ... >>>>>>>> >> Run completed in 15 minutes, 12 seconds. >>>>>>>> >> Total number of tests run: 1940 >>>>>>>> >> Suites: completed 206, aborted 0 >>>>>>>> >> Tests: succeeded 1940, failed 0, canceled 4, ignored 8, pending 0 >>>>>>>> >> All tests passed. >>>>>>>> >> [INFO] >>>>>>>> >> ------------------------------------------------------------ >>>>>>>> ------------ >>>>>>>> >> [INFO] BUILD SUCCESS >>>>>>>> >> [INFO] >>>>>>>> >> ------------------------------------------------------------ >>>>>>>> ------------ >>>>>>>> >> [INFO] Total time: 16:51 min >>>>>>>> >> [INFO] Finished at: 2017-05-09T17:51:04+09:00 >>>>>>>> >> [INFO] Final Memory: 53M/514M >>>>>>>> >> [INFO] >>>>>>>> >> ------------------------------------------------------------ >>>>>>>> ------------ >>>>>>>> >> [WARNING] The requested profile "hive" could not be activated >>>>>>>> because it >>>>>>>> >> does not exist. >>>>>>>> >> >>>>>>>> >> >>>>>>>> >> Kazuaki Ishizaki, >>>>>>>> >> >>>>>>>> >> >>>>>>>> >> >>>>>>>> >> From: Michael Armbrust <mich...@databricks.com> >>>>>>>> >> To: "dev@spark.apache.org" <dev@spark.apache.org> >>>>>>>> >> Date: 2017/05/05 02:08 >>>>>>>> >> Subject: [VOTE] Apache Spark 2.2.0 (RC2) >>>>>>>> >> ________________________________ >>>>>>>> >> >>>>>>>> >> >>>>>>>> >> >>>>>>>> >> Please vote on releasing the following candidate as Apache Spark >>>>>>>> version >>>>>>>> >> 2.2.0. The vote is open until Tues, May 9th, 2017 at 12:00 PST >>>>>>>> and passes if >>>>>>>> >> a majority of at least 3 +1 PMC votes are cast. >>>>>>>> >> >>>>>>>> >> [ ] +1 Release this package as Apache Spark 2.2.0 >>>>>>>> >> [ ] -1 Do not release this package because ... >>>>>>>> >> >>>>>>>> >> >>>>>>>> >> To learn more about Apache Spark, please see >>>>>>>> http://spark.apache.org/ >>>>>>>> >> >>>>>>>> >> The tag to be voted on is v2.2.0-rc2 >>>>>>>> >> (1d4017b44d5e6ad156abeaae6371747f111dd1f9) >>>>>>>> >> >>>>>>>> >> List of JIRA tickets resolved can be found with this filter. >>>>>>>> >> >>>>>>>> >> The release files, including signatures, digests, etc. can be >>>>>>>> found at: >>>>>>>> >> http://home.apache.org/~pwendell/spark-releases/spark-2.2.0- >>>>>>>> rc2-bin/ >>>>>>>> >> >>>>>>>> >> Release artifacts are signed with the following key: >>>>>>>> >> https://people.apache.org/keys/committer/pwendell.asc >>>>>>>> >> >>>>>>>> >> The staging repository for this release can be found at: >>>>>>>> >> https://repository.apache.org/content/repositories/orgapache >>>>>>>> spark-1236/ >>>>>>>> >> >>>>>>>> >> The documentation corresponding to this release can be found at: >>>>>>>> >> http://people.apache.org/~pwendell/spark-releases/spark-2.2. >>>>>>>> 0-rc2-docs/ >>>>>>>> >> >>>>>>>> >> >>>>>>>> >> FAQ >>>>>>>> >> >>>>>>>> >> How can I help test this release? >>>>>>>> >> >>>>>>>> >> If you are a Spark user, you can help us test this release by >>>>>>>> taking an >>>>>>>> >> existing Spark workload and running on this release candidate, >>>>>>>> then >>>>>>>> >> reporting any regressions. >>>>>>>> >> >>>>>>>> >> What should happen to JIRA tickets still targeting 2.2.0? >>>>>>>> >> >>>>>>>> >> Committers should look at those and triage. Extremely important >>>>>>>> bug fixes, >>>>>>>> >> documentation, and API tweaks that impact compatibility should >>>>>>>> be worked on >>>>>>>> >> immediately. Everything else please retarget to 2.3.0 or 2.2.1. >>>>>>>> >> >>>>>>>> >> But my bug isn't fixed!??! >>>>>>>> >> >>>>>>>> >> In order to make timely releases, we will typically not hold the >>>>>>>> release >>>>>>>> >> unless the bug in question is a regression from 2.1.1. >>>>>>>> >> >>>>>>>> > >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Marcelo >>>>>>>> >>>>>>>> ------------------------------------------------------------ >>>>>>>> --------- >>>>>>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org >>>>>>>> >>>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Joseph Bradley >>>>>>> Software Engineer - Machine Learning >>>>>>> Databricks, Inc. >>>>>>> [image: http://databricks.com] <http://databricks.com/> >>>>>>> >>>>>> >>>> >>>> >>>> >>> >> >