+1 ---- Ricardo Almeida
On 20 May 2016 at 18:33, Mark Hamstra <m...@clearstorydata.com> wrote: > This is isn't yet a release candidate since, as Reynold mention in his > opening post, preview releases are "not meant to be functional, i.e. they > can and highly likely will contain critical bugs or documentation errors." > Once we're at the point where we expect there not to be such bugs and > errors, then the release candidates will start. > > On Fri, May 20, 2016 at 4:40 AM, Ross Lawley <ross.law...@gmail.com> > wrote: > >> +1 Having an rc1 would help me get stable feedback on using my library >> with Spark, compared to relying on 2.0.0-SNAPSHOT. >> >> >> On Fri, 20 May 2016 at 05:57 Xiao Li <gatorsm...@gmail.com> wrote: >> >>> Changed my vote to +1. Thanks! >>> >>> 2016-05-19 13:28 GMT-07:00 Xiao Li <gatorsm...@gmail.com>: >>> >>>> Will do. Thanks! >>>> >>>> 2016-05-19 13:26 GMT-07:00 Reynold Xin <r...@databricks.com>: >>>> >>>>> Xiao thanks for posting. Please file a bug in JIRA. Again as I said in >>>>> the email this is not meant to be a functional release and will contain >>>>> bugs. >>>>> >>>>> On Thu, May 19, 2016 at 1:20 PM, Xiao Li <gatorsm...@gmail.com> wrote: >>>>> >>>>>> -1 >>>>>> >>>>>> Unable to use Hive meta-store in pyspark shell. Tried both >>>>>> HiveContext and SparkSession. Both failed. It always uses in-memory >>>>>> catalog. Anybody else hit the same issue? >>>>>> >>>>>> >>>>>> Method 1: SparkSession >>>>>> >>>>>> >>> from pyspark.sql import SparkSession >>>>>> >>>>>> >>> spark = SparkSession.builder.enableHiveSupport().getOrCreate() >>>>>> >>>>>> >>> >>>>>> >>>>>> >>> spark.sql("CREATE TABLE IF NOT EXISTS src (key INT, value >>>>>> STRING)") >>>>>> >>>>>> DataFrame[] >>>>>> >>>>>> >>> spark.sql("LOAD DATA LOCAL INPATH >>>>>> 'examples/src/main/resources/kv1.txt' INTO TABLE src") >>>>>> >>>>>> Traceback (most recent call last): >>>>>> >>>>>> File "<stdin>", line 1, in <module> >>>>>> >>>>>> File >>>>>> "/Users/xiaoli/IdeaProjects/sparkDelivery/python/pyspark/sql/session.py", >>>>>> line 494, in sql >>>>>> >>>>>> return DataFrame(self._jsparkSession.sql(sqlQuery), self._wrapped) >>>>>> >>>>>> File >>>>>> "/Users/xiaoli/IdeaProjects/sparkDelivery/python/lib/py4j-0.10.1-src.zip/py4j/java_gateway.py", >>>>>> line 933, in __call__ >>>>>> >>>>>> File >>>>>> "/Users/xiaoli/IdeaProjects/sparkDelivery/python/pyspark/sql/utils.py", >>>>>> line 57, in deco >>>>>> >>>>>> return f(*a, **kw) >>>>>> >>>>>> File >>>>>> "/Users/xiaoli/IdeaProjects/sparkDelivery/python/lib/py4j-0.10.1-src.zip/py4j/protocol.py", >>>>>> line 312, in get_return_value >>>>>> >>>>>> py4j.protocol.Py4JJavaError: An error occurred while calling o21.sql. >>>>>> >>>>>> : java.lang.UnsupportedOperationException: loadTable is not >>>>>> implemented >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.catalyst.catalog.InMemoryCatalog.loadTable(InMemoryCatalog.scala:297) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.catalyst.catalog.SessionCatalog.loadTable(SessionCatalog.scala:280) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.command.LoadData.run(tables.scala:263) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:57) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:55) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:69) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:115) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:115) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:136) >>>>>> >>>>>> at >>>>>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:133) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:114) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:85) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:85) >>>>>> >>>>>> at org.apache.spark.sql.Dataset.<init>(Dataset.scala:187) >>>>>> >>>>>> at org.apache.spark.sql.Dataset.<init>(Dataset.scala:168) >>>>>> >>>>>> at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:63) >>>>>> >>>>>> at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:541) >>>>>> >>>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>>>> >>>>>> at >>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>>>>> >>>>>> at >>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>>>> >>>>>> at java.lang.reflect.Method.invoke(Method.java:606) >>>>>> >>>>>> at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:237) >>>>>> >>>>>> at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357) >>>>>> >>>>>> at py4j.Gateway.invoke(Gateway.java:280) >>>>>> >>>>>> at >>>>>> py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:128) >>>>>> >>>>>> at py4j.commands.CallCommand.execute(CallCommand.java:79) >>>>>> >>>>>> at py4j.GatewayConnection.run(GatewayConnection.java:211) >>>>>> >>>>>> at java.lang.Thread.run(Thread.java:745) >>>>>> >>>>>> >>>>>> Method 2: Using HiveContext: >>>>>> >>>>>> >>> from pyspark.sql import HiveContext >>>>>> >>>>>> >>> sqlContext = HiveContext(sc) >>>>>> >>>>>> >>> sqlContext.sql("CREATE TABLE IF NOT EXISTS src (key INT, value >>>>>> STRING)") >>>>>> >>>>>> DataFrame[] >>>>>> >>>>>> >>> sqlContext.sql("LOAD DATA LOCAL INPATH >>>>>> 'examples/src/main/resources/kv1.txt' INTO TABLE src") >>>>>> >>>>>> Traceback (most recent call last): >>>>>> >>>>>> File "<stdin>", line 1, in <module> >>>>>> >>>>>> File >>>>>> "/Users/xiaoli/IdeaProjects/sparkDelivery/python/pyspark/sql/context.py", >>>>>> line 346, in sql >>>>>> >>>>>> return self.sparkSession.sql(sqlQuery) >>>>>> >>>>>> File >>>>>> "/Users/xiaoli/IdeaProjects/sparkDelivery/python/pyspark/sql/session.py", >>>>>> line 494, in sql >>>>>> >>>>>> return DataFrame(self._jsparkSession.sql(sqlQuery), self._wrapped) >>>>>> >>>>>> File >>>>>> "/Users/xiaoli/IdeaProjects/sparkDelivery/python/lib/py4j-0.10.1-src.zip/py4j/java_gateway.py", >>>>>> line 933, in __call__ >>>>>> >>>>>> File >>>>>> "/Users/xiaoli/IdeaProjects/sparkDelivery/python/pyspark/sql/utils.py", >>>>>> line 57, in deco >>>>>> >>>>>> return f(*a, **kw) >>>>>> >>>>>> File >>>>>> "/Users/xiaoli/IdeaProjects/sparkDelivery/python/lib/py4j-0.10.1-src.zip/py4j/protocol.py", >>>>>> line 312, in get_return_value >>>>>> >>>>>> py4j.protocol.Py4JJavaError: An error occurred while calling o21.sql. >>>>>> >>>>>> : java.lang.UnsupportedOperationException: loadTable is not >>>>>> implemented >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.catalyst.catalog.InMemoryCatalog.loadTable(InMemoryCatalog.scala:297) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.catalyst.catalog.SessionCatalog.loadTable(SessionCatalog.scala:280) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.command.LoadData.run(tables.scala:263) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:57) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:55) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:69) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:115) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:115) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:136) >>>>>> >>>>>> at >>>>>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:133) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:114) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:85) >>>>>> >>>>>> at >>>>>> org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:85) >>>>>> >>>>>> at org.apache.spark.sql.Dataset.<init>(Dataset.scala:187) >>>>>> >>>>>> at org.apache.spark.sql.Dataset.<init>(Dataset.scala:168) >>>>>> >>>>>> at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:63) >>>>>> >>>>>> at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:541) >>>>>> >>>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>>>> >>>>>> at >>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>>>>> >>>>>> at >>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>>>> >>>>>> at java.lang.reflect.Method.invoke(Method.java:606) >>>>>> >>>>>> at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:237) >>>>>> >>>>>> at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357) >>>>>> >>>>>> at py4j.Gateway.invoke(Gateway.java:280) >>>>>> >>>>>> at >>>>>> py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:128) >>>>>> >>>>>> at py4j.commands.CallCommand.execute(CallCommand.java:79) >>>>>> >>>>>> at py4j.GatewayConnection.run(GatewayConnection.java:211) >>>>>> >>>>>> at java.lang.Thread.run(Thread.java:745) >>>>>> >>>>>> 2016-05-19 12:49 GMT-07:00 Herman van Hövell tot Westerflier < >>>>>> hvanhov...@questtec.nl>: >>>>>> >>>>>>> +1 >>>>>>> >>>>>>> >>>>>>> 2016-05-19 18:20 GMT+02:00 Xiangrui Meng <m...@databricks.com>: >>>>>>> >>>>>>>> +1 >>>>>>>> >>>>>>>> On Thu, May 19, 2016 at 9:18 AM Joseph Bradley < >>>>>>>> jos...@databricks.com> wrote: >>>>>>>> >>>>>>>>> +1 >>>>>>>>> >>>>>>>>> On Wed, May 18, 2016 at 10:49 AM, Reynold Xin <r...@databricks.com >>>>>>>>> > wrote: >>>>>>>>> >>>>>>>>>> Hi Ovidiu-Cristian , >>>>>>>>>> >>>>>>>>>> The best source of truth is change the filter with target version >>>>>>>>>> to 2.1.0. Not a lot of tickets have been targeted yet, but I'd >>>>>>>>>> imagine as >>>>>>>>>> we get closer to 2.0 release, more will be retargeted at 2.1.0. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Wed, May 18, 2016 at 10:43 AM, Ovidiu-Cristian MARCU < >>>>>>>>>> ovidiu-cristian.ma...@inria.fr> wrote: >>>>>>>>>> >>>>>>>>>>> Yes, I can filter.. >>>>>>>>>>> Did that and for example: >>>>>>>>>>> >>>>>>>>>>> https://issues.apache.org/jira/browse/SPARK-15370?jql=project%20%3D%20SPARK%20AND%20resolution%20%3D%20Unresolved%20AND%20affectedVersion%20%3D%202.0.0 >>>>>>>>>>> <https://issues.apache.org/jira/browse/SPARK-15370?jql=project%20=%20SPARK%20AND%20resolution%20=%20Unresolved%20AND%20affectedVersion%20=%202.0.0> >>>>>>>>>>> >>>>>>>>>>> To rephrase: for 2.0 do you have specific issues that are not a >>>>>>>>>>> priority and will released maybe with 2.1 for example? >>>>>>>>>>> >>>>>>>>>>> Keep up the good work! >>>>>>>>>>> >>>>>>>>>>> On 18 May 2016, at 18:19, Reynold Xin <r...@databricks.com> >>>>>>>>>>> wrote: >>>>>>>>>>> >>>>>>>>>>> You can find that by changing the filter to target version = >>>>>>>>>>> 2.0.0. Cheers. >>>>>>>>>>> >>>>>>>>>>> On Wed, May 18, 2016 at 9:00 AM, Ovidiu-Cristian MARCU < >>>>>>>>>>> ovidiu-cristian.ma...@inria.fr> wrote: >>>>>>>>>>> >>>>>>>>>>>> +1 Great, I see the list of resolved issues, do you have a list >>>>>>>>>>>> of known issue you plan to stay with this release? >>>>>>>>>>>> >>>>>>>>>>>> with >>>>>>>>>>>> build/mvn -Pyarn -Phadoop-2.6 -Dhadoop.version=2.7.1 -Phive >>>>>>>>>>>> -Phive-thriftserver -DskipTests clean package >>>>>>>>>>>> >>>>>>>>>>>> mvn -version >>>>>>>>>>>> Apache Maven 3.3.9 (bb52d8502b132ec0a5a3f4c09453c07478323dc5; >>>>>>>>>>>> 2015-11-10T17:41:47+01:00) >>>>>>>>>>>> Maven home: /Users/omarcu/tools/apache-maven-3.3.9 >>>>>>>>>>>> Java version: 1.7.0_80, vendor: Oracle Corporation >>>>>>>>>>>> Java home: >>>>>>>>>>>> /Library/Java/JavaVirtualMachines/jdk1.7.0_80.jdk/Contents/Home/jre >>>>>>>>>>>> Default locale: en_US, platform encoding: UTF-8 >>>>>>>>>>>> OS name: "mac os x", version: "10.11.5", arch: "x86_64", >>>>>>>>>>>> family: “mac" >>>>>>>>>>>> >>>>>>>>>>>> [INFO] Reactor Summary: >>>>>>>>>>>> [INFO] >>>>>>>>>>>> [INFO] Spark Project Parent POM ........................... >>>>>>>>>>>> SUCCESS [ 2.635 s] >>>>>>>>>>>> [INFO] Spark Project Tags ................................. >>>>>>>>>>>> SUCCESS [ 1.896 s] >>>>>>>>>>>> [INFO] Spark Project Sketch ............................... >>>>>>>>>>>> SUCCESS [ 2.560 s] >>>>>>>>>>>> [INFO] Spark Project Networking ........................... >>>>>>>>>>>> SUCCESS [ 6.533 s] >>>>>>>>>>>> [INFO] Spark Project Shuffle Streaming Service ............ >>>>>>>>>>>> SUCCESS [ 4.176 s] >>>>>>>>>>>> [INFO] Spark Project Unsafe ............................... >>>>>>>>>>>> SUCCESS [ 4.809 s] >>>>>>>>>>>> [INFO] Spark Project Launcher ............................. >>>>>>>>>>>> SUCCESS [ 6.242 s] >>>>>>>>>>>> [INFO] Spark Project Core ................................. >>>>>>>>>>>> SUCCESS [01:20 min] >>>>>>>>>>>> [INFO] Spark Project GraphX ............................... >>>>>>>>>>>> SUCCESS [ 9.148 s] >>>>>>>>>>>> [INFO] Spark Project Streaming ............................ >>>>>>>>>>>> SUCCESS [ 22.760 s] >>>>>>>>>>>> [INFO] Spark Project Catalyst ............................. >>>>>>>>>>>> SUCCESS [ 50.783 s] >>>>>>>>>>>> [INFO] Spark Project SQL .................................. >>>>>>>>>>>> SUCCESS [01:05 min] >>>>>>>>>>>> [INFO] Spark Project ML Local Library ..................... >>>>>>>>>>>> SUCCESS [ 4.281 s] >>>>>>>>>>>> [INFO] Spark Project ML Library ........................... >>>>>>>>>>>> SUCCESS [ 54.537 s] >>>>>>>>>>>> [INFO] Spark Project Tools ................................ >>>>>>>>>>>> SUCCESS [ 0.747 s] >>>>>>>>>>>> [INFO] Spark Project Hive ................................. >>>>>>>>>>>> SUCCESS [ 33.032 s] >>>>>>>>>>>> [INFO] Spark Project HiveContext Compatibility ............ >>>>>>>>>>>> SUCCESS [ 3.198 s] >>>>>>>>>>>> [INFO] Spark Project REPL ................................. >>>>>>>>>>>> SUCCESS [ 3.573 s] >>>>>>>>>>>> [INFO] Spark Project YARN Shuffle Service ................. >>>>>>>>>>>> SUCCESS [ 4.617 s] >>>>>>>>>>>> [INFO] Spark Project YARN ................................. >>>>>>>>>>>> SUCCESS [ 7.321 s] >>>>>>>>>>>> [INFO] Spark Project Hive Thrift Server ................... >>>>>>>>>>>> SUCCESS [ 16.496 s] >>>>>>>>>>>> [INFO] Spark Project Assembly ............................. >>>>>>>>>>>> SUCCESS [ 2.300 s] >>>>>>>>>>>> [INFO] Spark Project External Flume Sink .................. >>>>>>>>>>>> SUCCESS [ 4.219 s] >>>>>>>>>>>> [INFO] Spark Project External Flume ....................... >>>>>>>>>>>> SUCCESS [ 6.987 s] >>>>>>>>>>>> [INFO] Spark Project External Flume Assembly .............. >>>>>>>>>>>> SUCCESS [ 1.465 s] >>>>>>>>>>>> [INFO] Spark Integration for Kafka 0.8 .................... >>>>>>>>>>>> SUCCESS [ 6.891 s] >>>>>>>>>>>> [INFO] Spark Project Examples ............................. >>>>>>>>>>>> SUCCESS [ 13.465 s] >>>>>>>>>>>> [INFO] Spark Project External Kafka Assembly .............. >>>>>>>>>>>> SUCCESS [ 2.815 s] >>>>>>>>>>>> [INFO] >>>>>>>>>>>> ------------------------------------------------------------------------ >>>>>>>>>>>> [INFO] BUILD SUCCESS >>>>>>>>>>>> [INFO] >>>>>>>>>>>> ------------------------------------------------------------------------ >>>>>>>>>>>> [INFO] Total time: 07:04 min >>>>>>>>>>>> [INFO] Finished at: 2016-05-18T17:55:33+02:00 >>>>>>>>>>>> [INFO] Final Memory: 90M/824M >>>>>>>>>>>> [INFO] >>>>>>>>>>>> ------------------------------------------------------------------------ >>>>>>>>>>>> >>>>>>>>>>>> On 18 May 2016, at 16:28, Sean Owen <so...@cloudera.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>> I think it's a good idea. Although releases have been preceded >>>>>>>>>>>> before >>>>>>>>>>>> by release candidates for developers, it would be good to get a >>>>>>>>>>>> formal >>>>>>>>>>>> preview/beta release ratified for public consumption ahead of a >>>>>>>>>>>> new >>>>>>>>>>>> major release. Better to have a little more testing in the wild >>>>>>>>>>>> to >>>>>>>>>>>> identify problems before 2.0.0 is finalized. >>>>>>>>>>>> >>>>>>>>>>>> +1 to the release. License, sigs, etc check out. On Ubuntu 16 + >>>>>>>>>>>> Java >>>>>>>>>>>> 8, compilation and tests succeed for "-Pyarn -Phive >>>>>>>>>>>> -Phive-thriftserver -Phadoop-2.6". >>>>>>>>>>>> >>>>>>>>>>>> On Wed, May 18, 2016 at 6:40 AM, Reynold Xin <r...@apache.org> >>>>>>>>>>>> wrote: >>>>>>>>>>>> >>>>>>>>>>>> Hi, >>>>>>>>>>>> >>>>>>>>>>>> In the past the Apache Spark community have created preview >>>>>>>>>>>> packages (not >>>>>>>>>>>> official releases) and used those as opportunities to ask >>>>>>>>>>>> community members >>>>>>>>>>>> to test the upcoming versions of Apache Spark. Several people >>>>>>>>>>>> in the Apache >>>>>>>>>>>> community have suggested we conduct votes for these preview >>>>>>>>>>>> packages and >>>>>>>>>>>> turn them into formal releases by the Apache foundation's >>>>>>>>>>>> standard. Preview >>>>>>>>>>>> releases are not meant to be functional, i.e. they can and >>>>>>>>>>>> highly likely >>>>>>>>>>>> will contain critical bugs or documentation errors, but we will >>>>>>>>>>>> be able to >>>>>>>>>>>> post them to the project's website to get wider feedback. They >>>>>>>>>>>> should >>>>>>>>>>>> satisfy the legal requirements of Apache's release policy >>>>>>>>>>>> (http://www.apache.org/dev/release.html) such as having proper >>>>>>>>>>>> licenses. >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Please vote on releasing the following candidate as Apache >>>>>>>>>>>> Spark version >>>>>>>>>>>> 2.0.0-preview. The vote is open until Friday, May 20, 2015 at >>>>>>>>>>>> 11:00 PM PDT >>>>>>>>>>>> and passes if a majority of at least 3 +1 PMC votes are cast. >>>>>>>>>>>> >>>>>>>>>>>> [ ] +1 Release this package as Apache Spark 2.0.0-preview >>>>>>>>>>>> [ ] -1 Do not release this package because ... >>>>>>>>>>>> >>>>>>>>>>>> To learn more about Apache Spark, please see >>>>>>>>>>>> http://spark.apache.org/ >>>>>>>>>>>> >>>>>>>>>>>> The tag to be voted on is 2.0.0-preview >>>>>>>>>>>> (8f5a04b6299e3a47aca13cbb40e72344c0114860) >>>>>>>>>>>> >>>>>>>>>>>> The release files, including signatures, digests, etc. can be >>>>>>>>>>>> found at: >>>>>>>>>>>> >>>>>>>>>>>> http://home.apache.org/~pwendell/spark-releases/spark-2.0.0-preview-bin/ >>>>>>>>>>>> >>>>>>>>>>>> Release artifacts are signed with the following key: >>>>>>>>>>>> https://people.apache.org/keys/committer/pwendell.asc >>>>>>>>>>>> >>>>>>>>>>>> The documentation corresponding to this release can be found at: >>>>>>>>>>>> >>>>>>>>>>>> http://home.apache.org/~pwendell/spark-releases/spark-2.0.0-preview-docs/ >>>>>>>>>>>> >>>>>>>>>>>> The list of resolved issues are: >>>>>>>>>>>> >>>>>>>>>>>> https://issues.apache.org/jira/browse/SPARK-15351?jql=project%20%3D%20SPARK%20AND%20fixVersion%20%3D%202.0.0 >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> If you are a Spark user, you can help us test this release by >>>>>>>>>>>> taking an >>>>>>>>>>>> existing Apache Spark workload and running on this candidate, >>>>>>>>>>>> then reporting >>>>>>>>>>>> any regressions. >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> --------------------------------------------------------------------- >>>>>>>>>>>> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org >>>>>>>>>>>> For additional commands, e-mail: dev-h...@spark.apache.org >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >