If this is something we expect to mostly impact new users I think we can
push them towards Spark 3 instead of introducing a behaviour change in 2.4.6

On Wed, Jun 3, 2020 at 12:34 PM Mridul Muralidharan <mri...@gmail.com>
wrote:

>
>   Is this a behavior change in 2.4.x from earlier version ?
> Or are we proposing to introduce  a functionality to help with adoption ?
>
> Regards,
> Mridul
>
>
> On Wed, Jun 3, 2020 at 10:32 AM Xiao Li <gatorsm...@gmail.com> wrote:
>
>> Yes. Spark 3.0 RC2 works well.
>>
>> I think the current behavior in Spark 2.4 affects the adoption,
>> especially for the new users who want to try Spark in their local
>> environment.
>>
>> It impacts all our built-in clients, like Scala Shell and PySpark. Should
>> we consider back-porting it to 2.4?
>>
>> Although this fixes the bug, it will also introduce the behavior change.
>> We should publicly document it and mention it in the release note. Let us
>> review it more carefully and understand the risk and impact.
>>
>> Thanks,
>>
>> Xiao
>>
>> Nicholas Chammas <nicholas.cham...@gmail.com> 于2020年6月3日周三 上午10:12写道:
>>
>>> I believe that was fixed in 3.0 and there was a decision not to backport
>>> the fix: SPARK-31170 <https://issues.apache.org/jira/browse/SPARK-31170>
>>>
>>> On Wed, Jun 3, 2020 at 1:04 PM Xiao Li <gatorsm...@gmail.com> wrote:
>>>
>>>> Just downloaded it in my local macbook. Trying to create a table using
>>>> the pre-built PySpark. It sounds like the conf "spark.sql.warehouse.dir"
>>>> does not take an effect. It is trying to create a directory in
>>>> "file:/user/hive/warehouse/t1". I have not done any investigation yet. Have
>>>> any of you hit the same issue?
>>>>
>>>> C02XT0U7JGH5:bin lixiao$ ./pyspark --conf
>>>> spark.sql.warehouse.dir="/Users/lixiao/Downloads/spark-2.4.6-bin-hadoop2.6"
>>>>
>>>> Python 2.7.16 (default, Jan 27 2020, 04:46:15)
>>>>
>>>> [GCC 4.2.1 Compatible Apple LLVM 10.0.1 (clang-1001.0.37.14)] on darwin
>>>>
>>>> Type "help", "copyright", "credits" or "license" for more information.
>>>>
>>>> 20/06/03 09:56:11 WARN NativeCodeLoader: Unable to load native-hadoop
>>>> library for your platform... using builtin-java classes where applicable
>>>>
>>>> Using Spark's default log4j profile:
>>>> org/apache/spark/log4j-defaults.properties
>>>>
>>>> Setting default log level to "WARN".
>>>>
>>>> To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use
>>>> setLogLevel(newLevel).
>>>>
>>>> Welcome to
>>>>
>>>>       ____              __
>>>>
>>>>      / __/__  ___ _____/ /__
>>>>
>>>>     _\ \/ _ \/ _ `/ __/  '_/
>>>>
>>>>    /__ / .__/\_,_/_/ /_/\_\   version 2.4.6
>>>>
>>>>       /_/
>>>>
>>>>
>>>> Using Python version 2.7.16 (default, Jan 27 2020 04:46:15)
>>>>
>>>> SparkSession available as 'spark'.
>>>>
>>>> >>> spark.sql("set spark.sql.warehouse.dir").show(truncate=False)
>>>>
>>>>
>>>> +-----------------------+-------------------------------------------------+
>>>>
>>>> |key                    |value
>>>>     |
>>>>
>>>>
>>>> +-----------------------+-------------------------------------------------+
>>>>
>>>>
>>>> |spark.sql.warehouse.dir|/Users/lixiao/Downloads/spark-2.4.6-bin-hadoop2.6|
>>>>
>>>>
>>>> +-----------------------+-------------------------------------------------+
>>>>
>>>>
>>>> >>> spark.sql("create table t1 (col1 int)")
>>>>
>>>> 20/06/03 09:56:29 WARN HiveMetaStore: Location:
>>>> file:/user/hive/warehouse/t1 specified for non-external table:t1
>>>>
>>>> Traceback (most recent call last):
>>>>
>>>>   File "<stdin>", line 1, in <module>
>>>>
>>>>   File
>>>> "/Users/lixiao/Downloads/spark-2.4.6-bin-hadoop2.6/python/pyspark/sql/session.py",
>>>> line 767, in sql
>>>>
>>>>     return DataFrame(self._jsparkSession.sql(sqlQuery), self._wrapped)
>>>>
>>>>   File
>>>> "/Users/lixiao/Downloads/spark-2.4.6-bin-hadoop2.6/python/lib/py4j-0.10.7-src.zip/py4j/java_gateway.py",
>>>> line 1257, in __call__
>>>>
>>>>   File
>>>> "/Users/lixiao/Downloads/spark-2.4.6-bin-hadoop2.6/python/pyspark/sql/utils.py",
>>>> line 69, in deco
>>>>
>>>>     raise AnalysisException(s.split(': ', 1)[1], stackTrace)
>>>>
>>>> pyspark.sql.utils.AnalysisException:
>>>> u'org.apache.hadoop.hive.ql.metadata.HiveException:
>>>> MetaException(message:file:/user/hive/warehouse/t1 is not a directory or
>>>> unable to create one);'
>>>>
>>>> Dongjoon Hyun <dongjoon.h...@gmail.com> 于2020年6月3日周三 上午9:18写道:
>>>>
>>>>> +1
>>>>>
>>>>> Bests,
>>>>> Dongjoon
>>>>>
>>>>> On Wed, Jun 3, 2020 at 5:59 AM Tom Graves <tgraves...@yahoo.com.invalid>
>>>>> wrote:
>>>>>
>>>>>>  +1
>>>>>>
>>>>>> Tom
>>>>>>
>>>>>> On Sunday, May 31, 2020, 06:47:09 PM CDT, Holden Karau <
>>>>>> hol...@pigscanfly.ca> wrote:
>>>>>>
>>>>>>
>>>>>> Please vote on releasing the following candidate as Apache Spark
>>>>>> version 2.4.6.
>>>>>>
>>>>>> The vote is open until June 5th at 9AM PST and passes if a majority
>>>>>> +1 PMC votes are cast, with a minimum of 3 +1 votes.
>>>>>>
>>>>>> [ ] +1 Release this package as Apache Spark 2.4.6
>>>>>> [ ] -1 Do not release this package because ...
>>>>>>
>>>>>> To learn more about Apache Spark, please see http://spark.apache.org/
>>>>>>
>>>>>> There are currently no issues targeting 2.4.6 (try project = SPARK
>>>>>> AND "Target Version/s" = "2.4.6" AND status in (Open, Reopened, "In
>>>>>> Progress"))
>>>>>>
>>>>>> The tag to be voted on is v2.4.6-rc8 (commit
>>>>>> 807e0a484d1de767d1f02bd8a622da6450bdf940):
>>>>>> https://github.com/apache/spark/tree/v2.4.6-rc8
>>>>>>
>>>>>> The release files, including signatures, digests, etc. can be found
>>>>>> at:
>>>>>> https://dist.apache.org/repos/dist/dev/spark/v2.4.6-rc8-bin/
>>>>>>
>>>>>> Signatures used for Spark RCs can be found in this file:
>>>>>> https://dist.apache.org/repos/dist/dev/spark/KEYS
>>>>>>
>>>>>> The staging repository for this release can be found at:
>>>>>>
>>>>>> https://repository.apache.org/content/repositories/orgapachespark-1349/
>>>>>>
>>>>>> The documentation corresponding to this release can be found at:
>>>>>> https://dist.apache.org/repos/dist/dev/spark/v2.4.6-rc8-docs/
>>>>>>
>>>>>> The list of bug fixes going into 2.4.6 can be found at the following
>>>>>> URL:
>>>>>> https://issues.apache.org/jira/projects/SPARK/versions/12346781
>>>>>>
>>>>>> This release is using the release script of the tag v2.4.6-rc8.
>>>>>>
>>>>>> FAQ
>>>>>>
>>>>>> =========================
>>>>>> What happened to the other RCs?
>>>>>> =========================
>>>>>>
>>>>>> The parallel maven build caused some flakiness so I wasn't
>>>>>> comfortable releasing them. I backported the fix from the 3.0 branch for
>>>>>> this release. I've got a proposed change to the build script so that we
>>>>>> only push tags when once the build is a success for the future, but it 
>>>>>> does
>>>>>> not block this release.
>>>>>>
>>>>>> =========================
>>>>>> How can I help test this release?
>>>>>> =========================
>>>>>>
>>>>>> If you are a Spark user, you can help us test this release by taking
>>>>>> an existing Spark workload and running on this release candidate, then
>>>>>> reporting any regressions.
>>>>>>
>>>>>> If you're working in PySpark you can set up a virtual env and install
>>>>>> the current RC and see if anything important breaks, in the Java/Scala
>>>>>> you can add the staging repository to your projects resolvers and test
>>>>>> with the RC (make sure to clean up the artifact cache before/after so
>>>>>> you don't end up building with an out of date RC going forward).
>>>>>>
>>>>>> ===========================================
>>>>>> What should happen to JIRA tickets still targeting 2.4.6?
>>>>>> ===========================================
>>>>>>
>>>>>> The current list of open tickets targeted at 2.4.6 can be found at:
>>>>>> https://issues.apache.org/jira/projects/SPARK and search for "Target
>>>>>> Version/s" = 2.4.6
>>>>>>
>>>>>> Committers should look at those and triage. Extremely important bug
>>>>>> fixes, documentation, and API tweaks that impact compatibility should
>>>>>> be worked on immediately. Everything else please retarget to an
>>>>>> appropriate release.
>>>>>>
>>>>>> ==================
>>>>>> But my bug isn't fixed?
>>>>>> ==================
>>>>>>
>>>>>> In order to make timely releases, we will typically not hold the
>>>>>> release unless the bug in question is a regression from the previous
>>>>>> release. That being said, if there is something which is a regression
>>>>>> that has not been correctly targeted please ping me or a committer to
>>>>>> help target the issue.
>>>>>>
>>>>>>
>>>>>> --
>>>>>> Twitter: https://twitter.com/holdenkarau
>>>>>> Books (Learning Spark, High Performance Spark, etc.):
>>>>>> https://amzn.to/2MaRAG9  <https://amzn.to/2MaRAG9>
>>>>>> YouTube Live Streams: https://www.youtube.com/user/holdenkarau
>>>>>>
>>>>> --
Twitter: https://twitter.com/holdenkarau
Books (Learning Spark, High Performance Spark, etc.):
https://amzn.to/2MaRAG9  <https://amzn.to/2MaRAG9>
YouTube Live Streams: https://www.youtube.com/user/holdenkarau

Reply via email to