It On Mon, Apr 24, 2017 at 10:33 AM, Michael Allman <mich...@videoamp.com> wrote:
> The trouble we ran into is that this upgrade was blocking access to our > tables, and we didn't know why. This sounds like a kind of migration > operation, but it was not apparent that this was the case. It took an > expert examining a stack trace and source code to figure this out. Would a > more naive end user be able to debug this issue? Maybe we're an unusual > case, but our particular experience was pretty bad. I have my doubts that > the schema inference on our largest tables would ever complete without > throwing some kind of timeout (which we were in fact receiving) or the end > user just giving up and killing our job. We ended up doing a rollback while > we investigated the source of the issue. In our case, INFER_NEVER is > clearly the best configuration. We're going to add that to our default > configuration files. > > My expectation is that a minor point release is a pretty safe bug fix > release. We were a bit hasty in not doing better due diligence pre-upgrade. > > One suggestion the Spark team might consider is releasing 2.1.1 with > INVER_NEVER and 2.2.0 with INFER_AND_SAVE. Clearly some kind of up-front > migration notes would help in identifying this new behavior in 2.2. > > Thanks, > > Michael > > > On Apr 24, 2017, at 2:09 AM, Wenchen Fan <wenc...@databricks.com> wrote: > > see https://issues.apache.org/jira/browse/SPARK-19611 > > On Mon, Apr 24, 2017 at 2:22 PM, Holden Karau <hol...@pigscanfly.ca> > wrote: > >> Whats the regression this fixed in 2.1 from 2.0? >> >> On Fri, Apr 21, 2017 at 7:45 PM, Wenchen Fan <wenc...@databricks.com> >> wrote: >> >>> IIRC, the new "spark.sql.hive.caseSensitiveInferenceMode" stuff will >>> only scan all table files only once, and write back the inferred schema to >>> metastore so that we don't need to do the schema inference again. >>> >>> So technically this will introduce a performance regression for the >>> first query, but compared to branch-2.0, it's not performance regression. >>> And this patch fixed a regression in branch-2.1, which can run in >>> branch-2.0. Personally, I think we should keep INFER_AND_SAVE as the >>> default mode. >>> >>> + [Eric], what do you think? >>> >>> On Sat, Apr 22, 2017 at 1:37 AM, Michael Armbrust < >>> mich...@databricks.com> wrote: >>> >>>> Thanks for pointing this out, Michael. Based on the conversation on >>>> the PR >>>> <https://github.com/apache/spark/pull/16944#issuecomment-285529275> >>>> this seems like a risky change to include in a release branch with a >>>> default other than NEVER_INFER. >>>> >>>> +Wenchen? What do you think? >>>> >>>> On Thu, Apr 20, 2017 at 4:14 PM, Michael Allman <mich...@videoamp.com> >>>> wrote: >>>> >>>>> We've identified the cause of the change in behavior. It is related to >>>>> the SQL conf key "spark.sql.hive.caseSensitiveInferenceMode". This >>>>> key and its related functionality was absent from our previous build. The >>>>> default setting in the current build was causing Spark to attempt to scan >>>>> all table files during query analysis. Changing this setting to >>>>> NEVER_INFER >>>>> disabled this operation and resolved the issue we had. >>>>> >>>>> Michael >>>>> >>>>> >>>>> On Apr 20, 2017, at 3:42 PM, Michael Allman <mich...@videoamp.com> >>>>> wrote: >>>>> >>>>> I want to caution that in testing a build from this morning's >>>>> branch-2.1 we found that Hive partition pruning was not working. We found >>>>> that Spark SQL was fetching all Hive table partitions for a very simple >>>>> query whereas in a build from several weeks ago it was fetching only the >>>>> required partitions. I cannot currently think of a reason for the >>>>> regression outside of some difference between branch-2.1 from our previous >>>>> build and branch-2.1 from this morning. >>>>> >>>>> That's all I know right now. We are actively investigating to find the >>>>> root cause of this problem, and specifically whether this is a problem in >>>>> the Spark codebase or not. I will report back when I have an answer to >>>>> that >>>>> question. >>>>> >>>>> Michael >>>>> >>>>> >>>>> On Apr 18, 2017, at 11:59 AM, Michael Armbrust <mich...@databricks.com> >>>>> wrote: >>>>> >>>>> Please vote on releasing the following candidate as Apache Spark >>>>> version 2.1.1. The vote is open until Fri, April 21st, 2018 at 13:00 >>>>> PST and passes if a majority of at least 3 +1 PMC votes are cast. >>>>> >>>>> [ ] +1 Release this package as Apache Spark 2.1.1 >>>>> [ ] -1 Do not release this package because ... >>>>> >>>>> >>>>> To learn more about Apache Spark, please see http://spark.apache.org/ >>>>> >>>>> The tag to be voted on is v2.1.1-rc3 >>>>> <https://github.com/apache/spark/tree/v2.1.1-rc3> (2ed19cff2f6ab79 >>>>> a718526e5d16633412d8c4dd4) >>>>> >>>>> List of JIRA tickets resolved can be found with this filter >>>>> <https://issues.apache.org/jira/browse/SPARK-20134?jql=project%20%3D%20SPARK%20AND%20fixVersion%20%3D%202.1.1> >>>>> . >>>>> >>>>> The release files, including signatures, digests, etc. can be found at: >>>>> http://home.apache.org/~pwendell/spark-releases/spark-2.1.1-rc3-bin/ >>>>> >>>>> Release artifacts are signed with the following key: >>>>> https://people.apache.org/keys/committer/pwendell.asc >>>>> >>>>> The staging repository for this release can be found at: >>>>> https://repository.apache.org/content/repositories/orgapache >>>>> spark-1230/ >>>>> >>>>> The documentation corresponding to this release can be found at: >>>>> http://people.apache.org/~pwendell/spark-releases/spark-2.1. >>>>> 1-rc3-docs/ >>>>> >>>>> >>>>> *FAQ* >>>>> >>>>> *How can I help test this release?* >>>>> >>>>> If you are a Spark user, you can help us test this release by taking >>>>> an existing Spark workload and running on this release candidate, then >>>>> reporting any regressions. >>>>> >>>>> *What should happen to JIRA tickets still targeting 2.1.1?* >>>>> >>>>> Committers should look at those and triage. Extremely important bug >>>>> fixes, documentation, and API tweaks that impact compatibility should be >>>>> worked on immediately. Everything else please retarget to 2.1.2 or 2.2.0. >>>>> >>>>> *But my bug isn't fixed!??!* >>>>> >>>>> In order to make timely releases, we will typically not hold the >>>>> release unless the bug in question is a regression from 2.1.0. >>>>> >>>>> *What happened to RC1?* >>>>> >>>>> There were issues with the release packaging and as a result was >>>>> skipped. >>>>> >>>>> >>>>> >>>>> >>>> >>> >> >> >> -- >> Cell : 425-233-8271 <(425)%20233-8271> >> Twitter: https://twitter.com/holdenkarau >> > > > -- Cell : 425-233-8271 Twitter: https://twitter.com/holdenkarau