GitHub user dongjoon-hyun opened a pull request:
https://github.com/apache/spark/pull/16463
[SPARK-18877][SQL][BACKPORT-2.1] `CSVInferSchema.inferField` on DecimalType
should find a common type with `typeSoFar`
## What changes were proposed in this pull request?
CSV type inferencing causes `IllegalArgumentException` on decimal numbers
with heterogeneous precisions and scales because the current logic uses the
last decimal type in a **partition**. Specifically, `inferRowType`, the
**seqOp** of **aggregate**, returns the last decimal type. This PR fixes it to
use `findTightestCommonType`.
**decimal.csv**
```
9.03E+12
1.19E+11
```
**BEFORE**
```scala
scala> spark.read.format("csv").option("inferSchema",
true).load("decimal.csv").printSchema
root
|-- _c0: decimal(3,-9) (nullable = true)
scala> spark.read.format("csv").option("inferSchema",
true).load("decimal.csv").show
16/12/16 14:32:49 ERROR Executor: Exception in task 0.0 in stage 4.0 (TID 4)
java.lang.IllegalArgumentException: requirement failed: Decimal precision 4
exceeds max precision 3
```
**AFTER**
```scala
scala> spark.read.format("csv").option("inferSchema",
true).load("decimal.csv").printSchema
root
|-- _c0: decimal(4,-9) (nullable = true)
scala> spark.read.format("csv").option("inferSchema",
true).load("decimal.csv").show
+---------+
| _c0|
+---------+
|9.030E+12|
| 1.19E+11|
+---------+
```
## How was this patch tested?
Pass the newly add test case.
You can merge this pull request into a Git repository by running:
$ git pull https://github.com/dongjoon-hyun/spark SPARK-18877-BACKPORT-21
Alternatively you can review and apply these changes as the patch at:
https://github.com/apache/spark/pull/16463.patch
To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:
This closes #16463
----
commit 7d4f31eb32b35f89ea5d3c7c8fa9c08e025d3ee1
Author: Dongjoon Hyun <[email protected]>
Date: 2017-01-03T15:06:50Z
[SPARK-18877][SQL][BACKPORT-2.1] `CSVInferSchema.inferField` on DecimalType
should find a common type with `typeSoFar`
CSV type inferencing causes `IllegalArgumentException` on decimal numbers
with heterogeneous precisions and scales because the current logic uses the
last decimal type in a **partition**. Specifically, `inferRowType`, the
**seqOp** of **aggregate**, returns the last decimal type. This PR fixes it to
use `findTightestCommonType`.
**decimal.csv**
```
9.03E+12
1.19E+11
```
**BEFORE**
```scala
scala> spark.read.format("csv").option("inferSchema",
true).load("decimal.csv").printSchema
root
|-- _c0: decimal(3,-9) (nullable = true)
scala> spark.read.format("csv").option("inferSchema",
true).load("decimal.csv").show
16/12/16 14:32:49 ERROR Executor: Exception in task 0.0 in stage 4.0 (TID 4)
java.lang.IllegalArgumentException: requirement failed: Decimal precision 4
exceeds max precision 3
```
**AFTER**
```scala
scala> spark.read.format("csv").option("inferSchema",
true).load("decimal.csv").printSchema
root
|-- _c0: decimal(4,-9) (nullable = true)
scala> spark.read.format("csv").option("inferSchema",
true).load("decimal.csv").show
+---------+
| _c0|
+---------+
|9.030E+12|
| 1.19E+11|
+---------+
```
Pass the newly add test case.
Author: Dongjoon Hyun <[email protected]>
Closes #16320 from dongjoon-hyun/SPARK-18877.
----
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]