[
https://issues.apache.org/jira/browse/SPARK-32021?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Dongjoon Hyun updated SPARK-32021:
----------------------------------
Fix Version/s: (was: 3.0.0)
3.0.1
> make_interval does not accept seconds >100
> ------------------------------------------
>
> Key: SPARK-32021
> URL: https://issues.apache.org/jira/browse/SPARK-32021
> Project: Spark
> Issue Type: Bug
> Components: SQL
> Affects Versions: 3.0.0
> Reporter: Juliusz Sompolski
> Assignee: Maxim Gekk
> Priority: Major
> Fix For: 3.0.1, 3.1.0
>
>
> In make_interval(years, months, weeks, days, hours, mins, secs), secs are
> defined as Decimal(8, 6), which turns into null if the value of the
> expression overflows 100 seconds.
> Larger seconds values should be allowed.
> This has been reported by Simba, who wants to use make_interval to implement
> translation for TIMESTAMP_ADD ODBC function in Spark 3.0.
> ODBC {fn TIMESTAMPADD(SECOND, integer_exp, timestamp} fails when integer_exp
> returns seconds values >= 100.
--
This message was sent by Atlassian Jira
(v8.3.4#803005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]