Github user dmateusp commented on the issue:
https://github.com/apache/spark/pull/21706
Just added it to the FunctionRegistry:
```scala
scala> spark.sql("DESC function calendarinterval").show(truncate=false)
+--------------------------------------------------------------------------------------------------+
|function_desc
|
+--------------------------------------------------------------------------------------------------+
|Function: calendarinterval
|
|Class: org.apache.spark.sql.catalyst.expressions.Cast
|
|Usage: calendarinterval(expr) - Casts the value `expr` to the target data
type `calendarinterval`.|
+--------------------------------------------------------------------------------------------------+
scala> spark.sql("select calendarinterval('interval 10 days')").show()
+------------------------------------------+
|CAST(interval 10 days AS CALENDARINTERVAL)|
+------------------------------------------+
| interval 1 weeks ...|
+------------------------------------------+
```
I sent an email to the dev mailing list because I'm having troubles
understanding how the sql-tests work (the structure, and how to run them). I'll
add tests in there as soon as I figure that out :)
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]