Github user BryanCutler commented on the issue:
https://github.com/apache/spark/pull/22913
I'm a little against adding this because the Arrow Java Vectors used so far
were done to match the internal data of Spark, to keep things simple and avoid
lots of conversions on the Java side. Conversions to supported types are being
done before reading the data in Java.
For instance, there are lots of timestamp types for other time units, but
we only accept microseconds with tz in Java (to match Spark) and do any
necessary conversions in Python before writing the Arrow data. Can this also be
done in R?
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]