David Mollitor created HIVE-24693:
-------------------------------------

             Summary: Parquet Timestamp Values Read/Write Very Slow
                 Key: HIVE-24693
                 URL: https://issues.apache.org/jira/browse/HIVE-24693
             Project: Hive
          Issue Type: Improvement
            Reporter: David Mollitor
            Assignee: David Mollitor


Parquet {{DataWriteableWriter}} relias on {{NanoTimeUtils}} to convert a 
timestamp object into a binary value.  The way in which it does this,... it 
calls {{toString()}} on the timestamp object, and then parses the String.  This 
particular timestamp do not carry a timezone, so the string is something like:

{{2021-21-03 12:32:23.0000...}}

The parse code tries to parse the string assuming there is a time zone, and if 
not, falls-back and applies the provided "default time zone".  As was noted in 
[HIVE-24353], if something fails to parse, it is very expensive to try to parse 
again.  So, for each timestamp in the Parquet file, it:

* Builds a string from the time stamp
* Parses it (throws an exception, parses again)

There is no need to do this kind of string manipulations/parsing, it should 
just be using the epoch millis/seconds/time stored internal to the Timestamp 
object.

{code:java}
  // Converts Timestamp to TimestampTZ.
  public static TimestampTZ convert(Timestamp ts, ZoneId defaultTimeZone) {
    return parse(ts.toString(), defaultTimeZone);
  }
{code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to