Hi, below sounds like something that someone will have experienced... I have external tables of parquet files with a hive table defined on top of the data. I dont manage/know the details of how the data lands. For some tables no issues when querying through spark. But for others there is an issue: looks like the datatype for hive is timestamp, but the parquet file contains an integer number =microseconds: if I access the table in spark sql I get:
Unable to create Parquet converter for data type “timestamp” whose Parquet type is optional int64 airac_wef (TIMESTAMP(MICROS,false)) OR Parquet column cannot be converted in file abfs://somelocation/table/partition=484/000000_0. Column: [atimefield], Expected: timestamp, Found: INT64 Anyone encountered this? I tried several sorts of CAST but no success yet. I see similar problems on forums (like this: https://stackoverflow.com/questions/59096125/spark-2-4-parquet-column-cannot-be-converted-in-file-column-impressions-exp ) but no solution. Thanks for input!