[
https://issues.apache.org/jira/browse/HUDI-5452?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Jonathan Vexler closed HUDI-5452.
---------------------------------
Resolution: Not A Problem
> Spark-sql long datatype conversion to bigint in hive causes issues with alter
> table
> -----------------------------------------------------------------------------------
>
> Key: HUDI-5452
> URL: https://issues.apache.org/jira/browse/HUDI-5452
> Project: Apache Hudi
> Issue Type: Bug
> Components: spark-sql
> Reporter: Jonathan Vexler
> Priority: Major
> Attachments: AlterTableIssue.txt
>
>
> Commands run to get this error: [^AlterTableIssue.txt] . When trying to alter
> the table with long in the schema we get this error.
> When calling describe table we get
> {code:java}
> spark-sql> describe test_table;
> _hoodie_commit_time string
> _hoodie_commit_seqno string
> _hoodie_record_key string
> _hoodie_partition_path string
> _hoodie_file_name string
> id int
> name string
> price double
> ts bigint
> dt string {code}
> We think that it is having issues with long turning into bigint. When we
> created a different table with the same schema but ts had type int instead of
> long, we were able to alter the schema.
--
This message was sent by Atlassian Jira
(v8.20.10#820010)