sketchmind commented on PR #25536: URL: https://github.com/apache/spark/pull/25536#issuecomment-1195128550
> But people can always create an empty table first and then insert? If we really want to support this case, we should allow people to define table schema in CTAS (or Scala createOrReplace) so that it can overwrite dataframe schema. CreateOrReplace(create an empty table first and then insert) is a very frequently used function,it's just that most tables are not sensitive to whether a column is NULL, especially Hive, so this issue is not widely mentioned. I think it is better for spark to leave this part of the logic to the catalog developers to manage instead of using NULL to create tables for all databases. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: [email protected] For queries about this service, please contact Infrastructure at: [email protected] --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
