jhchee opened a new issue, #8499: URL: https://github.com/apache/hudi/issues/8499
**Describe the problem you faced** Currently, if the insert statement specifies less columns than the target table size, the following exception will be thrown: ``` Exception in thread "main" java.lang.AssertionError: assertion failed: The number of insert assignments[3] must equal to the targetTable field size[4] ``` For a wide table that has 1000 columns, the user is required to specifies the all the columns with default values null to avoid this exception. Can we support partial insert in merge (default to null if not specified) into command so the developer can maintain clean SQL statement. E.g. Delta merge command has already supported for this <img width="830" alt="image" src="https://user-images.githubusercontent.com/33974196/233024353-6c44effa-4f0a-4f3c-ad34-7504bab9f4ae.png"> https://docs.databricks.com/sql/language-manual/delta-merge-into.html Delata **To Reproduce** Steps to reproduce the behavior: 1. Using merge into command in spark.sql(). 2. Insert into command specifies only subset of the target table columns. **Expected behavior** Partial insert in merge into command and default to null if not specified. **Environment Description** * Hudi version : 0.13.0 * Spark version : 3.3.1 * Hive version : Irrelevant * Hadoop version : Irrelevant * Storage (HDFS/S3/GCS..) : Irrelevant * Running on Docker? (yes/no) : Irrelevant **Additional context** Add any other context about the problem here. **Stacktrace** ```Add the stacktrace of the error.``` -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: [email protected] For queries about this service, please contact Infrastructure at: [email protected]
