stevenzwu edited a comment on issue #4190:
URL: https://github.com/apache/iceberg/issues/4190#issuecomment-1079336184


   I am not sure this would the universally desired behavior. if data stream 
contains incompatible schema change (like removing a required field), it will 
break the downstream consumers.
   
   there is a value of automatically syncing input data schema to Iceberg table 
schema (for compatible schema evolution). Personally, I would like to keep it 
at the control plane, which would be more natural if there is a schema registry 
for tracking input data schema change. Control plane can then update Iceberg 
table schema and restart the Flink job to pick up new Iceberg table schema for 
write path.
   
   It is tricky to support in automatic schema sync in the data plane. There 
would be parallel Iceberg writers (like hundreds) for a single sink table. 
Coordinating metadata (like schema) change is very tricky. 


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]



---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to