amoghrajesh commented on PR #51059:
URL: https://github.com/apache/airflow/pull/51059#issuecomment-2979011497

   > there was a discussion on whether it's good that we are depending on other 
dependencies (say pandas, deltalake, iceberg, kubernetes) which are part of the 
"core" airflow - but also we have "providers" for many of those that provide 
operators / hooks and other "core extensions" related to the respective 
"external entity".
   
   In my view, serializers for kubernetes, should come from kubernetes 
provider. Deltalake -> should come from databricks (or deltalake provider if we 
decide to have one), iceberg should come from iceberg provider.
   
   
   
   Good pointers, i absolutely agree that the kubernetes, pandas, deltalake and 
iceberg should not belong in core and should be safely moved to providers, i 
would love to hear what @bolkedebruin thinks on this. Is it the versioning that 
is stopping us from doing that? 
   
   
   > For me the Pydantic thing comes as very similar thing - it should be 
"extension" that should be IMHO implemented outside of core. So maybe it's the 
right time to do this kind of refactoring -> Implement "discovery" mechanism in 
providers manager to discover which serializers are installed (similarly as all 
other extensions) - and similarly speciic "pydantic" model could be provided as 
"extension" - by a provider or manually.
   
   However, i think pydantic is more of a core thing, it not essentially 
belongs to a provider, it can be consumed and used in core without additional 
dependencies. So there's no stopping anybody from returning a pydantic 
dataclass object as xcom.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to