shahar1 commented on issue #11323:
URL: https://github.com/apache/airflow/issues/11323#issuecomment-3835916805

   > Hi @shahar1,
   > I have a concern about this implementation.
   > Since the `XCom` has a hard limit to 1GB (due to the DB field type 
`JSONB`) and the return value from a Operator stored in the `XCom`, do we need 
to do something extra? I know the 1GB is a huge size but still... 
   > 
   > Lets assume you are migration a bucket to another bucket (this can be 
trained AI model data) and try to move the data between buckets. I think we can 
hit that limit in this kind of scenario
   
   Hey, thanks for raising these concerns!
   
   
   1. Some operators included in this issue (still) transfer only one file, so 
I wouldn't worry about them.
   
   2. Specificaly for transferring an entire bucket, especially if it includes 
a lot of files, I don't think that that the "regular" GCS2GCS transfer operator 
is right tool to use (using Airflow's resources for transferring), but I'd 
rather use the GCP's storage transfer service instead (deferring to GCP).
   
   3. If you insist to use the GCS2GCS operator, you could always pass 
`do_xcom_push = False` and avoid the xcom being pushed at all :)


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to