potiuk commented on code in PR #40894:
URL: https://github.com/apache/airflow/pull/40894#discussion_r1685506885
##########
airflow/serialization/pydantic/taskinstance.py:
##########
@@ -458,9 +458,9 @@ def schedule_downstream_tasks(self, session: Session | None
= None, max_tis_per_
:meta: private
"""
- return TaskInstance._schedule_downstream_tasks(
- ti=self, session=session, max_tis_per_query=max_tis_per_query
- )
+ # we should not schedule downstream tasks with Pydantic model because
it will not be able to
+ # get the DAG object (we do not serialize it currently).
+ return
Review Comment:
But the problem is that this method is a "db" method and it needs to be
executed in the "internal_api" component - so we would have to effectively pass
the parsed DAG over the RPC call (or parse the DAG in the "internal_api". This
method however intermixes `dag` object and DB access. This is also why I think
mini-scheduler in AIP-72 is going to be very difficult, because (at least
currently) you would essentially have to have both "parsed DAG" and DB in the
same place (with AIP-72 and internal API here - the worker has DAG, but it has
no DB.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]