bertrand-buffat opened a new issue, #26394:
URL: https://github.com/apache/airflow/issues/26394

   ### Apache Airflow version
   
   2.3.4
   
   ### What happened
   
   Tasks are no longer scheduled since upgrading to airflow.2.3.4, with logs:
   ```
   {base_executor.py:211} INFO - task TaskInstanceKey(dag_id='sys_liveness', 
task_id='liveness', run_id='scheduled__2022-09-14T12:15:00+00:00', 
try_number=1, map_index=-1) is still running
   {base_executor.py:215} ERROR - could not queue task 
TaskInstanceKey(dag_id='sys_liveness', task_id='liveness', 
run_id='scheduled__2022-09-14T12:15:00+00:00', try_number=1, map_index=-1) 
(still running after 4 attempts)
   ```
   Even with only one dag running in a simple deployment, and enough slot in 
the default pool.
   
   ### What you think should happen instead
   
   Tasks should be executed.
   
   ### How to reproduce
   
   _No response_
   
   ### Operating System
   
   Debian GNU/Linux
   
   ### Versions of Apache Airflow Providers
   
   ```
   apache-airflow-providers-amazon = "4.1.0"
   apache-airflow-providers-cncf-kubernetes = "4.0.2"
   apache-airflow-providers-http = "2.1.2"
   apache-airflow-providers-mysql = "2.2.3"
   apache-airflow-providers-postgres = "4.1.0"
   apache-airflow-providers-ssh = "2.4.4"
   apache-airflow-providers-sqlite = "2.1.3"
   ```
   
   ### Deployment
   
   Official Apache Airflow Helm Chart
   
   ### Deployment details
   
   `kubernetes executor`
   ```
   [core]
   dags_folder = /data-foundation-airflow/
   hostname_callable = socket.getfqdn
   default_timezone = utc
   executor = KubernetesExecutor
   parallelism = 512
   max_active_tasks_per_dag = 128
   dags_are_paused_at_creation = False
   max_active_runs_per_dag = 1
   load_examples = False
   plugins_folder = /home/airflow/plugins
   execute_tasks_new_python_interpreter = False
   fernet_key = 46BKJoQYlPPOexq0OhDZnIlNepKFf87WFwLbfzqDDho=
   
   donot_pickle = False
   dagbag_import_timeout = 30.0
   dagbag_import_error_tracebacks = True
   dagbag_import_error_traceback_depth = 2
   dag_file_processor_timeout = 300
   task_runner = StandardTaskRunner
   default_impersonation =
   security =
   unit_test_mode = False
   enable_xcom_pickling = False
   killed_task_cleanup_time = 60
   dag_run_conf_overrides_params = True
   dag_discovery_safe_mode = True
   dag_ignore_file_syntax = regexp
   default_task_retries = 0
   default_task_weight_rule = downstream
   default_task_execution_timeout =
   min_serialized_dag_update_interval = 30
   compress_serialized_dags = False
   min_serialized_dag_fetch_interval = 30
   max_num_rendered_ti_fields_per_task = 30
   check_slas = True
   xcom_backend = airflow.models.xcom.BaseXCom
   lazy_load_plugins = True
   lazy_discover_providers = True
   hide_sensitive_var_conn_fields = True
   sensitive_var_conn_names =
   default_pool_task_slot_count = 128
   max_map_length = 1024
   daemon_umask = 0o077
   secure_mode = False
   dag_concurrency = 256
   non_pooled_task_slot_count = 512
   sql_alchemy_conn = 
postgresql+psycopg2://airflow:lc9c2w74uhyqi...@airflow-postgres-instance-1-cluster.ctyrh80ang3y.eu-central-1.rds.amazonaws.com:5432/airflow?options=-csearch_path%3Dairflow
   
   sql_alchemy_schema = airflow
   
   [logging]
   base_log_folder = /home/airflow/logs
   remote_logging = True
   remote_log_conn_id =
   google_key_path =
   remote_base_log_folder = 
s3://doctolib-data-foundation-staging/airflow/staging/logs/
   encrypt_s3_logs = True
   logging_level = INFO
   celery_logging_level =
   fab_logging_level = WARNING
   logging_config_class = log_config.LOGGING_CONFIG
   colored_console_log = True
   colored_log_format = [%(blue)s%(asctime)s%(reset)s] 
{%(blue)s%(filename)s:%(reset)s%(lineno)d} %(log_color)s%(levelname)s%(reset)s 
- %(log_color)s%(message)s%(reset)s
   colored_formatter_class = 
airflow.utils.log.colored_log.CustomTTYColoredFormatter
   log_format = [%(asctime)s] {%(filename)s:%(lineno)d} %(levelname)s - 
%(message)s
   simple_log_format = %(asctime)s %(levelname)s - %(message)s
   log_formatter_class = airflow.utils.log.timezone_aware.TimezoneAware
   task_log_prefix_template =
   log_filename_template = dag_id={{ ti.dag_id }}/run_id={{ ti.run_id 
}}/task_id={{ ti.task_id }}/{% if ti.map_index >= 0 %}map_index={{ ti.map_index 
}}/{% endif %}attempt={{ try_number }}.log
   log_processor_filename_template = {{ filename }}.log
   dag_processor_manager_log_location = 
/home/airflow/logs/dag_processor_manager/dag_processor_manager.log
   task_log_reader = task
   extra_logger_names =
   worker_log_server_port = 8793
   
   [api]
   enable_experimental_api = False
   auth_backends = 
airflow.api.auth.backend.basic_auth,airflow.api.auth.backend.session
   maximum_page_limit = 100
   fallback_page_limit = 100
   google_oauth2_audience =
   google_key_path =
   access_control_allow_headers =
   access_control_allow_methods =
   access_control_allow_origins =
   auth_backend = airflow.api.auth.backend.basic_auth
   
   [webserver]
   base_url = https://airflow-staging.doctolibdata.com
   default_ui_timezone = UTC
   web_server_host = 0.0.0.0
   web_server_port = 8080
   web_server_ssl_cert = /home/airflow/ssl/cert.pem
   web_server_ssl_key = /home/airflow/ssl/key.pem
   session_backend = database
   web_server_master_timeout = 120
   web_server_worker_timeout = 120
   worker_refresh_batch_size = 1
   worker_refresh_interval = 30
   reload_on_plugin_change = False
   secret_key = kCjfcp+BFuDIGas1zi3NkQ==
   workers = 4
   worker_class = sync
   access_logfile = -
   error_logfile = -
   access_logformat =
   expose_config = True
   expose_hostname = True
   expose_stacktrace = True
   dag_default_view = grid
   dag_orientation = LR
   log_fetch_timeout_sec = 5
   log_fetch_delay_sec = 2
   log_auto_tailing_offset = 30
   log_animation_speed = 1000
   hide_paused_dags_by_default = False
   page_size = 100
   navbar_color = #F1C40F
   default_dag_run_display_number = 25
   enable_proxy_fix = False
   proxy_fix_x_for = 1
   proxy_fix_x_proto = 1
   proxy_fix_x_host = 1
   proxy_fix_x_port = 1
   proxy_fix_x_prefix = 1
   cookie_secure = False
   cookie_samesite = Lax
   default_wrap = False
   x_frame_enabled = True
   show_recent_stats_for_completed_runs = True
   update_fab_perms = True
   session_lifetime_minutes = 43200
   instance_name_has_markup = True
   auto_refresh_interval = 3
   warn_deployment_exposure = True
   audit_view_excluded_events = 
gantt,landing_times,tries,duration,calendar,graph,grid,tree,tree_data
   filter_by_owner = True
   authenticate = True
   auth_backend = airflow.contrib.auth.backends.google_auth
   rbac = True
   
   [smtp]
   smtp_host = smtp.gmail.com
   smtp_starttls = True
   smtp_ssl = False
   smtp_port = 587
   smtp_mail_from = [email protected]
   smtp_timeout = 30
   smtp_retry_limit = 5
   smtp_user = [email protected]
   smtp_password = svjfxmprcyxsytdu
   
   
   [scheduler]
   job_heartbeat_sec = 5
   scheduler_heartbeat_sec = 5
   num_runs = -1
   scheduler_idle_sleep_time = 1
   min_file_process_interval = 400
   deactivate_stale_dags_interval = 60
   dag_dir_list_interval = 300
   print_stats_interval = 30
   pool_metrics_interval = 5.0
   scheduler_health_check_threshold = 30
   orphaned_tasks_check_interval = 300.0
   child_process_log_directory = /home/airflow/logs/scheduler
   scheduler_zombie_task_threshold = 300
   zombie_detection_interval = 10.0
   catchup_by_default = True
   ignore_first_depends_on_past_by_default = True
   max_tis_per_query = 512
   use_row_level_locking = True
   max_dagruns_to_create_per_loop = 10
   max_dagruns_per_loop_to_schedule = 20
   schedule_after_task_execution = True
   parsing_processes = 4
   file_parsing_sort_mode = modified_time
   standalone_dag_processor = False
   max_callbacks_per_loop = 20
   use_job_schedule = True
   allow_trigger_in_future = False
   dependency_detector = 
airflow.serialization.serialized_objects.DependencyDetector
   trigger_timeout_check_interval = 15
   max_threads = 5
   ```
   
   ### Anything else
   
   When we restart the scheduler, any queued tasks are executed. But the next 
task after isn't and is stucked in queued again.
   
   ### Are you willing to submit PR?
   
   - [ ] Yes I am willing to submit a PR!
   
   ### Code of Conduct
   
   - [X] I agree to follow this project's [Code of 
Conduct](https://github.com/apache/airflow/blob/main/CODE_OF_CONDUCT.md)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to