truonglac2603 opened a new issue, #39184:
URL: https://github.com/apache/airflow/issues/39184

   ### Apache Airflow version
   
   Other Airflow 2 version (please specify below)
   
   ### If "Other Airflow 2 version" selected, which one?
   
   2.8.3
   
   ### What happened?
   
   I'm having an issue when using SparkKubernetesOperator to execute Spark job. 
Whenever i set do_xcom_push to True, the driver pod is created and run 
perfectly, but the xcom sidecar pod is nowhere to be found. Airflow log is thus 
stuck with this message:  
   
   > [2024-04-23, 02:47:38 UTC] {custom_object_launcher.py:301} WARNING - Spark 
job submitted but not yet started. job_id: spark-custome-task-5r3pt2vk
   > [2024-04-23, 02:47:48 UTC] {pod_manager.py:529} ERROR - container base 
whose logs were requested not found in the pod 
spark-custome-task-5r3pt2vk-driver
   > [2024-04-23, 02:47:48 UTC] {pod_manager.py:718} INFO - Checking if xcom 
sidecar container is started.
   > [2024-04-23, 02:47:48 UTC] {pod_manager.py:724} WARNING - The xcom sidecar 
container is not yet started.
   > [2024-04-23, 02:51:24 UTC] {local_task_job_runner.py:296} WARNING - DagRun 
timed out after 0:05:03.242442.
   > [2024-04-23, 02:51:29 UTC] {local_task_job_runner.py:296} WARNING - DagRun 
timed out after 0:05:08.319418.
   > [2024-04-23, 02:51:29 UTC] {local_task_job_runner.py:302} WARNING - State 
of this instance has been externally set to skipped. Terminating instance.
   > [2024-04-23, 02:51:29 UTC] {process_utils.py:131} INFO - Sending 15 to 
group 30. PIDs of all processes in the group: [30]
   > [2024-04-23, 02:51:29 UTC] {process_utils.py:86} INFO - Sending the signal 
15 to group 30
   > [2024-04-23, 02:51:29 UTC] {taskinstance.py:2483} ERROR - Received 
SIGTERM. Terminating subprocesses.
   > [2024-04-23, 02:51:29 UTC] {process_utils.py:79} INFO - Process 
psutil.Process(pid=30, status='terminated', exitcode=0, started='02:46:43') 
(30) terminated with exit code 0
   
   In my opinion, there should be something wrong with xcom push for this 
operator. Any help would be much appreciated. Thanks in advance.
   
   ### What you think should happen instead?
   
   xcom pushed perfectly with any case of spark, maybe there will be 
placeholder for spark job with no return value or whatsoever
   
   ### How to reproduce
   
   submit a basic spark application by using SparkKubernetesOperator to K8s 
cluster and set do_xcom_push to True
   
   ### Operating System
   
   Ubuntu 22.04.4 LTS
   
   ### Versions of Apache Airflow Providers
   
   apache-airflow-providers-cncf-kubernetes==8.0.1
   
   ### Deployment
   
   Official Apache Airflow Helm Chart
   
   ### Deployment details
   
   _No response_
   
   ### Anything else?
   
   _No response_
   
   ### Are you willing to submit PR?
   
   - [ ] Yes I am willing to submit a PR!
   
   ### Code of Conduct
   
   - [X] I agree to follow this project's [Code of 
Conduct](https://github.com/apache/airflow/blob/main/CODE_OF_CONDUCT.md)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to