This is an automated email from the ASF dual-hosted git repository.

gopidesu pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/airflow.git


The following commit(s) were added to refs/heads/main by this push:
     new 40116e30d7e Cleanup unused args example_pyspark.py (#52492)
40116e30d7e is described below

commit 40116e30d7e6db74cf16328ea85940649f261183
Author: GPK <[email protected]>
AuthorDate: Sun Jun 29 20:26:04 2025 +0100

    Cleanup unused args example_pyspark.py (#52492)
    
    * Cleanup unused args example_pyspark.py
    
    * Cleanup unused args example_pyspark.py
---
 providers/apache/spark/tests/system/apache/spark/example_pyspark.py | 3 +--
 1 file changed, 1 insertion(+), 2 deletions(-)

diff --git 
a/providers/apache/spark/tests/system/apache/spark/example_pyspark.py 
b/providers/apache/spark/tests/system/apache/spark/example_pyspark.py
index d2d8ef6983f..d8ed823d23c 100644
--- a/providers/apache/spark/tests/system/apache/spark/example_pyspark.py
+++ b/providers/apache/spark/tests/system/apache/spark/example_pyspark.py
@@ -23,7 +23,6 @@ import pendulum
 
 if typing.TYPE_CHECKING:
     import pandas as pd
-    from pyspark import SparkContext
     from pyspark.sql import SparkSession
 
 from airflow.decorators import dag, task
@@ -43,7 +42,7 @@ def example_pyspark():
 
     # [START task_pyspark]
     @task.pyspark(conn_id="spark-local")
-    def spark_task(spark: SparkSession, sc: SparkContext) -> pd.DataFrame:
+    def spark_task(spark: SparkSession) -> pd.DataFrame:
         df = spark.createDataFrame(
             [
                 (1, "John Doe", 21),

Reply via email to