This is an automated email from the ASF dual-hosted git repository.
gopidesu pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/airflow.git
The following commit(s) were added to refs/heads/main by this push:
new 40116e30d7e Cleanup unused args example_pyspark.py (#52492)
40116e30d7e is described below
commit 40116e30d7e6db74cf16328ea85940649f261183
Author: GPK <[email protected]>
AuthorDate: Sun Jun 29 20:26:04 2025 +0100
Cleanup unused args example_pyspark.py (#52492)
* Cleanup unused args example_pyspark.py
* Cleanup unused args example_pyspark.py
---
providers/apache/spark/tests/system/apache/spark/example_pyspark.py | 3 +--
1 file changed, 1 insertion(+), 2 deletions(-)
diff --git
a/providers/apache/spark/tests/system/apache/spark/example_pyspark.py
b/providers/apache/spark/tests/system/apache/spark/example_pyspark.py
index d2d8ef6983f..d8ed823d23c 100644
--- a/providers/apache/spark/tests/system/apache/spark/example_pyspark.py
+++ b/providers/apache/spark/tests/system/apache/spark/example_pyspark.py
@@ -23,7 +23,6 @@ import pendulum
if typing.TYPE_CHECKING:
import pandas as pd
- from pyspark import SparkContext
from pyspark.sql import SparkSession
from airflow.decorators import dag, task
@@ -43,7 +42,7 @@ def example_pyspark():
# [START task_pyspark]
@task.pyspark(conn_id="spark-local")
- def spark_task(spark: SparkSession, sc: SparkContext) -> pd.DataFrame:
+ def spark_task(spark: SparkSession) -> pd.DataFrame:
df = spark.createDataFrame(
[
(1, "John Doe", 21),