[
https://issues.apache.org/jira/browse/SPARK-35299?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Yusheng Ding updated SPARK-35299:
---------------------------------
Description:
To reproduce:
test_table path: s3a://test_bucket/test_table/
df = spark_session.sql("SELECT * FROM test_table")
df.count() # produce row number 1000
#####S3 operation######
s3 = boto3.client("s3")
s3.put_object(
Bucket="test_bucket", Body="", Key=f"test_table/"
)
#####S3 operation######
df.write.insertInto(test_table, overwrite=True)
# Same goes to df.write.save(mode="overwrite", format="parquet",
path="s3a://test_bucket/test_table")
df = spark_session.sql("SELECT * FROM test_table")
df.count() # produce row number 2000
Overwrite is not functioning correctly. Old files will not be deleted on S3.
was:
To reproduce:
test_table path: s3a://test_bucket/test_table/
df = spark_session.sql("SELECT * FROM test_table")
df.count() # produce row number 1000
#####S3 operation######
s3 = boto3.client("s3")
s3.put_object(
Bucket="test_bucket", Body="", Key=f"test_table/"
)
#####S3 operation######
df.write.insertInto(test_table, overwrite=True)
# Same goes to df.write.save(mode="overwrite", format="parquet",
path="s3a://test_bucket/test_table")
df = spark_session.sql("SELECT * FROM test_table")
df.count() # produce row number 2000
Overwrite is not functioning correctly. Old files will not be deleted on S3.
> Dataframe overwrite on S3 does not delete old files with S3 object-put to
> table path
> ------------------------------------------------------------------------------------
>
> Key: SPARK-35299
> URL: https://issues.apache.org/jira/browse/SPARK-35299
> Project: Spark
> Issue Type: Bug
> Components: Spark Core
> Affects Versions: 2.2.0
> Reporter: Yusheng Ding
> Priority: Major
> Labels: aws-s3, dataframe, hive, spark
>
> To reproduce:
> test_table path: s3a://test_bucket/test_table/
>
> df = spark_session.sql("SELECT * FROM test_table")
> df.count() # produce row number 1000
> #####S3 operation######
> s3 = boto3.client("s3")
> s3.put_object(
> Bucket="test_bucket", Body="", Key=f"test_table/"
> )
> #####S3 operation######
> df.write.insertInto(test_table, overwrite=True)
> # Same goes to df.write.save(mode="overwrite", format="parquet",
> path="s3a://test_bucket/test_table")
> df = spark_session.sql("SELECT * FROM test_table")
> df.count() # produce row number 2000
>
> Overwrite is not functioning correctly. Old files will not be deleted on S3.
>
>
>
--
This message was sent by Atlassian Jira
(v8.3.4#803005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]