Fred Teunissen created FLINK-37749: -------------------------------------- Summary: FileSystem Connector fails writing to S3 bucket Key: FLINK-37749 URL: https://issues.apache.org/jira/browse/FLINK-37749 Project: Flink Issue Type: Bug Components: Connectors / FileSystem Affects Versions: 1.20.1, 1.19.2, 1.20.0 Environment: Docker running: - Flink 1.20.1 - minio Reporter: Fred Teunissen Attachments: filesystem-connector-s3-issue.tgz
Writing to an output file on S3 using the filesystem connector results in a crash. The output file is defined as follows. {code:java} CREATE TABLE <tableName>> ( bool BOOLEAN, `someInt` INT ) PARTITIONED BY (someInt) WITH ( 'connector' = 'filesystem', 'path' = 's3a://tdata', 'format' = 'csv', -- 'format' = 'parquet', 'partition.default-name' = 'null' );{code} Note that the 'path' refers to an existing bucket that is hosted using Minio. The job starts OK but as soon as the first checkpoint is being triggered the job crashes with the following exception: {code:java} "Caused by: org.apache.flink.util.SerializedThrowable: java.io.IOException: Stream closed." (full details in the attached zip file).{code} The issue can be reproduced by unzipping the attached zip file that contains a docker-compose file to spin up the Flink cluster and Minio container and using the instructions in the README.MD to generate test data resulting in the job crash. Cause is unknown but I suspect more (or other) .jars need to be added to the plugins/s3 directory as documented here: https://nightlies.apache.org/flink/flink-docs-master/docs/deployment/filesystems/plugins/ -- This message was sent by Atlassian Jira (v8.20.10#820010)