Samrat002 commented on code in PR #27187:
URL: https://github.com/apache/flink/pull/27187#discussion_r2822325818


##########
flink-filesystems/flink-s3-fs-native/src/main/java/org/apache/flink/fs/s3native/NativeS3OutputStream.java:
##########
@@ -0,0 +1,149 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.fs.s3native;
+
+import org.apache.flink.core.fs.FSDataOutputStream;
+
+import software.amazon.awssdk.core.sync.RequestBody;
+import software.amazon.awssdk.services.s3.S3Client;
+import software.amazon.awssdk.services.s3.model.PutObjectRequest;
+
+import javax.annotation.Nullable;
+
+import java.io.File;
+import java.io.FileOutputStream;
+import java.io.IOException;
+import java.nio.file.Files;
+import java.util.UUID;
+
+public class NativeS3OutputStream extends FSDataOutputStream {
+
+    private final S3Client s3Client;
+    private final String bucketName;
+    private final String key;
+    private final File tmpFile;
+    private final FileOutputStream localStream;
+    private final S3EncryptionConfig encryptionConfig;
+
+    private long position;
+    private volatile boolean closed;
+
+    public NativeS3OutputStream(
+            S3Client s3Client, String bucketName, String key, String 
localTmpDir)
+            throws IOException {
+        this(s3Client, bucketName, key, localTmpDir, null);
+    }
+
+    public NativeS3OutputStream(
+            S3Client s3Client,
+            String bucketName,
+            String key,
+            String localTmpDir,
+            @Nullable S3EncryptionConfig encryptionConfig)
+            throws IOException {
+        this.s3Client = s3Client;
+        this.bucketName = bucketName;
+        this.key = key;
+        this.encryptionConfig =
+                encryptionConfig != null ? encryptionConfig : 
S3EncryptionConfig.none();
+
+        File tmpDir = new File(localTmpDir);
+        if (!tmpDir.exists()) {
+            tmpDir.mkdirs();
+        }
+
+        this.tmpFile = new File(tmpDir, "s3-upload-" + UUID.randomUUID());
+        this.localStream = new FileOutputStream(tmpFile);
+        this.position = 0;
+        this.closed = false;
+    }
+
+    @Override
+    public long getPos() throws IOException {
+        return position;
+    }
+
+    @Override
+    public void write(int b) throws IOException {
+        if (closed) {
+            throw new IOException("Stream is closed");
+        }
+        localStream.write(b);
+        position++;
+    }
+
+    @Override
+    public void write(byte[] b, int off, int len) throws IOException {
+        if (closed) {
+            throw new IOException("Stream is closed");
+        }
+        if (b == null) {
+            throw new NullPointerException();
+        }
+        if (off < 0 || len < 0 || len > b.length - off) {
+            throw new IndexOutOfBoundsException();
+        }
+
+        localStream.write(b, off, len);
+        position += len;
+    }
+
+    @Override
+    public void flush() throws IOException {
+        if (!closed) {
+            localStream.flush();
+        }
+    }
+
+    @Override
+    public void sync() throws IOException {
+        flush();
+    }
+
+    @Override
+    public void close() throws IOException {
+        if (!closed) {
+            closed = true;
+
+            try {
+                localStream.close();
+
+                PutObjectRequest.Builder putRequestBuilder =
+                        PutObjectRequest.builder().bucket(bucketName).key(key);
+
+                // Apply encryption settings
+                if (encryptionConfig.isEnabled()) {
+                    putRequestBuilder.serverSideEncryption(
+                            encryptionConfig.getServerSideEncryption());
+                    if (encryptionConfig.getEncryptionType()
+                                    == 
S3EncryptionConfig.EncryptionType.SSE_KMS
+                            && encryptionConfig.getKmsKeyId() != null) {
+                        
putRequestBuilder.ssekmsKeyId(encryptionConfig.getKmsKeyId());
+                    }
+                }
+
+                s3Client.putObject(putRequestBuilder.build(), 
RequestBody.fromFile(tmpFile));

Review Comment:
   Valid concern. PutObject has a hard 5GB ceiling, and callers like 
FsCheckpointStreamFactory can write state files that approach or exceed that. 
   
   The simplest fix here is to switch to `S3TransferManager.uploadFile()` which 
automatically uses multipart upload for large files. The transfer manager is 
already created in `S3ClientProvider`. 
   
   I can refactor NativeS3OutputStream to accept either the transfer manager or 
the S3ClientProvider instead of a raw S3Client, and use the transfer manager 
for the upload in close(). This removes the 5GB limit without adding manual MPU 
orchestration.
   
   For this pr, I would prefer to go via the current code as 5GB is a fairly 
large number. i will pick it up definitely as a follow-up. 
   
   WDUT ?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to