Samrat002 commented on code in PR #27187:
URL: https://github.com/apache/flink/pull/27187#discussion_r2877941290


##########
flink-filesystems/flink-s3-fs-native/src/main/java/org/apache/flink/fs/s3native/writer/NativeS3RecoverableFsDataOutputStream.java:
##########
@@ -0,0 +1,273 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.fs.s3native.writer;
+
+import org.apache.flink.core.fs.RecoverableFsDataOutputStream;
+import org.apache.flink.core.fs.RecoverableWriter;
+import org.apache.flink.fs.s3native.writer.NativeS3Recoverable.PartETag;
+
+import javax.annotation.concurrent.NotThreadSafe;
+
+import java.io.File;
+import java.io.FileOutputStream;
+import java.io.IOException;
+import java.nio.file.Files;
+import java.util.ArrayList;
+import java.util.List;
+import java.util.UUID;
+import java.util.concurrent.locks.ReentrantLock;
+
+/**
+ * A recoverable output stream writes data to S3 using multipart uploads.
+ *
+ * <p>This class is NOT thread-safe. All write operations ({@link #write}, 
{@link #flush}, {@link
+ * #persist}, {@link #closeForCommit}) must be called from a single thread 
(the Flink operator
+ * thread). This is consistent with Flink's {@link 
RecoverableFsDataOutputStream} contract where
+ * output streams are confined to a single task thread.
+ *
+ * <p>The {@link #close()} method may be called concurrently (e.g., during 
task cancellation). A
+ * {@link ReentrantLock} guards the critical sections in {@link #close()}, 
{@link
+ * #closeForCommit()}, and {@link #persist()} to ensure safe cleanup of local 
resources without
+ * corrupting S3 state.
+ */
+@NotThreadSafe
+public class NativeS3RecoverableFsDataOutputStream extends 
RecoverableFsDataOutputStream {
+
+    /** Lock to guard close/persist/commit paths against concurrent close() 
during cancellation. */
+    private final ReentrantLock lock = new ReentrantLock();
+
+    private final NativeS3AccessHelper s3AccessHelper;
+    private final String key;
+    private final String uploadId;
+    private final String localTmpDir;
+    private final long minPartSize;
+
+    private final List<PartETag> completedParts;
+    private long numBytesInParts;
+
+    private File currentTempFile;
+    private FileOutputStream currentOutputStream;
+    private long currentPartSize;
+    private int nextPartNumber;
+
+    private volatile boolean closed;
+
+    public NativeS3RecoverableFsDataOutputStream(
+            NativeS3AccessHelper s3AccessHelper,
+            String key,
+            String uploadId,
+            String localTmpDir,
+            long minPartSize)
+            throws IOException {
+        this(s3AccessHelper, key, uploadId, localTmpDir, minPartSize, new 
ArrayList<>(), 0L);
+    }
+
+    public NativeS3RecoverableFsDataOutputStream(
+            NativeS3AccessHelper s3AccessHelper,
+            String key,
+            String uploadId,
+            String localTmpDir,
+            long minPartSize,
+            List<PartETag> existingParts,
+            long numBytesInParts)
+            throws IOException {
+        this.s3AccessHelper = s3AccessHelper;
+        this.key = key;
+        this.uploadId = uploadId;
+        this.localTmpDir = localTmpDir;
+        this.minPartSize = minPartSize;
+        this.completedParts = new ArrayList<>(existingParts);
+        this.numBytesInParts = numBytesInParts;
+        this.nextPartNumber = existingParts.size() + 1;
+        this.currentPartSize = 0;
+        this.closed = false;
+
+        createNewTempFile();
+    }
+
+    private void createNewTempFile() throws IOException {
+        File tmpDir = new File(localTmpDir);
+        if (!tmpDir.exists()) {
+            tmpDir.mkdirs();
+        }
+
+        currentTempFile = new File(tmpDir, "s3-part-" + UUID.randomUUID());
+        currentOutputStream = new FileOutputStream(currentTempFile);
+        currentPartSize = 0;
+    }
+
+    @Override
+    public long getPos() throws IOException {
+        return numBytesInParts + currentPartSize;
+    }
+
+    @Override
+    public void write(int b) throws IOException {
+        if (closed) {
+            throw new IOException("Stream is closed");
+        }
+
+        currentOutputStream.write(b);
+        currentPartSize++;
+
+        if (currentPartSize >= minPartSize) {
+            uploadCurrentPart();
+            createNewTempFile();
+        }
+    }
+
+    @Override
+    public void write(byte[] b, int off, int len) throws IOException {
+        if (closed) {
+            throw new IOException("Stream is closed");
+        }
+        if (b == null) {
+            throw new NullPointerException();
+        }
+        if (off < 0 || len < 0 || len > b.length - off) {
+            throw new IndexOutOfBoundsException();
+        }
+
+        currentOutputStream.write(b, off, len);
+        currentPartSize += len;
+
+        if (currentPartSize >= minPartSize) {
+            uploadCurrentPart();
+            createNewTempFile();
+        }
+    }
+
+    @Override
+    public void flush() throws IOException {
+        if (!closed) {
+            currentOutputStream.flush();
+        }
+    }
+
+    @Override
+    public void sync() throws IOException {
+        flush();
+    }
+
+    private void uploadCurrentPart() throws IOException {
+        currentOutputStream.close();
+
+        int partNumber = nextPartNumber++;
+        NativeS3AccessHelper.UploadPartResult result =
+                s3AccessHelper.uploadPart(
+                        key, uploadId, partNumber, currentTempFile, 
currentPartSize);
+
+        completedParts.add(new PartETag(result.getPartNumber(), 
result.getETag()));
+        numBytesInParts += currentPartSize;
+
+        Files.delete(currentTempFile.toPath());
+    }
+
+    @Override
+    public Committer closeForCommit() throws IOException {
+        lock();
+        try {
+            if (closed) {
+                throw new IOException("Stream is already closed");
+            }
+
+            closed = true;
+            currentOutputStream.close();
+
+            if (currentPartSize > 0) {
+                uploadCurrentPart();
+            } else {
+                Files.delete(currentTempFile.toPath());
+            }
+
+            NativeS3Recoverable recoverable =
+                    new NativeS3Recoverable(
+                            key, uploadId, new ArrayList<>(completedParts), 
numBytesInParts);
+
+            return new NativeS3Committer(s3AccessHelper, recoverable);
+        } finally {
+            unlock();
+        }
+    }
+
+    @Override
+    public RecoverableWriter.ResumeRecoverable persist() throws IOException {
+        lock();
+        try {
+            flush();
+
+            String incompletePartKey = null;
+            long incompletePartLength = 0;
+
+            if (currentPartSize > 0) {
+                currentOutputStream.flush();
+                incompletePartKey = key + "/.incomplete/" + uploadId + "/" + 
UUID.randomUUID();
+                s3AccessHelper.putObject(incompletePartKey, currentTempFile);
+                incompletePartLength = currentPartSize;
+            }
+
+            return new NativeS3Recoverable(
+                    key,
+                    uploadId,
+                    new ArrayList<>(completedParts),
+                    numBytesInParts,
+                    incompletePartKey,
+                    incompletePartLength);
+        } finally {
+            unlock();
+        }
+    }
+
+    @Override
+    public void close() throws IOException {
+        lock();
+        try {
+            if (!closed) {
+                closed = true;
+                if (currentOutputStream != null) {
+                    currentOutputStream.close();
+                }
+                if (currentTempFile != null && currentTempFile.exists()) {
+                    Files.delete(currentTempFile.toPath());
+                }
+
+                try {
+                    s3AccessHelper.abortMultiPartUpload(key, uploadId);
+                } catch (IOException e) {
+                    // best-effort cleanup
+                }

Review Comment:
   
   > Can you elaborate on that? What "S3 lifecycle rules" do you mean?
   
   S3's AbortIncompleteMultipartUpload lifecycle rule, configured on the 
bucket, automatically deletes incomplete multipart uploads after a configurable 
number of days. This is the standard S3 mechanism for handling abandoned 
uploads. AWS recommend enabling it on all buckets as a best practice.
   
   
https://docs.aws.amazon.com/AmazonS3/latest/userguide/mpu-abort-incomplete-mpu-lifecycle-config.html
   
   
   
   > Ok. In that case I guess at worst we are discussing a potential resource 
leak.
   
   Yes, in the worst case this is a potential resource leak of orphaned 
multipart upload parts in S3, not a data correctness issue.
   
   I think not propagating the exception should not be an issue because of the 
following reason: 
   1. This is the disposal path (RecoverableFsDataOutputStream.close()), not 
the persistence path. The upload was never committed so no data is at risk.
   2. Propagating would mask the original failure that triggered `close()` in 
the first place. For example, task cancellation.
   3. The pre-existing `S3RecoverableFsDataOutputStream.close()` follows the 
same pattern with IOUtils.closeQuietly().



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to