swamirishi commented on code in PR #9489:
URL: https://github.com/apache/ozone/pull/9489#discussion_r2616326767


##########
hadoop-hdds/rocks-native/src/main/native/ManagedRawSSTFileIterator.cpp:
##########
@@ -16,57 +16,50 @@
  * limitations under the License.
  */
 
-#include "org_apache_hadoop_hdds_utils_db_managed_ManagedRawSSTFileIterator.h"
+#include "org_apache_hadoop_hdds_utils_db_ManagedRawSSTFileIterator.h"
 #include "rocksdb/options.h"
 #include "rocksdb/raw_iterator.h"
 #include <string>
 #include "cplusplus_to_java_convert.h"
 #include <iostream>
 
-jboolean 
Java_org_apache_hadoop_hdds_utils_db_managed_ManagedRawSSTFileIterator_hasNext(JNIEnv
 *env, jobject obj,
+template <class T>
+static jint copyToDirect(JNIEnv* env, T& source, jobject jtarget, jint 
jtarget_off, jint jtarget_len);
+
+jboolean 
Java_org_apache_hadoop_hdds_utils_db_ManagedRawSSTFileIterator_hasNext(JNIEnv 
*env, jobject obj,
                                                                                
            jlong native_handle) {
     return 
static_cast<jboolean>(reinterpret_cast<ROCKSDB_NAMESPACE::RawIterator*>(native_handle)->Valid());
 }
 
-void 
Java_org_apache_hadoop_hdds_utils_db_managed_ManagedRawSSTFileIterator_next(JNIEnv
 *env, jobject obj,
+void 
Java_org_apache_hadoop_hdds_utils_db_ManagedRawSSTFileIterator_next(JNIEnv 
*env, jobject obj,
                                                                                
        jlong native_handle) {
     reinterpret_cast<ROCKSDB_NAMESPACE::RawIterator*>(native_handle)->Next();
 }
 
-jbyteArray 
Java_org_apache_hadoop_hdds_utils_db_managed_ManagedRawSSTFileIterator_getKey(JNIEnv
 *env,
-                                                                               
                jobject obj,
-                                                                               
                jlong native_handle) {
+jint 
Java_org_apache_hadoop_hdds_utils_db_ManagedRawSSTFileIterator_getKey(JNIEnv 
*env,
+                                                                           
jobject obj,
+                                                                           
jlong native_handle,
+                                                                           
jobject jtarget,
+                                                                           
jint jtarget_off, jint jtarget_len) {
     ROCKSDB_NAMESPACE::Slice slice = 
reinterpret_cast<ROCKSDB_NAMESPACE::RawIterator*>(native_handle)->key();
-    jbyteArray jkey = env->NewByteArray(static_cast<jsize>(slice.size()));
-    if (jkey == nullptr) {
-        // exception thrown: OutOfMemoryError
-        return nullptr;
-    }
-    env->SetByteArrayRegion(
-            jkey, 0, static_cast<jsize>(slice.size()),
-            const_cast<jbyte*>(reinterpret_cast<const jbyte*>(slice.data())));
-    return jkey;
+    return copyToDirect(env, slice, jtarget,
+                                                      jtarget_off, 
jtarget_len);
 }
 
 
-jbyteArray 
Java_org_apache_hadoop_hdds_utils_db_managed_ManagedRawSSTFileIterator_getValue(JNIEnv
 *env,
-                                                                               
                jobject obj,
-                                                                               
                jlong native_handle) {
+jint 
Java_org_apache_hadoop_hdds_utils_db_ManagedRawSSTFileIterator_getValue(JNIEnv 
*env,
+                                                                             
jobject obj,
+                                                                             
jlong native_handle,
+                                                                             
jobject jtarget,
+                                                                             
jint jtarget_off, jint jtarget_len) {
     ROCKSDB_NAMESPACE::Slice slice = 
reinterpret_cast<ROCKSDB_NAMESPACE::RawIterator*>(native_handle)->value();
     jbyteArray jkey = env->NewByteArray(static_cast<jsize>(slice.size()));

Review Comment:
   done



##########
hadoop-hdds/rocks-native/src/main/native/ManagedRawSSTFileIterator.cpp:
##########
@@ -75,16 +68,38 @@ jlong 
Java_org_apache_hadoop_hdds_utils_db_managed_ManagedRawSSTFileIterator_get
 }
 
 
-jint 
Java_org_apache_hadoop_hdds_utils_db_managed_ManagedRawSSTFileIterator_getType(JNIEnv
 *env,
+jint 
Java_org_apache_hadoop_hdds_utils_db_ManagedRawSSTFileIterator_getType(JNIEnv 
*env,
                                                                                
           jobject obj,
                                                                                
           jlong native_handle) {
     uint32_t type = 
reinterpret_cast<ROCKSDB_NAMESPACE::RawIterator*>(native_handle)->type();
     return static_cast<jint>(type);
 }
 
 
-void 
Java_org_apache_hadoop_hdds_utils_db_managed_ManagedRawSSTFileIterator_closeInternal(JNIEnv
 *env,
+void 
Java_org_apache_hadoop_hdds_utils_db_ManagedRawSSTFileIterator_closeInternal(JNIEnv
 *env,
                                                                                
                 jobject obj,
                                                                                
                 jlong native_handle) {
     delete reinterpret_cast<ROCKSDB_NAMESPACE::RawIterator*>(native_handle);
 }
+
+template <class T>
+static jint copyToDirect(JNIEnv* env, T& source, jobject jtarget,
+                         jint jtarget_off, jint jtarget_len) {
+  char* target = reinterpret_cast<char*>(env->GetDirectBufferAddress(jtarget));
+  if (target == nullptr || env->GetDirectBufferCapacity(jtarget) < 
(jtarget_off + jtarget_len)) {
+    jclass exClass = env->FindClass("java/lang/IllegalArgumentException");
+    if (exClass == nullptr) {

Review Comment:
   done



##########
hadoop-hdds/rocksdb-checkpoint-differ/src/main/java/org/apache/hadoop/hdds/utils/db/ManagedSstFileIterator.java:
##########
@@ -0,0 +1,90 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements. See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License. You may obtain a copy of the License at
+ *
+ *      http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.hadoop.hdds.utils.db;
+
+import org.apache.hadoop.hdds.utils.db.managed.ManagedOptions;
+import org.apache.hadoop.hdds.utils.db.managed.ManagedReadOptions;
+import org.apache.hadoop.hdds.utils.db.managed.ManagedSstFileReader;
+import org.apache.hadoop.hdds.utils.db.managed.ManagedSstFileReaderIterator;
+import org.apache.hadoop.ozone.util.ClosableIterator;
+import org.rocksdb.RocksDBException;
+
+/**
+ * ManagedSstFileIterator is an abstract class designed to provide a managed, 
resource-safe
+ * iteration over SST (Sorted String Table) files leveraging RocksDB. It 
implements the
+ * {@link ClosableIterator} interface to support resource management and 
ensures proper
+ * cleanup of resources upon closure. This class binds together a 
ManagedSstFileReader,
+ * ManagedSstFileReaderIterator, and Buffers for keys and values, while 
allowing specific
+ * implementations to define how the iterator values are derived.
+ *
+ * @param <T> The type of the element to be returned by the iterator.
+ */
+abstract class ManagedSstFileIterator<T> implements ClosableIterator<T> {
+  private final ManagedSstFileReader fileReader;
+  private final ManagedSstFileReaderIterator fileReaderIterator;
+  private final IteratorType type;
+  private boolean closed;
+  private final Buffer keyBuffer;
+  private final Buffer valueBuffer;
+
+  ManagedSstFileIterator(String path, ManagedOptions options, 
ManagedReadOptions readOptions,
+      IteratorType type) throws RocksDatabaseException {
+    try {
+      this.fileReader = new ManagedSstFileReader(options);
+      this.fileReader.open(path);
+      this.fileReaderIterator = 
ManagedSstFileReaderIterator.managed(fileReader.newIterator(readOptions));
+      fileReaderIterator.get().seekToFirst();
+      this.closed = false;
+      this.type = type;
+      this.keyBuffer = new Buffer(
+          new CodecBuffer.Capacity(path + " iterator-key", 1 << 10),
+          this.type.readKey() ? buffer -> fileReaderIterator.get().key(buffer) 
: null);
+      this.valueBuffer = new Buffer(
+          new CodecBuffer.Capacity(path + " iterator-value", 4 << 10),
+          this.type.readValue() ? buffer -> 
fileReaderIterator.get().value(buffer) : null);
+    } catch (RocksDBException e) {
+      throw new RocksDatabaseException("Failed to open SST file: " + path, e);
+    }
+  }
+
+  @Override
+  public synchronized void close() {
+    if (!closed) {
+      this.fileReaderIterator.close();
+      this.fileReader.close();
+      keyBuffer.release();
+      valueBuffer.release();
+    }
+    closed = true;
+  }
+
+  @Override
+  public boolean hasNext() {
+    return fileReaderIterator.get().isValid();
+  }
+
+  protected abstract T getIteratorValue(CodecBuffer key, CodecBuffer value);
+
+  @Override
+  public T next() {

Review Comment:
   done



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to