ibessonov commented on code in PR #787: URL: https://github.com/apache/ignite-3/pull/787#discussion_r857443657
########## modules/storage-rocksdb/src/main/java/org/apache/ignite/internal/storage/rocksdb/RocksDbMvPartitionStorage.java: ########## @@ -0,0 +1,494 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one or more + * contributor license agreements. See the NOTICE file distributed with + * this work for additional information regarding copyright ownership. + * The ASF licenses this file to You under the Apache License, Version 2.0 + * (the "License"); you may not use this file except in compliance with + * the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.ignite.internal.storage.rocksdb; + +import static java.lang.ThreadLocal.withInitial; +import static java.nio.ByteBuffer.allocateDirect; +import static java.nio.ByteOrder.BIG_ENDIAN; +import static org.apache.ignite.internal.storage.IgniteRowId.MAX_ROW_ID_SIZE; +import static org.apache.ignite.internal.util.ArrayUtils.BYTE_EMPTY_ARRAY; + +import java.nio.ByteBuffer; +import java.util.Arrays; +import java.util.NoSuchElementException; +import java.util.UUID; +import java.util.function.Predicate; +import org.apache.ignite.internal.schema.BinaryRow; +import org.apache.ignite.internal.schema.ByteBufferRow; +import org.apache.ignite.internal.storage.IgniteRowId; +import org.apache.ignite.internal.storage.MvPartitionStorage; +import org.apache.ignite.internal.storage.StorageException; +import org.apache.ignite.internal.storage.TxIdMismatchException; +import org.apache.ignite.internal.tx.Timestamp; +import org.apache.ignite.internal.util.Cursor; +import org.apache.ignite.internal.util.GridUnsafe; +import org.apache.ignite.internal.util.IgniteUtils; +import org.apache.ignite.lang.IgniteInternalException; +import org.jetbrains.annotations.Nullable; +import org.rocksdb.ColumnFamilyHandle; +import org.rocksdb.ReadOptions; +import org.rocksdb.RocksDB; +import org.rocksdb.RocksDBException; +import org.rocksdb.RocksIterator; +import org.rocksdb.Slice; +import org.rocksdb.WriteOptions; + +/** + * Mult-versioned partition storage implementation based on RocksDB. Stored data has the following format: + * <pre><code> + * | partId (2 bytes, BE) | rowId ({@link #igniteRowIdSize} bytes) |</code></pre> + * or + * <pre><code> + * | partId (2 bytes, BE) | rowId ({@link #igniteRowIdSize} bytes) | timestamp (16 bytes, DESC) |</code></pre> + * depending on transaction status. Pending transactions data doesn't have a timestamp assigned. + * + * <p/>BE means Big Endian, meaning that lexicographical bytes order matches a natural order of partitions. + * + * <p/>DESC means that timestamps are sorted from newest to oldest (N2O). Please refer to {@link #putTimestamp(ByteBuffer, Timestamp)} to + * see how it's achieved. Missing timestamp could be interpreted as a moment infinitely far away in the future. + */ +public class RocksDbMvPartitionStorage implements MvPartitionStorage { + /** Position of row id inside of the key. */ + private static final int ROW_ID_OFFSET = Short.BYTES; + + /** Timestamp size in bytes. */ + private static final int TIMESTAMP_SIZE = 2 * Long.BYTES; + + /** Maximum possible size of the key. */ + private static final int MAX_KEY_SIZE = /* partId */ ROW_ID_OFFSET + /* rowId */ MAX_ROW_ID_SIZE + /* timestamp */ TIMESTAMP_SIZE; + + /** Threadlocal direct buffer instance to read keys from RocksDB. */ + private static final ThreadLocal<ByteBuffer> MV_KEY_BUFFER = withInitial(() -> allocateDirect(MAX_KEY_SIZE).order(BIG_ENDIAN)); + + /** Threadlocal on-heap byte buffer instance to use for key manipulations. */ + private final ThreadLocal<ByteBuffer> heapKeyBuffer; + + /** Size of the {@link IgniteRowId} identifiers for this particular partition storage. */ + private final int igniteRowIdSize; + + /** + * Partition ID (should be treated as an unsigned short). + * + * <p/>Partition IDs are always stored in the big endian order, since they need to be compared lexicographically. + */ + private final int partId; + + /** RocksDb instance. */ + private final RocksDB db; + + /** Partitions column family. */ + private final ColumnFamilyHandle cf; + + /** Write options. */ + private final WriteOptions writeOpts = new WriteOptions().setDisableWAL(true); + + /** Upper bound for scans and reads. */ + private final Slice upperBound; + + /** + * Constructor. + * + * @param igniteRowIdSize Size of row id. + * @param partId Partition id. + * @param db RocksDB instance. + * @param cf Column family handle to store partition data. + */ + public RocksDbMvPartitionStorage(int igniteRowIdSize, int partId, RocksDB db, ColumnFamilyHandle cf) { + this.igniteRowIdSize = igniteRowIdSize; + this.partId = partId; + this.db = db; + this.cf = cf; + + heapKeyBuffer = withInitial(() -> + ByteBuffer.allocate(Short.BYTES + igniteRowIdSize + ROW_ID_OFFSET * Long.BYTES) + .order(BIG_ENDIAN) + .putShort((short) partId) + ); + + upperBound = new Slice(partitionEndPrefix()); + } + + /** {@inheritDoc} */ + @Override + public void addWrite(IgniteRowId rowId, @Nullable BinaryRow row, UUID txId) throws TxIdMismatchException, StorageException { + // Prepare a buffer with partition id and row id. + ByteBuffer keyBuf = prepareHeapKeyBuf(rowId); + + try { + // Check concurrent transaction data. + if (RocksDB.NOT_FOUND != db.get(cf, keyBuf.array(), 0, keyBuf.position(), BYTE_EMPTY_ARRAY, 0, 0)) { + throw new TxIdMismatchException(); + } + + if (row == null) { + // Write empty array as a tombstone. + db.put(cf, writeOpts, keyBuf.array(), 0, keyBuf.position(), BYTE_EMPTY_ARRAY, 0, 0); + } else { + byte[] valueBytes = row.bytes(); + + // Write binary row data as a value. + db.put(cf, writeOpts, keyBuf.array(), 0, keyBuf.position(), valueBytes, 0, valueBytes.length); + } + } catch (RocksDBException e) { + throw new StorageException(e); + } + } + + /** {@inheritDoc} */ + @Override + public void abortWrite(IgniteRowId rowId) throws StorageException { + // Prepare a buffer with partition id and row id. + ByteBuffer keyBuf = prepareHeapKeyBuf(rowId); + + try { + // Perform unconditional remove for the key without associated timestamp. + db.delete(cf, writeOpts, keyBuf.array(), 0, keyBuf.position()); + } catch (RocksDBException e) { + throw new StorageException(e); + } + } + + /** {@inheritDoc} */ + @Override + public void commitWrite(IgniteRowId rowId, Timestamp timestamp) throws StorageException { + // Prepare a buffer with partition id and row id. + ByteBuffer keyBuf = prepareHeapKeyBuf(rowId); + + try { + // Read a value associated with pending write. + byte[] valueBytes = db.get(cf, keyBuf.array(), 0, keyBuf.position()); + + // Delete pending write. + db.delete(cf, writeOpts, keyBuf.array(), 0, keyBuf.position()); + + // Add timestamp to the key put the value back into the storage. + putTimestamp(keyBuf, timestamp); + + db.put(cf, writeOpts, keyBuf.array(), 0, keyBuf.position(), valueBytes, 0, valueBytes.length); + } catch (RocksDBException e) { + throw new StorageException(e); + } + } + + /** {@inheritDoc} */ + @Override + public @Nullable BinaryRow read(IgniteRowId rowId, @Nullable Timestamp timestamp) { + // Prepare a buffer with partition id and row id. + ByteBuffer keyBuf = prepareHeapKeyBuf(rowId); + + try ( + // Set next partition as an upper bound. + var readOpts = new ReadOptions().setIterateUpperBound(upperBound); + RocksIterator it = db.newIterator(cf, readOpts) + ) { + if (timestamp == null) { + // Seek to the first appearance of row id if timestamp isn't set. + // Since timestamps are sorted from newest to oldest, first occurance will always be the latest version. + // Unfortunately, copy here is unavoidable with current API. + it.seek(Arrays.copyOf(keyBuf.array(), keyBuf.position())); + } else { + // Put timestamp restriction according to N2O timestamps order. + putTimestamp(keyBuf, timestamp); + + // This seek will either find a key with timestamp that's less or equal than required value, or a different key whatsoever. + // It is guaranteed by descending order of timestamps. + it.seek(keyBuf.array()); + } + + // Return null if nothing was found. + if (invalid(it)) { + return null; + } + + // There's no guarantee that required key even exists. If it doesn't, then "seek" will point to a different key, obviously. + // To avoid returning its value, we have to check that actual key matches what we need. + // Here we prepare direct buffer to read key without timestamp. Shared direct buffer is used to avoid extra memory allocations. + ByteBuffer directBuffer = MV_KEY_BUFFER.get().position(0).limit(ROW_ID_OFFSET + igniteRowIdSize); + + it.key(directBuffer); + + // Comparison starts from the position of the row id. + directBuffer.position(ROW_ID_OFFSET); + + // Return null if seek found a wrong key. + if (rowId.compareTo(directBuffer, false) != 0) { + return null; + } + + // Get binary row from the iterator. It has the exact payload that we need. + byte[] valueBytes = it.value(); + + assert valueBytes != null; + + // Empty array means a tombstone. + if (valueBytes.length == 0) { + return null; + } + + return new ByteBufferRow(valueBytes); + } + } + + //TODO Play with prefix settings and benchmark results. + /** {@inheritDoc} */ + @Override + public Cursor<BinaryRow> scan(Predicate<BinaryRow> keyFilter, @Nullable Timestamp timestamp) throws StorageException { + // Set next partition as an upper bound. + var options = new ReadOptions().setIterateUpperBound(upperBound).setTotalOrderSeek(true); + + RocksIterator it = db.newIterator(cf, options); + + // Seek iterator to the beginning of the partition. + it.seek(partitionStartPrefix()); + + // Size of the buffer to seek values. Should fit partition id, row id and maybe timestamp, if it's not null. + int seekKeyBufSize = ROW_ID_OFFSET + igniteRowIdSize + (timestamp == null ? 0 : TIMESTAMP_SIZE); + + // Here's seek buffer itself. Originally it contains a valid partition id, row id payload that's filled with zeroes, and maybe + // a timestamp value. Zero row id guarantees that it's lexicographically less then or equal to any other row id stored in the + // partition. + // Byte buffer from a thread-local field can't be used here, because of two reasons: + // - no one guarantees that there will only be a single cursor; + // - no one guarantees that returned cursor will not be used other threads. + // The thing is, we need this buffer to preserve its content between invocactions of "hasNext" method. + ByteBuffer seekKeyBuf = ByteBuffer.allocate(seekKeyBufSize).order(BIG_ENDIAN).putShort((short) partId); + + if (timestamp != null) { + putTimestamp(seekKeyBuf.position(ROW_ID_OFFSET + igniteRowIdSize), timestamp); + } + + // Version without timestamp to compare row ids. + ByteBuffer seekKeyBufSliceWithoutTimestamp = timestamp == null + ? seekKeyBuf + : seekKeyBuf.position(0).limit(ROW_ID_OFFSET + igniteRowIdSize).slice().order(BIG_ENDIAN); + + return new Cursor<BinaryRow>() { + /** Cached value for {@link #next()} method. Also optimizes the code of {@link #hasNext()}. */ + private BinaryRow next = null; + + /** {@inheritDoc} */ + @Override + public boolean hasNext() { + // Fastpath for consecutive invocations. + if (next != null) { + return true; + } + + // Prepare direct buffer slice to read keys from the iterator. It's goind to be used to read row ids without timestamps. + // To enforce this, we limit it to a proper capacity. + ByteBuffer directBuffer = MV_KEY_BUFFER.get() + .position(0).limit(ROW_ID_OFFSET + igniteRowIdSize).slice().order(BIG_ENDIAN); + + while (true) { + // This flag is used to skip row ids that were created after required timestamp. + boolean found = true; + + // We should do it all the time. Here in particular it means one of two things: + // - partition is empty; + // - iterator exhausted all the data in partition. + if (invalid(it)) { + return false; + } + + // At this point, seekKeyBuf should contain row id that's above the one we already scanned, but not greater than any + // other row id in partition. When we start, row id is filled with zeroes. Value during the iteration is described later + // in this code. Now let's descrie what we'll find, assuming that iterator found something: + // - if timestamp is null: + // - this seek will find the newest version of the next row in iterator. Exactly what we need. + // - if timestamp is not null: + // - suppose that seek key buffer has the following value: "| P0 | R0 | T0 |" (partition, row id, timestamp) + // and iterator finds something, let's denote it as "| P0 | R1 | T1 |" (partition must match). Again, there are + // few possibilities here: + // - R1 == R0, this means a match. By the rules of ordering we derive that T1 >= T0. Timestamps are stored in + // descending order, this means that we found exactly what's needed. + // - R1 > R0, this means that we found next row and T1 is either missing (pending row) or represents the latest + // version of the row. It doesn't matter in this case, because this row id will be reused to find its value + // at time T0. Additional "seek" will be required to do it. + it.seek(seekKeyBuf.array()); + + // Finish scan if nothing was found. + if (invalid(it)) { + return false; + } + + // Read the actual key into a direct buffer. + it.key(directBuffer); + + // To understand this condition please read that huge comment above. + if (timestamp == null || !directBuffer.equals(seekKeyBufSliceWithoutTimestamp)) { Review Comment: Ok, I give up, let's see if it gets better -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: [email protected] For queries about this service, please contact Infrastructure at: [email protected]
