mike-tr-adamson commented on code in PR #2409: URL: https://github.com/apache/cassandra/pull/2409#discussion_r1242367814
########## src/java/org/apache/cassandra/index/sai/disk/v1/bbtree/BlockBalancedTreeReader.java: ########## @@ -0,0 +1,455 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.cassandra.index.sai.disk.v1.bbtree; + +import java.io.Closeable; +import java.io.IOException; +import java.lang.invoke.MethodHandles; +import java.util.Comparator; +import java.util.PriorityQueue; +import java.util.concurrent.TimeUnit; + +import com.google.common.base.Preconditions; +import com.google.common.base.Stopwatch; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + +import org.apache.cassandra.exceptions.QueryCancelledException; +import org.apache.cassandra.index.sai.IndexContext; +import org.apache.cassandra.index.sai.QueryContext; +import org.apache.cassandra.index.sai.disk.io.IndexFileUtils; +import org.apache.cassandra.index.sai.disk.io.SeekingRandomAccessInput; +import org.apache.cassandra.index.sai.disk.v1.DirectReaders; +import org.apache.cassandra.index.sai.disk.v1.postings.FilteringPostingList; +import org.apache.cassandra.index.sai.disk.v1.postings.MergePostingList; +import org.apache.cassandra.index.sai.disk.v1.postings.PostingsReader; +import org.apache.cassandra.index.sai.metrics.QueryEventListener; +import org.apache.cassandra.index.sai.postings.PeekablePostingList; +import org.apache.cassandra.index.sai.postings.PostingList; +import org.apache.cassandra.io.util.FileHandle; +import org.apache.cassandra.io.util.FileUtils; +import org.apache.cassandra.utils.Throwables; +import org.apache.lucene.index.CorruptIndexException; +import org.apache.lucene.index.PointValues.Relation; +import org.apache.lucene.store.IndexInput; +import org.apache.lucene.util.BytesRef; +import org.apache.lucene.util.FixedBitSet; +import org.apache.lucene.util.FutureArrays; +import org.apache.lucene.util.packed.DirectWriter; + +/** + * Handles intersection of a point or point range with a block balanced tree previously written with + * {@link BlockBalancedTreeWriter}. + */ +public class BlockBalancedTreeReader extends TraversingBlockBalancedTreeReader implements Closeable +{ + private static final Logger logger = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass()); + + private static final Comparator<PeekablePostingList> COMPARATOR = Comparator.comparingLong(PeekablePostingList::peek); + + private final IndexContext indexContext; + private final FileHandle postingsFile; + private final BlockBalancedTreePostingsIndex postingsIndex; + private final DirectReaders.Reader leafOrderMapReader; + + /** + * Performs a blocking read. + */ + public BlockBalancedTreeReader(IndexContext indexContext, + FileHandle treeIndexFile, + long treeIndexRoot, + FileHandle postingsFile, + long treePostingsRoot) throws IOException + { + super(treeIndexFile, treeIndexRoot); + this.indexContext = indexContext; + this.postingsFile = postingsFile; + this.postingsIndex = new BlockBalancedTreePostingsIndex(postingsFile, treePostingsRoot); + leafOrderMapReader = DirectReaders.getReaderForBitsPerValue((byte) DirectWriter.unsignedBitsRequired(maxPointsInLeafNode - 1)); + } + + public int getBytesPerValue() + { + return bytesPerValue; + } + + public long getPointCount() + { + return pointCount; + } + + @Override + public void close() + { + super.close(); + FileUtils.closeQuietly(postingsFile); + } + + @SuppressWarnings({"resource", "RedundantSuppression"}) + public PostingList intersect(IntersectVisitor visitor, QueryEventListener.BalancedTreeEventListener listener, QueryContext context) + { + Relation relation = visitor.compare(minPackedValue, maxPackedValue); + + if (relation == Relation.CELL_OUTSIDE_QUERY) + { + listener.onIntersectionEarlyExit(); + return null; + } + + listener.onSegmentHit(); + IndexInput treeInput = IndexFileUtils.instance.openInput(treeIndexFile); + IndexInput postingsInput = IndexFileUtils.instance.openInput(postingsFile); + IndexInput postingsSummaryInput = IndexFileUtils.instance.openInput(postingsFile); + PackedIndexTree index = new PackedIndexTree(); + + Intersection intersection = relation == Relation.CELL_INSIDE_QUERY + ? new Intersection(treeInput, postingsInput, postingsSummaryInput, index, listener, context) + : new FilteringIntersection(treeInput, postingsInput, postingsSummaryInput, index, visitor, listener, context); + + return intersection.execute(); + } + + /** + * Synchronous intersection of a point or point range with a block balanced tree previously written + * with {@link BlockBalancedTreeWriter}. + */ + private class Intersection + { + private final Stopwatch queryExecutionTimer = Stopwatch.createStarted(); + final QueryContext context; + + final IndexInput treeInput; + final IndexInput postingsInput; + final IndexInput postingsSummaryInput; + final PackedIndexTree index; + final QueryEventListener.BalancedTreeEventListener listener; + + Intersection(IndexInput treeInput, IndexInput postingsInput, IndexInput postingsSummaryInput, + PackedIndexTree index, QueryEventListener.BalancedTreeEventListener listener, QueryContext context) + { + this.treeInput = treeInput; + this.postingsInput = postingsInput; + this.postingsSummaryInput = postingsSummaryInput; + this.index = index; + this.listener = listener; + this.context = context; + } + + public PostingList execute() + { + try + { + PriorityQueue<PeekablePostingList> postingLists = new PriorityQueue<>(100, COMPARATOR); + + executeInternal(postingLists); + + FileUtils.closeQuietly(treeInput); + + return mergePostings(postingLists); + } + catch (Throwable t) + { + if (!(t instanceof QueryCancelledException)) + logger.error(indexContext.logMessage("Balanced tree intersection failed on {}"), treeIndexFile.path(), t); + + closeOnException(); + throw Throwables.cleaned(t); + } + } + + protected void executeInternal(final PriorityQueue<PeekablePostingList> postingLists) throws IOException + { + collectPostingLists(postingLists); + } + + protected void closeOnException() + { + FileUtils.closeQuietly(treeInput); + FileUtils.closeQuietly(postingsInput); + FileUtils.closeQuietly(postingsSummaryInput); + } + + protected PostingList mergePostings(PriorityQueue<PeekablePostingList> postingLists) + { + final long elapsedMicros = queryExecutionTimer.stop().elapsed(TimeUnit.MICROSECONDS); + + listener.onIntersectionComplete(elapsedMicros, TimeUnit.MICROSECONDS); + listener.postingListsHit(postingLists.size()); + + if (postingLists.isEmpty()) + { + FileUtils.closeQuietly(postingsInput); + FileUtils.closeQuietly(postingsSummaryInput); + return null; + } + else + { + if (logger.isTraceEnabled()) + logger.trace(indexContext.logMessage("[{}] Intersection completed in {} microseconds. {} leaf and internal posting lists hit."), + treeIndexFile.path(), elapsedMicros, postingLists.size()); + return MergePostingList.merge(postingLists, () -> FileUtils.close(postingsInput, postingsSummaryInput)); + } + } + + private void collectPostingLists(PriorityQueue<PeekablePostingList> postingLists) throws IOException + { + context.checkpoint(); + + final int nodeID = index.getNodeID(); + + // if there is pre-built posting for entire subtree + if (postingsIndex.exists(nodeID)) + { + postingLists.add(initPostingReader(postingsIndex.getPostingsFilePointer(nodeID))); + return; + } + + Preconditions.checkState(!index.isLeafNode(), "Leaf node %s does not have balanced tree postings.", index.getNodeID()); + + // Recurse on left subtree: + index.pushLeft(); + collectPostingLists(postingLists); + index.pop(); + + // Recurse on right subtree: + index.pushRight(); + collectPostingLists(postingLists); + index.pop(); + } + + @SuppressWarnings({"resource", "RedundantSuppression"}) + private PeekablePostingList initPostingReader(long offset) throws IOException + { + final PostingsReader.BlocksSummary summary = new PostingsReader.BlocksSummary(postingsSummaryInput, offset); + return PeekablePostingList.makePeekable(new PostingsReader(postingsInput, summary, listener.postingListEventListener())); + } + } + + private class FilteringIntersection extends Intersection + { + private final IntersectVisitor visitor; + private final byte[] packedValue; + private final short[] origIndex; + + FilteringIntersection(IndexInput treeInput, IndexInput postingsInput, IndexInput postingsSummaryInput, + PackedIndexTree index, IntersectVisitor visitor, + QueryEventListener.BalancedTreeEventListener listener, QueryContext context) + { + super(treeInput, postingsInput, postingsSummaryInput, index, listener, context); + this.visitor = visitor; + this.packedValue = new byte[packedBytesLength]; + this.origIndex = new short[maxPointsInLeafNode]; + } + + @Override + public void executeInternal(final PriorityQueue<PeekablePostingList> postingLists) throws IOException + { + collectPostingLists(postingLists, minPackedValue, maxPackedValue); + } + + private void collectPostingLists(PriorityQueue<PeekablePostingList> postingLists, byte[] cellMinPacked, byte[] cellMaxPacked) throws IOException + { + context.checkpoint(); + + final Relation r = visitor.compare(cellMinPacked, cellMaxPacked); + + if (r == Relation.CELL_OUTSIDE_QUERY) + { + // This cell is fully outside the query shape: stop recursing + return; + } + + if (r == Relation.CELL_INSIDE_QUERY) + { + // This cell is fully inside the query shape: recursively add all points in this cell without filtering + super.collectPostingLists(postingLists); + return; + } + + if (index.isLeafNode()) + { + if (index.nodeExists()) + filterLeaf(postingLists); + return; + } + + visitNode(postingLists, cellMinPacked, cellMaxPacked); + } + + private void filterLeaf(PriorityQueue<PeekablePostingList> postingLists) throws IOException + { + treeInput.seek(index.getLeafBlockFP()); + + int count = treeInput.readVInt(); + + // loading doc ids occurred here prior + + int orderMapLength = treeInput.readVInt(); + + long orderMapPointer = treeInput.getFilePointer(); + + SeekingRandomAccessInput randomAccessInput = new SeekingRandomAccessInput(treeInput); + for (int x = 0; x < count; x++) + { + origIndex[x] = (short) LeafOrderMap.getValue(randomAccessInput, orderMapPointer, x, leafOrderMapReader); + } + + // seek beyond the ordermap + treeInput.seek(orderMapPointer + orderMapLength); + + FixedBitSet fixedBitSet = visitDocValues(treeInput, count, visitor, origIndex); Review Comment: Yes, I like that. I have carried that through to `buildPostingsFilterForSingleValueLeaf` and `buildPostingsFilterForMultValueLeaf`. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: [email protected] For queries about this service, please contact Infrastructure at: [email protected] --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]

