GeorgeJahad commented on code in PR #4045:
URL: https://github.com/apache/ozone/pull/4045#discussion_r1049082423
##########
hadoop-hdds/rocksdb-checkpoint-differ/src/main/java/org/apache/ozone/rocksdiff/RocksDBCheckpointDiffer.java:
##########
@@ -939,6 +978,114 @@ private void populateCompactionDAG(List<String>
inputFiles,
}
+ /**
+ * This is the task definition which is run periodically by the service
+ * executor at fixed delay.
+ * It looks for snapshots in compaction DAG which are older than the allowed
+ * time to be in compaction DAG and removes them from the DAG.
+ */
+ public void pruneOlderSnapshotsWithCompactionHistory() {
+ String snapshotDir = null;
+ long currentTimeMillis = System.currentTimeMillis();
+
+ while (!snapshots.isEmpty() &&
+ (currentTimeMillis - snapshots.peek().getLeft())
+ > maxAllowedTimeInDag) {
+ snapshotDir = snapshots.poll().getRight();
+ }
+
+ if (snapshotDir != null) {
+ pruneSnapshotFileNodesFromDag(snapshotDir);
+ }
+ }
+
+ /**
+ * Prunes forward and backward DAGs when oldest snapshot with compaction
+ * history gets deleted.
+ */
+ public void pruneSnapshotFileNodesFromDag(String snapshotDir) {
+ Set<String> snapshotSstFiles = readRocksDBLiveFiles(snapshotDir);
Review Comment:
The pruning is driving by the list of sst files retrieved from the snapshot
dirs, but the background deletion task may have already deleted some of those
files. Could that lead to gaps in the nodes that get pruned depending on when
the background deletion task runs?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]