Uladzislau Blok created KAFKA-19943:
---------------------------------------

             Summary: Stale values in State Store after tombstone was compacted
                 Key: KAFKA-19943
                 URL: https://issues.apache.org/jira/browse/KAFKA-19943
             Project: Kafka
          Issue Type: Bug
          Components: streams
    Affects Versions: 3.9.1, 4.1.1
            Reporter: Uladzislau Blok


h3. *Summary*

When a Kafka Streams application with a local *state store* (backed by RocksDB) 
restarts after a period exceeding the changelog topic's 
{*}{{delete.retention.ms}}{*}, it can lead to previously deleted entities 
"magically" reappearing. This happens because the *tombstones* required to mark 
these deletions are no longer present in the compacted changelog topic.
----
h3. *Details and Observed Behavior*

The issue typically occurs in environments without *shared storage* for state 
stores (like Kubernetes with local volumes) after a *failover* or prolonged 
shutdown.
 # *Original Instance:* An entity is processed and subsequently {*}deleted{*}. 
A *tombstone* (a record with a null value) is written to the state store's 
compacted changelog topic.

 # *Downtime/Failover:* The original instance is shut down, and a new instance 
(or pod) starts after a period longer than the changelog topic's 
{{{}delete.retention.ms{}}}.

 # *Tombstone Removal:* Since the tombstone has aged past 
{{{}delete.retention.ms{}}}, the Kafka broker removes it during log compaction.

 # *Restart and Rehydration:* The new instance starts with its own, empty local 
RocksDB. It begins to *rebuild* its state store by consuming the compacted 
changelog topic.

 # *The Bug:* The deleted entity's key, while removed from the changelog, may 
still exist in the local RocksDB of the _old_ (now failed-over) instance. 
Critically, if the old instance was running a long time ago, the key/value pair 
might have existed _before_ the deletion. Since the *tombstone* is gone, there 
is nothing in the changelog to tell the new instance to *delete* that key.

 # *Symptom:* The previously deleted entity is unexpectedly revived in the new 
state store. We observed this because a {*}punctuator{*}, which scans the 
{*}entire state store{*}, began processing these revived, outdated entities.

 
----
h3. *Reproduce issue*

I was able to reproduce an issue, while doing local testing with state store 
and aggressive compaction config

Entire changelog topic:
{code:java}
/opt/kafka/bin $ ./kafka-console-consumer.sh --bootstrap-server localhost:9092 
--topic ks-state-store-issue-1-example-state-store-changelog --property 
"print.key=true" -- from-beginning
10    string10
12    string12
6    null
9    null
3    m
2    b
7    c
5    null
11    null
13    string13
4    y
10    null
3    g
2    m
7    we
4    o
7    jh
7    yt
7    vbx
7    kgf
7    cbvn {code}
 There is no entity with key: *1*

Application logs:
{code:java}
15:29:27.311 
[ks-state-store-issue-1-473580d9-4588-428b-a01e-8b5a9dbddf56-StreamThread-1] 
WARN  org.bloku.SaveAndLogProcessor - Read from state store KV: KeyValue(13, 
string13)
15:29:27.311 
[ks-state-store-issue-1-473580d9-4588-428b-a01e-8b5a9dbddf56-StreamThread-1] 
WARN  org.bloku.SaveAndLogProcessor - Read from state store KV: KeyValue(4, o)
15:29:27.608 
[ks-state-store-issue-1-473580d9-4588-428b-a01e-8b5a9dbddf56-StreamThread-2] 
WARN  org.bloku.SaveAndLogProcessor - Read from state store KV: KeyValue(1, n) 
{code}
*Read from state store KV: KeyValue(1, n)*



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to