Hi Visha, If I remember correctly, the behavior of the Kafka consumer was changed in Flink 1.8 to account for such situations. Please check the release notes [1] and the corresponding Jira issue [2].
If this is not the behavior you need, please feel free to create a new Jira issue and start a discussion there. Thanks, Fabian [1] https://flink.apache.org/news/2019/04/09/release-1.8.0.html#important-changes [2] https://issues.apache.org/jira/browse/FLINK-10342 Am Do., 13. Juni 2019 um 22:31 Uhr schrieb Vishal Santoshi < vishal.santo...@gmail.com>: > I guess, adjusting the pattern ( blacklisting the topic/s ) would work.... > > On Thu, Jun 13, 2019 at 3:02 PM Vishal Santoshi <vishal.santo...@gmail.com> > wrote: > >> Given >> >> >> https://github.com/apache/flink/blob/master/flink-connectors/flink-connector-kafka-base/src/main/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaConsumerBase.java#L471 >> >> it seems that if >> >> * I have a regex pattern for consuming a bunch of topics >> * auto create is turned on >> >> then even if I do this >> >> * suspend the pipe with SP >> * delete the topic >> * start the pipe >> >> It will recreate the topic, b'coz the state has it and the check is >> against the pattern rather then against the cluster metadata....in case >> auto create is turned off it will throw an exception ( I think ) >> >> >> The question thus is, as to how can we safely delete a topic from our >> kafka cluster without either a recreation ( auto create on ) or exception ( >> auto create off ) when restarting from a SP, where the restore state had >> that topic.. >> >> Thanks much, >> >> Visha. >> >> >>