GitHub user huishougongming added a comment to the discussion: Distributed cluster support
Thank you for your clarification It's like this, there is currently a Kafka cluster with 20 nodes, and each topic has 10 or more partitions. There are three issues How to fill in multiple host names for. adapter and sink? When the streams service restarts, it always receives this data from the latest location, including the mqtt adapter, which leads to data loss during the downtime. How can this be resolved? 3. Data received through the adapter will enter the topic automatically created by local Kafka, and then output through sink. What is the logic for creating this topic in local Kafka? For example, how to set the number of partitions? GitHub link: https://github.com/apache/streampipes/discussions/2912#discussioncomment-9814114 ---- This is an automatically sent email for [email protected]. To unsubscribe, please send an email to: [email protected]
