This is an automated email from the ASF dual-hosted git repository.

guozhang pushed a commit to branch 2.0
in repository https://gitbox.apache.org/repos/asf/kafka.git


The following commit(s) were added to refs/heads/2.0 by this push:
     new 38fd13d  MINOR: standby task test throughput too low 2.0 (#6062)
38fd13d is described below

commit 38fd13d9c08778bf3e9a950c0062ef8e41c5282f
Author: Bill Bejeck <bbej...@gmail.com>
AuthorDate: Sat Dec 22 01:41:28 2018 -0500

    MINOR: standby task test throughput too low 2.0 (#6062)
    
    Previous PR #6043 reduced throughput for VerifiableProducer in base class, 
but the streams_standby_replica_test needs higher throughput for consumer to 
complete verification in 60 seconds. Same update as #6060 and #6061
    
    Reviewers: Guozhang Wang <wangg...@gmail.com>
---
 tests/kafkatest/tests/streams/base_streams_test.py            | 4 ++--
 tests/kafkatest/tests/streams/streams_standby_replica_test.py | 2 +-
 2 files changed, 3 insertions(+), 3 deletions(-)

diff --git a/tests/kafkatest/tests/streams/base_streams_test.py 
b/tests/kafkatest/tests/streams/base_streams_test.py
index 6e005dd..9a9704e 100644
--- a/tests/kafkatest/tests/streams/base_streams_test.py
+++ b/tests/kafkatest/tests/streams/base_streams_test.py
@@ -38,14 +38,14 @@ class BaseStreamsTest(KafkaTest):
                                   client_id,
                                   max_messages=num_messages)
 
-    def get_producer(self, topic, num_messages, repeating_keys=None):
+    def get_producer(self, topic, num_messages, throughput=1000, 
repeating_keys=None):
         return VerifiableProducer(self.test_context,
                                   1,
                                   self.kafka,
                                   topic,
                                   max_messages=num_messages,
                                   acks=1,
-                                  throughput=1000,
+                                  throughput=throughput,
                                   repeating_keys=repeating_keys)
 
     def assert_produce_consume(self,
diff --git a/tests/kafkatest/tests/streams/streams_standby_replica_test.py 
b/tests/kafkatest/tests/streams/streams_standby_replica_test.py
index 416a110..8425e14 100644
--- a/tests/kafkatest/tests/streams/streams_standby_replica_test.py
+++ b/tests/kafkatest/tests/streams/streams_standby_replica_test.py
@@ -46,7 +46,7 @@ class StreamsStandbyTask(BaseStreamsTest):
                                                                                
     self.streams_sink_topic_1,
                                                                                
     self.streams_sink_topic_2))
 
-        producer = self.get_producer(self.streams_source_topic, 
self.num_messages, repeating_keys=6)
+        producer = self.get_producer(self.streams_source_topic, 
self.num_messages, throughput=15000, repeating_keys=6)
         producer.start()
 
         processor_1 = StreamsStandbyTaskService(self.test_context, self.kafka, 
configs)

Reply via email to