http://git-wip-us.apache.org/repos/asf/kafka/blob/d50499a0/system_test/replication_testsuite/testcase_4015/testcase_4015_properties.json ---------------------------------------------------------------------- diff --git a/system_test/replication_testsuite/testcase_4015/testcase_4015_properties.json b/system_test/replication_testsuite/testcase_4015/testcase_4015_properties.json deleted file mode 100644 index 7841273..0000000 --- a/system_test/replication_testsuite/testcase_4015/testcase_4015_properties.json +++ /dev/null @@ -1,113 +0,0 @@ -{ - "description": {"01":"Broker Log Retention (Replica Factor < No. of Brokers) : 1. sync => false, acks => -1, comp => 0", - "02":"Produce and consume messages to 2 topics - 2 partitions", - "03":"This test sends messages to 2 replicas", - "04":"To trigger leader election: find the leader and terminate by controlled failure (kill -15)", - "05":"Restart the terminated broker", - "06":"Lookup brokers' log4j messages and verify that leader is re-elected successfully", - "07":"At the end it verifies the log size and contents", - "08":"Use a consumer to verify no message loss.", - "09":"Producer dimensions : mode:async, acks:-1, comp:0", - "10":"Log segment size : 102400" - }, - "testcase_args": { - "broker_type": "leader", - "bounce_broker": "true", - "replica_factor": "2", - "num_partition": "2", - "num_iteration": "1", - "sleep_seconds_between_producer_calls": "1", - "broker_down_time_in_sec": "5", - "message_producing_free_time_sec": "15", - "log_retention_test": "true" - }, - "entities": [ - { - "entity_id": "0", - "clientPort": "2188", - "dataDir": "/tmp/zookeeper_0", - "log_filename": "zookeeper_2188.log", - "config_filename": "zookeeper_2188.properties" - }, - { - "entity_id": "1", - "port": "9091", - "broker.id": "1", - "log.segment.bytes": "102400", - "log.retention.size": "1048576", - "log.dir": "/tmp/kafka_server_1_logs", - "default.replication.factor": "2", - "num.partitions": "2", - "log_filename": "kafka_server_9091.log", - "config_filename": "kafka_server_9091.properties" - }, - { - "entity_id": "2", - "port": "9092", - "broker.id": "2", - "log.segment.bytes": "102400", - "log.retention.size": "1048576", - "log.dir": "/tmp/kafka_server_2_logs", - "default.replication.factor": "2", - "num.partitions": "2", - "log_filename": "kafka_server_9092.log", - "config_filename": "kafka_server_9092.properties" - }, - { - "entity_id": "3", - "port": "9093", - "broker.id": "3", - "log.segment.bytes": "102400", - "log.retention.size": "1048576", - "log.dir": "/tmp/kafka_server_3_logs", - "default.replication.factor": "2", - "num.partitions": "2", - "log_filename": "kafka_server_9093.log", - "config_filename": "kafka_server_9093.properties" - }, - { - "entity_id": "4", - "new-producer":"true", - "topic": "test_1", - "threads": "5", - "compression-codec": "0", - "message-size": "500", - "message": "500", - "request-num-acks": "-1", - "producer-retry-backoff-ms": "300", - "sync":"false", - "log_filename": "producer_performance_4.log", - "config_filename": "producer_performance_4.properties" - }, - { - "entity_id": "5", - "new-producer":"true", - "topic": "test_2", - "threads": "5", - "compression-codec": "0", - "message-size": "500", - "message": "500", - "request-num-acks": "-1", - "producer-retry-backoff-ms": "300", - "sync":"false", - "log_filename": "producer_performance_5.log", - "config_filename": "producer_performance_5.properties" - }, - { - "entity_id": "6", - "topic": "test_1", - "group.id": "mytestgroup", - "consumer-timeout-ms": "60000", - "log_filename": "console_consumer_6.log", - "config_filename": "console_consumer_6.properties" - }, - { - "entity_id": "7", - "topic": "test_2", - "group.id": "mytestgroup", - "consumer-timeout-ms": "60000", - "log_filename": "console_consumer_7.log", - "config_filename": "console_consumer_7.properties" - } - ] -}
http://git-wip-us.apache.org/repos/asf/kafka/blob/d50499a0/system_test/replication_testsuite/testcase_4016/cluster_config.json ---------------------------------------------------------------------- diff --git a/system_test/replication_testsuite/testcase_4016/cluster_config.json b/system_test/replication_testsuite/testcase_4016/cluster_config.json deleted file mode 100644 index 9e733cf..0000000 --- a/system_test/replication_testsuite/testcase_4016/cluster_config.json +++ /dev/null @@ -1,76 +0,0 @@ -{ - "cluster_config": [ - { - "entity_id": "0", - "hostname": "localhost", - "role": "zookeeper", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9990" - }, - { - "entity_id": "1", - "hostname": "localhost", - "role": "broker", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9991" - }, - { - "entity_id": "2", - "hostname": "localhost", - "role": "broker", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9992" - }, - { - "entity_id": "3", - "hostname": "localhost", - "role": "broker", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9993" - }, - { - "entity_id": "4", - "hostname": "localhost", - "role": "producer_performance", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9997" - }, - { - "entity_id": "5", - "hostname": "localhost", - "role": "producer_performance", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9998" - }, - { - "entity_id": "6", - "hostname": "localhost", - "role": "console_consumer", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9999" - }, - { - "entity_id": "7", - "hostname": "localhost", - "role": "console_consumer", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9191" - } - ] -} http://git-wip-us.apache.org/repos/asf/kafka/blob/d50499a0/system_test/replication_testsuite/testcase_4016/testcase_4016_properties.json ---------------------------------------------------------------------- diff --git a/system_test/replication_testsuite/testcase_4016/testcase_4016_properties.json b/system_test/replication_testsuite/testcase_4016/testcase_4016_properties.json deleted file mode 100644 index 0519d27..0000000 --- a/system_test/replication_testsuite/testcase_4016/testcase_4016_properties.json +++ /dev/null @@ -1,113 +0,0 @@ -{ - "description": {"01":"Broker Log Retention (Replica Factor < No. of Brokers) : 1. acks => 1", - "02":"Produce and consume messages to 2 topics - 2 partitions", - "03":"This test sends messages to 2 replicas", - "04":"To trigger leader election: find the leader and terminate by controlled failure (kill -15)", - "05":"Restart the terminated broker", - "06":"Lookup brokers' log4j messages and verify that leader is re-elected successfully", - "07":"At the end it verifies the log size and contents", - "08":"Use a consumer to verify no message loss.", - "09":"Producer dimensions : mode:async, acks:1, comp:0", - "10":"Log segment size : 102400" - }, - "testcase_args": { - "broker_type": "leader", - "bounce_broker": "true", - "replica_factor": "2", - "num_partition": "2", - "num_iteration": "1", - "sleep_seconds_between_producer_calls": "1", - "broker_down_time_in_sec": "5", - "message_producing_free_time_sec": "15", - "log_retention_test": "true" - }, - "entities": [ - { - "entity_id": "0", - "clientPort": "2188", - "dataDir": "/tmp/zookeeper_0", - "log_filename": "zookeeper_2188.log", - "config_filename": "zookeeper_2188.properties" - }, - { - "entity_id": "1", - "port": "9091", - "broker.id": "1", - "log.segment.bytes": "102400", - "log.retention.size": "1048576", - "log.dir": "/tmp/kafka_server_1_logs", - "default.replication.factor": "2", - "num.partitions": "2", - "log_filename": "kafka_server_9091.log", - "config_filename": "kafka_server_9091.properties" - }, - { - "entity_id": "2", - "port": "9092", - "broker.id": "2", - "log.segment.bytes": "102400", - "log.retention.size": "1048576", - "log.dir": "/tmp/kafka_server_2_logs", - "default.replication.factor": "2", - "num.partitions": "2", - "log_filename": "kafka_server_9092.log", - "config_filename": "kafka_server_9092.properties" - }, - { - "entity_id": "3", - "port": "9093", - "broker.id": "3", - "log.segment.bytes": "102400", - "log.retention.size": "1048576", - "log.dir": "/tmp/kafka_server_3_logs", - "default.replication.factor": "2", - "num.partitions": "2", - "log_filename": "kafka_server_9093.log", - "config_filename": "kafka_server_9093.properties" - }, - { - "entity_id": "4", - "new-producer":"true", - "topic": "test_1", - "threads": "5", - "compression-codec": "0", - "message-size": "500", - "message": "500", - "request-num-acks": "1", - "producer-retry-backoff-ms": "300", - "sync":"false", - "log_filename": "producer_performance_4.log", - "config_filename": "producer_performance_4.properties" - }, - { - "entity_id": "5", - "new-producer":"true", - "topic": "test_2", - "threads": "5", - "compression-codec": "0", - "message-size": "500", - "message": "500", - "request-num-acks": "1", - "producer-retry-backoff-ms": "300", - "sync":"false", - "log_filename": "producer_performance_5.log", - "config_filename": "producer_performance_5.properties" - }, - { - "entity_id": "6", - "topic": "test_1", - "group.id": "mytestgroup", - "consumer-timeout-ms": "60000", - "log_filename": "console_consumer_6.log", - "config_filename": "console_consumer_6.properties" - }, - { - "entity_id": "7", - "topic": "test_2", - "group.id": "mytestgroup", - "consumer-timeout-ms": "60000", - "log_filename": "console_consumer_7.log", - "config_filename": "console_consumer_7.properties" - } - ] -} http://git-wip-us.apache.org/repos/asf/kafka/blob/d50499a0/system_test/replication_testsuite/testcase_4017/cluster_config.json ---------------------------------------------------------------------- diff --git a/system_test/replication_testsuite/testcase_4017/cluster_config.json b/system_test/replication_testsuite/testcase_4017/cluster_config.json deleted file mode 100644 index 9e733cf..0000000 --- a/system_test/replication_testsuite/testcase_4017/cluster_config.json +++ /dev/null @@ -1,76 +0,0 @@ -{ - "cluster_config": [ - { - "entity_id": "0", - "hostname": "localhost", - "role": "zookeeper", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9990" - }, - { - "entity_id": "1", - "hostname": "localhost", - "role": "broker", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9991" - }, - { - "entity_id": "2", - "hostname": "localhost", - "role": "broker", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9992" - }, - { - "entity_id": "3", - "hostname": "localhost", - "role": "broker", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9993" - }, - { - "entity_id": "4", - "hostname": "localhost", - "role": "producer_performance", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9997" - }, - { - "entity_id": "5", - "hostname": "localhost", - "role": "producer_performance", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9998" - }, - { - "entity_id": "6", - "hostname": "localhost", - "role": "console_consumer", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9999" - }, - { - "entity_id": "7", - "hostname": "localhost", - "role": "console_consumer", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9191" - } - ] -} http://git-wip-us.apache.org/repos/asf/kafka/blob/d50499a0/system_test/replication_testsuite/testcase_4017/testcase_4017_properties.json ---------------------------------------------------------------------- diff --git a/system_test/replication_testsuite/testcase_4017/testcase_4017_properties.json b/system_test/replication_testsuite/testcase_4017/testcase_4017_properties.json deleted file mode 100644 index c29077b..0000000 --- a/system_test/replication_testsuite/testcase_4017/testcase_4017_properties.json +++ /dev/null @@ -1,113 +0,0 @@ -{ - "description": {"01":"Broker Log Retention (Replica Factor < No. of Brokers) : 1. acks => -1, 2. comp => 1", - "02":"Produce and consume messages to 2 topics - 2 partitions", - "03":"This test sends messages to 2 replicas", - "04":"To trigger leader election: find the leader and terminate by controlled failure (kill -15)", - "05":"Restart the terminated broker", - "06":"Lookup brokers' log4j messages and verify that leader is re-elected successfully", - "07":"At the end it verifies the log size and contents", - "08":"Use a consumer to verify no message loss.", - "09":"Producer dimensions : mode:async, acks:-1, comp:1", - "10":"Log segment size : 102400" - }, - "testcase_args": { - "broker_type": "leader", - "bounce_broker": "true", - "replica_factor": "2", - "num_partition": "2", - "num_iteration": "1", - "sleep_seconds_between_producer_calls": "1", - "broker_down_time_in_sec": "5", - "message_producing_free_time_sec": "15", - "log_retention_test": "true" - }, - "entities": [ - { - "entity_id": "0", - "clientPort": "2188", - "dataDir": "/tmp/zookeeper_0", - "log_filename": "zookeeper_2188.log", - "config_filename": "zookeeper_2188.properties" - }, - { - "entity_id": "1", - "port": "9091", - "broker.id": "1", - "log.segment.bytes": "102400", - "log.retention.size": "1048576", - "log.dir": "/tmp/kafka_server_1_logs", - "default.replication.factor": "2", - "num.partitions": "2", - "log_filename": "kafka_server_9091.log", - "config_filename": "kafka_server_9091.properties" - }, - { - "entity_id": "2", - "port": "9092", - "broker.id": "2", - "log.segment.bytes": "102400", - "log.retention.size": "1048576", - "log.dir": "/tmp/kafka_server_2_logs", - "default.replication.factor": "2", - "num.partitions": "2", - "log_filename": "kafka_server_9092.log", - "config_filename": "kafka_server_9092.properties" - }, - { - "entity_id": "3", - "port": "9093", - "broker.id": "3", - "log.segment.bytes": "102400", - "log.retention.size": "1048576", - "log.dir": "/tmp/kafka_server_3_logs", - "default.replication.factor": "2", - "num.partitions": "2", - "log_filename": "kafka_server_9093.log", - "config_filename": "kafka_server_9093.properties" - }, - { - "entity_id": "4", - "new-producer":"true", - "topic": "test_1", - "threads": "5", - "compression-codec": "1", - "message-size": "500", - "message": "500", - "request-num-acks": "-1", - "producer-retry-backoff-ms": "300", - "sync":"false", - "log_filename": "producer_performance_4.log", - "config_filename": "producer_performance_4.properties" - }, - { - "entity_id": "5", - "new-producer":"true", - "topic": "test_2", - "threads": "5", - "compression-codec": "1", - "message-size": "500", - "message": "500", - "request-num-acks": "-1", - "producer-retry-backoff-ms": "300", - "sync":"false", - "log_filename": "producer_performance_5.log", - "config_filename": "producer_performance_5.properties" - }, - { - "entity_id": "6", - "topic": "test_1", - "group.id": "mytestgroup", - "consumer-timeout-ms": "60000", - "log_filename": "console_consumer_6.log", - "config_filename": "console_consumer_6.properties" - }, - { - "entity_id": "7", - "topic": "test_2", - "group.id": "mytestgroup", - "consumer-timeout-ms": "60000", - "log_filename": "console_consumer_7.log", - "config_filename": "console_consumer_7.properties" - } - ] -} http://git-wip-us.apache.org/repos/asf/kafka/blob/d50499a0/system_test/replication_testsuite/testcase_4018/cluster_config.json ---------------------------------------------------------------------- diff --git a/system_test/replication_testsuite/testcase_4018/cluster_config.json b/system_test/replication_testsuite/testcase_4018/cluster_config.json deleted file mode 100644 index 9e733cf..0000000 --- a/system_test/replication_testsuite/testcase_4018/cluster_config.json +++ /dev/null @@ -1,76 +0,0 @@ -{ - "cluster_config": [ - { - "entity_id": "0", - "hostname": "localhost", - "role": "zookeeper", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9990" - }, - { - "entity_id": "1", - "hostname": "localhost", - "role": "broker", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9991" - }, - { - "entity_id": "2", - "hostname": "localhost", - "role": "broker", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9992" - }, - { - "entity_id": "3", - "hostname": "localhost", - "role": "broker", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9993" - }, - { - "entity_id": "4", - "hostname": "localhost", - "role": "producer_performance", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9997" - }, - { - "entity_id": "5", - "hostname": "localhost", - "role": "producer_performance", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9998" - }, - { - "entity_id": "6", - "hostname": "localhost", - "role": "console_consumer", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9999" - }, - { - "entity_id": "7", - "hostname": "localhost", - "role": "console_consumer", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9191" - } - ] -} http://git-wip-us.apache.org/repos/asf/kafka/blob/d50499a0/system_test/replication_testsuite/testcase_4018/testcase_4018_properties.json ---------------------------------------------------------------------- diff --git a/system_test/replication_testsuite/testcase_4018/testcase_4018_properties.json b/system_test/replication_testsuite/testcase_4018/testcase_4018_properties.json deleted file mode 100644 index ab57e5a..0000000 --- a/system_test/replication_testsuite/testcase_4018/testcase_4018_properties.json +++ /dev/null @@ -1,113 +0,0 @@ -{ - "description": {"01":"Broker Log Retention (Replica Factor < No. of Brokers) : 1. acks => 1", - "02":"Produce and consume messages to 2 topics - 2 partitions", - "03":"This test sends messages to 2 replicas", - "04":"To trigger leader election: find the leader and terminate by controlled failure (kill -15)", - "05":"Restart the terminated broker", - "06":"Lookup brokers' log4j messages and verify that leader is re-elected successfully", - "07":"At the end it verifies the log size and contents", - "08":"Use a consumer to verify no message loss.", - "09":"Producer dimensions : mode:async, acks:1, comp:1", - "10":"Log segment size : 102400" - }, - "testcase_args": { - "broker_type": "leader", - "bounce_broker": "true", - "replica_factor": "2", - "num_partition": "2", - "num_iteration": "1", - "sleep_seconds_between_producer_calls": "1", - "broker_down_time_in_sec": "5", - "message_producing_free_time_sec": "15", - "log_retention_test": "true" - }, - "entities": [ - { - "entity_id": "0", - "clientPort": "2188", - "dataDir": "/tmp/zookeeper_0", - "log_filename": "zookeeper_2188.log", - "config_filename": "zookeeper_2188.properties" - }, - { - "entity_id": "1", - "port": "9091", - "broker.id": "1", - "log.segment.bytes": "102400", - "log.retention.size": "1048576", - "log.dir": "/tmp/kafka_server_1_logs", - "default.replication.factor": "2", - "num.partitions": "2", - "log_filename": "kafka_server_9091.log", - "config_filename": "kafka_server_9091.properties" - }, - { - "entity_id": "2", - "port": "9092", - "broker.id": "2", - "log.segment.bytes": "102400", - "log.retention.size": "1048576", - "log.dir": "/tmp/kafka_server_2_logs", - "default.replication.factor": "2", - "num.partitions": "2", - "log_filename": "kafka_server_9092.log", - "config_filename": "kafka_server_9092.properties" - }, - { - "entity_id": "3", - "port": "9093", - "broker.id": "3", - "log.segment.bytes": "102400", - "log.retention.size": "1048576", - "log.dir": "/tmp/kafka_server_3_logs", - "default.replication.factor": "2", - "num.partitions": "2", - "log_filename": "kafka_server_9093.log", - "config_filename": "kafka_server_9093.properties" - }, - { - "entity_id": "4", - "new-producer":"true", - "topic": "test_1", - "threads": "5", - "compression-codec": "1", - "message-size": "500", - "message": "500", - "request-num-acks": "1", - "producer-retry-backoff-ms": "300", - "sync":"false", - "log_filename": "producer_performance_4.log", - "config_filename": "producer_performance_4.properties" - }, - { - "entity_id": "5", - "new-producer":"true", - "topic": "test_2", - "threads": "5", - "compression-codec": "1", - "message-size": "500", - "message": "500", - "request-num-acks": "1", - "producer-retry-backoff-ms": "300", - "sync":"false", - "log_filename": "producer_performance_5.log", - "config_filename": "producer_performance_5.properties" - }, - { - "entity_id": "6", - "topic": "test_1", - "group.id": "mytestgroup", - "consumer-timeout-ms": "60000", - "log_filename": "console_consumer_6.log", - "config_filename": "console_consumer_6.properties" - }, - { - "entity_id": "7", - "topic": "test_2", - "group.id": "mytestgroup", - "consumer-timeout-ms": "60000", - "log_filename": "console_consumer_7.log", - "config_filename": "console_consumer_7.properties" - } - ] -} http://git-wip-us.apache.org/repos/asf/kafka/blob/d50499a0/system_test/replication_testsuite/testcase_9051/cluster_config.json ---------------------------------------------------------------------- diff --git a/system_test/replication_testsuite/testcase_9051/cluster_config.json b/system_test/replication_testsuite/testcase_9051/cluster_config.json deleted file mode 100644 index 8ed896b..0000000 --- a/system_test/replication_testsuite/testcase_9051/cluster_config.json +++ /dev/null @@ -1,58 +0,0 @@ -{ - "cluster_config": [ - { - "entity_id": "0", - "hostname": "localhost", - "role": "zookeeper", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9990" - }, - { - "entity_id": "1", - "hostname": "localhost", - "role": "broker", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9991" - }, - { - "entity_id": "2", - "hostname": "localhost", - "role": "broker", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9992" - }, - { - "entity_id": "3", - "hostname": "localhost", - "role": "broker", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9993" - }, - { - "entity_id": "4", - "hostname": "localhost", - "role": "producer_performance", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9997" - }, - { - "entity_id": "5", - "hostname": "localhost", - "role": "console_consumer", - "cluster_name": "source", - "kafka_home": "default", - "java_home": "default", - "jmx_port": "9998" - } - ] -} http://git-wip-us.apache.org/repos/asf/kafka/blob/d50499a0/system_test/replication_testsuite/testcase_9051/testcase_9051_properties.json ---------------------------------------------------------------------- diff --git a/system_test/replication_testsuite/testcase_9051/testcase_9051_properties.json b/system_test/replication_testsuite/testcase_9051/testcase_9051_properties.json deleted file mode 100644 index e959aed..0000000 --- a/system_test/replication_testsuite/testcase_9051/testcase_9051_properties.json +++ /dev/null @@ -1,86 +0,0 @@ -{ - "description": {"01":"To Test : 'Leader Failure in Replication'", - "02":"Produce and consume messages to 300 topics - 4 partitions.", - "03":"This test sends messages to 3 replicas", - "04":"To trigger leader election: find the leader and terminate by controlled failure (kill -15)", - "05":"Restart the terminated broker", - "06":"Lookup brokers' log4j messages and verify that leader is re-elected successfully", - "07":"At the end it verifies the log size and contents", - "08":"Use a consumer to verify no message loss.", - "09":"Producer dimensions : mode:sync, acks:-1, comp:0", - "10":"Log segment size : 1048576" - }, - "testcase_args": { - "broker_type": "leader", - "bounce_broker": "false", - "replica_factor": "3", - "num_partition": "2", - "num_iteration": "1", - "producer_multi_topics_mode": "true", - "consumer_multi_topics_mode": "true", - "sleep_seconds_between_producer_calls": "5", - "message_producing_free_time_sec": "15", - "num_topics_for_auto_generated_string": "20", - "num_messages_to_produce_per_producer_call": "50" - }, - "entities": [ - { - "entity_id": "0", - "clientPort": "2188", - "dataDir": "/tmp/zookeeper_0", - "log_filename": "zookeeper_2188.log", - "config_filename": "zookeeper_2188.properties" - }, - { - "entity_id": "1", - "port": "9091", - "broker.id": "1", - "log.segment.bytes": "1048576", - "log.dir": "/tmp/kafka_server_1_logs", - "log_filename": "kafka_server_9091.log", - "config_filename": "kafka_server_9091.properties" - }, - { - "entity_id": "2", - "port": "9092", - "broker.id": "2", - "log.segment.bytes": "1048576", - "log.dir": "/tmp/kafka_server_2_logs", - "log_filename": "kafka_server_9092.log", - "config_filename": "kafka_server_9092.properties" - }, - { - "entity_id": "3", - "port": "9093", - "broker.id": "3", - "log.segment.bytes": "1048576", - "log.dir": "/tmp/kafka_server_3_logs", - "log_filename": "kafka_server_9093.log", - "config_filename": "kafka_server_9093.properties" - }, - { - "entity_id": "4", - "new-producer":"true", - "topic": "t001", - "threads": "5", - "compression-codec": "0", - "message-size": "500", - "message": "500", - "request-num-acks": "-1", - "producer-retry-backoff-ms": "3500", - "producer-num-retries": "3", - "async":"false", - "log_filename": "producer_performance_9.log", - "config_filename": "producer_performance_9.properties" - }, - { - "entity_id": "5", - "topic": "t001", - "group.id": "mytestgroup", - "consumer-timeout-ms": "60000", - "zookeeper": "localhost:2188", - "log_filename": "console_consumer_10.log", - "config_filename": "console_consumer_10.properties" - } - ] -} http://git-wip-us.apache.org/repos/asf/kafka/blob/d50499a0/system_test/run_all.sh ---------------------------------------------------------------------- diff --git a/system_test/run_all.sh b/system_test/run_all.sh deleted file mode 100755 index 0c5c02d..0000000 --- a/system_test/run_all.sh +++ /dev/null @@ -1,7 +0,0 @@ -#!/bin/bash - -cp testcase_to_run_all.json testcase_to_run.json - -python -B system_test_runner.py - - http://git-wip-us.apache.org/repos/asf/kafka/blob/d50499a0/system_test/run_all_replica.sh ---------------------------------------------------------------------- diff --git a/system_test/run_all_replica.sh b/system_test/run_all_replica.sh deleted file mode 100755 index b3bce84..0000000 --- a/system_test/run_all_replica.sh +++ /dev/null @@ -1,7 +0,0 @@ -#!/bin/bash - -cp testcase_to_run_all_replica.json testcase_to_run.json - -python -B system_test_runner.py - - http://git-wip-us.apache.org/repos/asf/kafka/blob/d50499a0/system_test/run_sanity.sh ---------------------------------------------------------------------- diff --git a/system_test/run_sanity.sh b/system_test/run_sanity.sh deleted file mode 100755 index a301b96..0000000 --- a/system_test/run_sanity.sh +++ /dev/null @@ -1,7 +0,0 @@ -#!/bin/bash - -cp testcase_to_run_sanity.json testcase_to_run.json - -python -B system_test_runner.py - - http://git-wip-us.apache.org/repos/asf/kafka/blob/d50499a0/system_test/system_test_env.py ---------------------------------------------------------------------- diff --git a/system_test/system_test_env.py b/system_test/system_test_env.py deleted file mode 100644 index c24d3e8..0000000 --- a/system_test/system_test_env.py +++ /dev/null @@ -1,138 +0,0 @@ -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, -# software distributed under the License is distributed on an -# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY -# KIND, either express or implied. See the License for the -# specific language governing permissions and limitations -# under the License. -#!/usr/bin/env python - -# =================================== -# system_test_env.py -# =================================== - -import copy -import json -import os -import sys - -from utils import system_test_utils - -class SystemTestEnv(): - - # private: - _cwdFullPath = os.getcwd() - _thisScriptFullPathName = os.path.realpath(__file__) - _thisScriptBaseDir = os.path.abspath(os.path.join(os.path.dirname(sys.argv[0]))) - - # public: - SYSTEM_TEST_BASE_DIR = os.path.abspath(_thisScriptBaseDir) - SYSTEM_TEST_UTIL_DIR = os.path.abspath(SYSTEM_TEST_BASE_DIR + "/utils") - SYSTEM_TEST_SUITE_SUFFIX = "_testsuite" - SYSTEM_TEST_CASE_PREFIX = "testcase_" - SYSTEM_TEST_MODULE_EXT = ".py" - CLUSTER_CONFIG_FILENAME = "cluster_config.json" - CLUSTER_CONFIG_PATHNAME = os.path.abspath(SYSTEM_TEST_BASE_DIR + "/" + CLUSTER_CONFIG_FILENAME) - METRICS_FILENAME = "metrics.json" - METRICS_PATHNAME = os.path.abspath(SYSTEM_TEST_BASE_DIR + "/" + METRICS_FILENAME) - TESTCASE_TO_RUN_FILENAME = "testcase_to_run.json" - TESTCASE_TO_RUN_PATHNAME = os.path.abspath(SYSTEM_TEST_BASE_DIR + "/" + TESTCASE_TO_RUN_FILENAME) - TESTCASE_TO_SKIP_FILENAME = "testcase_to_skip.json" - TESTCASE_TO_SKIP_PATHNAME = os.path.abspath(SYSTEM_TEST_BASE_DIR + "/" + TESTCASE_TO_SKIP_FILENAME) - - clusterEntityConfigDictList = [] # cluster entity config for current level - clusterEntityConfigDictListInSystemTestLevel = [] # cluster entity config defined in system level - clusterEntityConfigDictListLastFoundInTestSuite = [] # cluster entity config last found in testsuite level - clusterEntityConfigDictListLastFoundInTestCase = [] # cluster entity config last found in testcase level - - systemTestResultsList = [] - testCaseToRunListDict = {} - testCaseToSkipListDict = {} - - printTestDescriptionsOnly = False - doNotValidateRemoteHost = False - - def __init__(self): - "Create an object with this system test session environment" - - # load the system level cluster config - system_test_utils.load_cluster_config(self.CLUSTER_CONFIG_PATHNAME, self.clusterEntityConfigDictList) - - # save the system level cluster config - self.clusterEntityConfigDictListInSystemTestLevel = copy.deepcopy(self.clusterEntityConfigDictList) - - # retrieve testcases to run from testcase_to_run.json - try: - testcaseToRunFileContent = open(self.TESTCASE_TO_RUN_PATHNAME, "r").read() - testcaseToRunData = json.loads(testcaseToRunFileContent) - for testClassName, caseList in testcaseToRunData.items(): - self.testCaseToRunListDict[testClassName] = caseList - except: - pass - - # retrieve testcases to skip from testcase_to_skip.json - try: - testcaseToSkipFileContent = open(self.TESTCASE_TO_SKIP_PATHNAME, "r").read() - testcaseToSkipData = json.loads(testcaseToSkipFileContent) - for testClassName, caseList in testcaseToSkipData.items(): - self.testCaseToSkipListDict[testClassName] = caseList - except: - pass - - def isTestCaseToSkip(self, testClassName, testcaseDirName): - testCaseToRunList = {} - testCaseToSkipList = {} - - try: - testCaseToRunList = self.testCaseToRunListDict[testClassName] - except: - # no 'testClassName' found => no need to run any cases for this test class - return True - - try: - testCaseToSkipList = self.testCaseToSkipListDict[testClassName] - except: - pass - - # if testCaseToRunList has elements, it takes precedence: - if len(testCaseToRunList) > 0: - #print "#### testClassName => ", testClassName - #print "#### testCaseToRunList => ", testCaseToRunList - #print "#### testcaseDirName => ", testcaseDirName - if not testcaseDirName in testCaseToRunList: - #self.log_message("Skipping : " + testcaseDirName) - return True - elif len(testCaseToSkipList) > 0: - #print "#### testClassName => ", testClassName - #print "#### testCaseToSkipList => ", testCaseToSkipList - #print "#### testcaseDirName => ", testcaseDirName - if testcaseDirName in testCaseToSkipList: - #self.log_message("Skipping : " + testcaseDirName) - return True - - return False - - - def getSystemTestEnvDict(self): - envDict = {} - envDict["system_test_base_dir"] = self.SYSTEM_TEST_BASE_DIR - envDict["system_test_util_dir"] = self.SYSTEM_TEST_UTIL_DIR - envDict["cluster_config_pathname"] = self.CLUSTER_CONFIG_PATHNAME - envDict["system_test_suite_suffix"] = self.SYSTEM_TEST_SUITE_SUFFIX - envDict["system_test_case_prefix"] = self.SYSTEM_TEST_CASE_PREFIX - envDict["system_test_module_ext"] = self.SYSTEM_TEST_MODULE_EXT - envDict["cluster_config_pathname"] = self.CLUSTER_CONFIG_PATHNAME - envDict["cluster_entity_config_dict_list"] = self.clusterEntityConfigDictList - envDict["system_test_results_list"] = self.systemTestResultsList - return envDict - - http://git-wip-us.apache.org/repos/asf/kafka/blob/d50499a0/system_test/system_test_runner.py ---------------------------------------------------------------------- diff --git a/system_test/system_test_runner.py b/system_test/system_test_runner.py deleted file mode 100644 index d6251b2..0000000 --- a/system_test/system_test_runner.py +++ /dev/null @@ -1,331 +0,0 @@ -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, -# software distributed under the License is distributed on an -# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY -# KIND, either express or implied. See the License for the -# specific language governing permissions and limitations -# under the License. -#!/usr/bin/evn python - -# ================================================================= -# system_test_runner.py -# -# - This script is the test driver for a distributed environment -# system testing framework. It is located at the top level of the -# framework hierachy (in this case - system_test/). -# -# - This test driver servers as an entry point to launch a series -# of test suites (module) with multiple functionally similar test -# cases which can be grouped together. -# -# - Please refer to system_test/README.txt for more details on -# how to add test suite and test case. -# -# - In most cases, it is not necessary to make any changes to this -# script. -# ================================================================= - -from optparse import OptionParser -from system_test_env import SystemTestEnv -from utils import system_test_utils - -import logging.config -import os -import pprint -import sys - - -# load the config file for logging -logging.config.fileConfig('logging.conf') - -# 'd' is an argument to be merged into the log message (see Python doc for logging). -# In this case, corresponding class name can be appended to the end of the logging -# message to facilitate debugging. -d = {'name_of_class': '(system_test_runner)'} - -class report: - systemTestEnv = None - reportString = "" - reportFileName = "system_test_report.html" - systemTestReport = None - header = """<head> - <title>Kafka System Test Report</title> - <meta charset="utf-8"> - <meta name="viewport" content="width=device-width, initial-scale=1"> - <link rel="stylesheet" href="http://maxcdn.bootstrapcdn.com/bootstrap/3.2.0/css/bootstrap.min.css"> - <script src="https://ajax.googleapis.com/ajax/libs/jquery/1.11.2/jquery.min.js"></script> - <script src="http://maxcdn.bootstrapcdn.com/bootstrap/3.3.2/js/bootstrap.min.js"></script> - </head>""" - footer = """ """ - - def __init__(self, systemTestEnv): - self.totalSkipped = 0 - self.totalPassed = 0 - self.totalTests = 0 - self.totalFailed = 0 - self.systemTestEnv = systemTestEnv - self.systemTestReport = open(self.reportFileName, 'w') - - def __del__(self): - self.systemTestReport.close() - self.systemTestReport = None - - def writeHtmlPage(self, body): - html = """ - <!DOCTYPE html> - <html lang="en"> - """ - html += self.header - html += body - html += self.footer - html += """ - </html> - """ - self.systemTestReport.write(html) - - def wrapIn(self, tag, content): - html = "\n<" + tag + ">" - html += "\n " + content - html += "\n</" + tag.split(" ")[0] + ">" - return html - - def genModal(self, className, caseName, systemTestResult): - key = "validation_status" - id = className + "_" + caseName - info = self.wrapIn("h4", "Validation Status") - for validatedItem in sorted(systemTestResult[key].iterkeys()): - testItemStatus = systemTestResult[key][validatedItem] - info += validatedItem + " : " + testItemStatus - return self.wrapIn("div class=\"modal fade\" id=\"" + id + "\" tabindex=\"-1\" role=\"dialog\" aria-labelledby=\"" + id + "Label\" aria-hidden=\"true\"", - self.wrapIn("div class=\"modal-dialog\"", - self.wrapIn("div class=\"modal-content\"", - self.wrapIn("div class=\"modal-header\"", - self.wrapIn("h4 class=\"modal-title\" id=\"" + id + "Label\"", - className + " - " + caseName)) + - self.wrapIn("div class=\"modal-body\"", - info) + - self.wrapIn("div class=\"modal-footer\"", - self.wrapIn("button type=\"button\" class=\"btn btn-default\" data-dismiss=\"modal\"", "Close"))))) - - def summarize(self): - testItemsTableHeader = self.wrapIn("thead", - self.wrapIn("tr", - self.wrapIn("th", "Test Class Name") + - self.wrapIn("th", "Test Case Name") + - self.wrapIn("th", "Validation Status"))) - testItemsTableBody = "" - modals = "" - - for systemTestResult in self.systemTestEnv.systemTestResultsList: - self.totalTests += 1 - if "_test_class_name" in systemTestResult: - testClassName = systemTestResult["_test_class_name"] - else: - testClassName = "" - - if "_test_case_name" in systemTestResult: - testCaseName = systemTestResult["_test_case_name"] - else: - testCaseName = "" - - if "validation_status" in systemTestResult: - testItemStatus = "SKIPPED" - for key in systemTestResult["validation_status"].iterkeys(): - testItemStatus = systemTestResult["validation_status"][key] - if "FAILED" == testItemStatus: - break; - if "FAILED" == testItemStatus: - self.totalFailed += 1 - validationStatus = self.wrapIn("div class=\"text-danger\" data-toggle=\"modal\" data-target=\"#" + testClassName + "_" + testCaseName + "\"", "FAILED") - modals += self.genModal(testClassName, testCaseName, systemTestResult) - elif "PASSED" == testItemStatus: - self.totalPassed += 1 - validationStatus = self.wrapIn("div class=\"text-success\"", "PASSED") - else: - self.totalSkipped += 1 - validationStatus = self.wrapIn("div class=\"text-warning\"", "SKIPPED") - else: - self.reportString += "|" - - testItemsTableBody += self.wrapIn("tr", - self.wrapIn("td", testClassName) + - self.wrapIn("td", testCaseName) + - self.wrapIn("td", validationStatus)) - - testItemsTableBody = self.wrapIn("tbody", testItemsTableBody) - testItemsTable = self.wrapIn("table class=\"table table-striped\"", testItemsTableHeader + testItemsTableBody) - - statsTblBody = self.wrapIn("tr class=\"active\"", self.wrapIn("td", "Total tests") + self.wrapIn("td", str(self.totalTests))) - statsTblBody += self.wrapIn("tr class=\"success\"", self.wrapIn("td", "Total tests passed") + self.wrapIn("td", str(self.totalPassed))) - statsTblBody += self.wrapIn("tr class=\"danger\"", self.wrapIn("td", "Total tests failed") + self.wrapIn("td", str(self.totalFailed))) - statsTblBody += self.wrapIn("tr class=\"warning\"", self.wrapIn("td", "Total tests skipped") + self.wrapIn("td", str(self.totalSkipped))) - testStatsTable = self.wrapIn("table class=\"table\"", statsTblBody) - - body = self.wrapIn("div class=\"container\"", - self.wrapIn("h2", "Kafka System Test Report") + - self.wrapIn("div class=\"row\"", self.wrapIn("div class=\"col-md-4\"", testStatsTable)) + - self.wrapIn("div class=\"row\"", self.wrapIn("div class=\"col-md-6\"", testItemsTable)) + - modals) - self.writeHtmlPage(self.wrapIn("body", body)) - -def main(): - nLogger = logging.getLogger('namedLogger') - aLogger = logging.getLogger('anonymousLogger') - - optionParser = OptionParser() - optionParser.add_option("-p", "--print-test-descriptions-only", - dest="printTestDescriptionsOnly", - default=False, - action="store_true", - help="print test descriptions only - don't run the test") - - optionParser.add_option("-n", "--do-not-validate-remote-host", - dest="doNotValidateRemoteHost", - default=False, - action="store_true", - help="do not validate remote host (due to different kafka versions are installed)") - - (options, args) = optionParser.parse_args() - - print "\n" - aLogger.info("=================================================") - aLogger.info(" System Regression Test Framework") - aLogger.info("=================================================") - print "\n" - - testSuiteClassDictList = [] - - # SystemTestEnv is a class to provide all environement settings for this session - # such as the SYSTEM_TEST_BASE_DIR, SYSTEM_TEST_UTIL_DIR, ... - systemTestEnv = SystemTestEnv() - - if options.printTestDescriptionsOnly: - systemTestEnv.printTestDescriptionsOnly = True - if options.doNotValidateRemoteHost: - systemTestEnv.doNotValidateRemoteHost = True - - if not systemTestEnv.printTestDescriptionsOnly: - if not systemTestEnv.doNotValidateRemoteHost: - if not system_test_utils.setup_remote_hosts(systemTestEnv): - nLogger.error("Remote hosts sanity check failed. Aborting test ...", extra=d) - print - sys.exit(1) - else: - nLogger.info("SKIPPING : checking remote machines", extra=d) - print - - # get all defined names within a module: - definedItemList = dir(SystemTestEnv) - aLogger.debug("=================================================") - aLogger.debug("SystemTestEnv keys:") - for item in definedItemList: - aLogger.debug(" " + item) - aLogger.debug("=================================================") - - aLogger.info("=================================================") - aLogger.info("looking up test suites ...") - aLogger.info("=================================================") - # find all test suites in SYSTEM_TEST_BASE_DIR - for dirName in os.listdir(systemTestEnv.SYSTEM_TEST_BASE_DIR): - - # make sure this is a valid testsuite directory - if os.path.isdir(dirName) and dirName.endswith(systemTestEnv.SYSTEM_TEST_SUITE_SUFFIX): - print - nLogger.info("found a testsuite : " + dirName, extra=d) - testModulePathName = os.path.abspath(systemTestEnv.SYSTEM_TEST_BASE_DIR + "/" + dirName) - - if not systemTestEnv.printTestDescriptionsOnly: - system_test_utils.setup_remote_hosts_with_testsuite_level_cluster_config(systemTestEnv, testModulePathName) - - # go through all test modules file in this testsuite - for moduleFileName in os.listdir(testModulePathName): - - # make sure it is a valid test module - if moduleFileName.endswith(systemTestEnv.SYSTEM_TEST_MODULE_EXT) \ - and not moduleFileName.startswith("__"): - - # found a test module file - nLogger.info("found a test module file : " + moduleFileName, extra=d) - - testModuleClassName = system_test_utils.sys_call("grep ^class " + testModulePathName + "/" + \ - moduleFileName + " | sed 's/^class //g' | sed 's/(.*):.*//g'") - testModuleClassName = testModuleClassName.rstrip('\n') - - # collect the test suite class data - testSuiteClassDict = {} - testSuiteClassDict["suite"] = dirName - extLenToRemove = systemTestEnv.SYSTEM_TEST_MODULE_EXT.__len__() * -1 - testSuiteClassDict["module"] = moduleFileName[:extLenToRemove] - testSuiteClassDict["class"] = testModuleClassName - testSuiteClassDictList.append(testSuiteClassDict) - - suiteName = testSuiteClassDict["suite"] - moduleName = testSuiteClassDict["module"] - className = testSuiteClassDict["class"] - - # add testsuite directory to sys.path such that the module can be loaded - sys.path.append(systemTestEnv.SYSTEM_TEST_BASE_DIR + "/" + suiteName) - - if not systemTestEnv.printTestDescriptionsOnly: - aLogger.info("=================================================") - aLogger.info("Running Test for : ") - aLogger.info(" suite : " + suiteName) - aLogger.info(" module : " + moduleName) - aLogger.info(" class : " + className) - aLogger.info("=================================================") - - # dynamically loading a module and starting the test class - mod = __import__(moduleName) - theClass = getattr(mod, className) - instance = theClass(systemTestEnv) - instance.runTest() - print - - report(systemTestEnv).summarize() - - if not systemTestEnv.printTestDescriptionsOnly: - totalFailureCount = 0 - print - print "========================================================" - print " TEST REPORTS" - print "========================================================" - for systemTestResult in systemTestEnv.systemTestResultsList: - for key in sorted(systemTestResult.iterkeys()): - if key == "validation_status": - print key, " : " - testItemStatus = None - for validatedItem in sorted(systemTestResult[key].iterkeys()): - testItemStatus = systemTestResult[key][validatedItem] - print " ", validatedItem, " : ", testItemStatus - if "FAILED" == testItemStatus: - totalFailureCount += 1 - else: - print key, " : ", systemTestResult[key] - print - print "========================================================" - print - - print "========================================================" - print "Total failures count : " + str(totalFailureCount) - print "========================================================" - print - return totalFailureCount - - return -1 - -# ========================= -# main entry point -# ========================= - -sys.exit(main()) - - http://git-wip-us.apache.org/repos/asf/kafka/blob/d50499a0/system_test/testcase_to_run_all.json ---------------------------------------------------------------------- diff --git a/system_test/testcase_to_run_all.json b/system_test/testcase_to_run_all.json deleted file mode 100644 index 3e80a1f..0000000 --- a/system_test/testcase_to_run_all.json +++ /dev/null @@ -1,139 +0,0 @@ -{ - "ReplicaBasicTest" : [ - "testcase_0001", - "testcase_0002", - "testcase_0003", - "testcase_0004", - "testcase_0005", - "testcase_0006", - "testcase_0007", - "testcase_0008", - "testcase_0009", - "testcase_0010", - - "testcase_0021", - "testcase_0022", - "testcase_0023", - - "testcase_0101", - "testcase_0102", - "testcase_0103", - "testcase_0104", - "testcase_0105", - "testcase_0106", - "testcase_0107", - "testcase_0108", - "testcase_0109", - "testcase_0110", - - "testcase_10101", - "testcase_10102", - "testcase_10103", - "testcase_10104", - "testcase_10105", - "testcase_10106", - "testcase_10107", - "testcase_10108", - "testcase_10109", - "testcase_10110", - - "testcase_0111", - "testcase_0112", - "testcase_0113", - "testcase_0114", - "testcase_0115", - "testcase_0116", - "testcase_0117", - "testcase_0118", - - "testcase_0121", - "testcase_0122", - "testcase_0123", - "testcase_0124", - "testcase_0125", - "testcase_0126", - "testcase_0127", - - "testcase_0131", - "testcase_0132", - "testcase_0133", - "testcase_0134", - - "testcase_10131", - "testcase_10132", - "testcase_10133", - "testcase_10134", - - "testcase_0151", - "testcase_0152", - "testcase_0153", - "testcase_0154", - "testcase_0155", - "testcase_0156", - "testcase_0157", - "testcase_0158", - - "testcase_0201", - "testcase_0202", - "testcase_0203", - "testcase_0204", - "testcase_0205", - "testcase_0206", - "testcase_0207", - "testcase_0208", - - "testcase_0251", - "testcase_0252", - "testcase_0253", - "testcase_0254", - "testcase_0255", - "testcase_0256", - "testcase_0257", - "testcase_0258", - - "testcase_0301", - "testcase_0302", - "testcase_0303", - "testcase_0304", - "testcase_0305", - "testcase_0306", - "testcase_0307", - "testcase_0308", - - "testcase_4001", - "testcase_4002", - "testcase_4003", - "testcase_4004", - "testcase_4005", - "testcase_4006", - "testcase_4007", - "testcase_4008", - - "testcase_4011", - "testcase_4012", - "testcase_4013", - "testcase_4014", - "testcase_4015", - "testcase_4016", - "testcase_4017", - "testcase_4018", - - "testcase_9051" - ], - - "MirrorMakerTest" : [ - "testcase_5001", - "testcase_5002", - "testcase_5003", - "testcase_5004", - "testcase_5005", - "testcase_5006", - - "testcase_15001", - "testcase_15002", - "testcase_15003", - "testcase_15004", - "testcase_15005", - "testcase_15006" - ] -} http://git-wip-us.apache.org/repos/asf/kafka/blob/d50499a0/system_test/testcase_to_run_all_replica.json ---------------------------------------------------------------------- diff --git a/system_test/testcase_to_run_all_replica.json b/system_test/testcase_to_run_all_replica.json deleted file mode 100644 index 34841f5..0000000 --- a/system_test/testcase_to_run_all_replica.json +++ /dev/null @@ -1,123 +0,0 @@ -{ - "ReplicaBasicTest" : [ - "testcase_0001", - "testcase_0002", - "testcase_0003", - "testcase_0004", - "testcase_0005", - "testcase_0006", - "testcase_0007", - "testcase_0008", - "testcase_0009", - "testcase_0010", - - "testcase_0021", - "testcase_0022", - "testcase_0023", - - "testcase_0101", - "testcase_0102", - "testcase_0103", - "testcase_0104", - "testcase_0105", - "testcase_0106", - "testcase_0107", - "testcase_0108", - "testcase_0109", - "testcase_0110", - - "testcase_10101", - "testcase_10102", - "testcase_10103", - "testcase_10104", - "testcase_10105", - "testcase_10106", - "testcase_10107", - "testcase_10108", - "testcase_10109", - "testcase_10110", - - "testcase_0111", - "testcase_0112", - "testcase_0113", - "testcase_0114", - "testcase_0115", - "testcase_0116", - "testcase_0117", - "testcase_0118", - - "testcase_0121", - "testcase_0122", - "testcase_0123", - "testcase_0124", - "testcase_0125", - "testcase_0126", - "testcase_0127", - - "testcase_0131", - "testcase_0132", - "testcase_0133", - "testcase_0134", - - "testcase_10131", - "testcase_10132", - "testcase_10133", - "testcase_10134", - - "testcase_0151", - "testcase_0152", - "testcase_0153", - "testcase_0154", - "testcase_0155", - "testcase_0156", - "testcase_0157", - "testcase_0158", - - "testcase_0201", - "testcase_0202", - "testcase_0203", - "testcase_0204", - "testcase_0205", - "testcase_0206", - "testcase_0207", - "testcase_0208", - - "testcase_0251", - "testcase_0252", - "testcase_0253", - "testcase_0254", - "testcase_0255", - "testcase_0256", - "testcase_0257", - "testcase_0258", - - "testcase_0301", - "testcase_0302", - "testcase_0303", - "testcase_0304", - "testcase_0305", - "testcase_0306", - "testcase_0307", - "testcase_0308", - - "testcase_4001", - "testcase_4002", - "testcase_4003", - "testcase_4004", - "testcase_4005", - "testcase_4006", - "testcase_4007", - "testcase_4008", - - "testcase_4011", - "testcase_4012", - "testcase_4013", - "testcase_4014", - "testcase_4015", - "testcase_4016", - "testcase_4017", - "testcase_4018", - - "testcase_9051" - ] -} http://git-wip-us.apache.org/repos/asf/kafka/blob/d50499a0/system_test/testcase_to_run_sanity.json ---------------------------------------------------------------------- diff --git a/system_test/testcase_to_run_sanity.json b/system_test/testcase_to_run_sanity.json deleted file mode 100644 index c6cf17e..0000000 --- a/system_test/testcase_to_run_sanity.json +++ /dev/null @@ -1,5 +0,0 @@ -{ - "ReplicaBasicTest" : [ - "testcase_1" - ] -} http://git-wip-us.apache.org/repos/asf/kafka/blob/d50499a0/system_test/testcase_to_skip.json ---------------------------------------------------------------------- diff --git a/system_test/testcase_to_skip.json b/system_test/testcase_to_skip.json deleted file mode 100644 index c24f839..0000000 --- a/system_test/testcase_to_skip.json +++ /dev/null @@ -1,3 +0,0 @@ -{ - "ReplicaBasicTest": [ "testcase_1" ] -} http://git-wip-us.apache.org/repos/asf/kafka/blob/d50499a0/system_test/utils/__init__.py ---------------------------------------------------------------------- diff --git a/system_test/utils/__init__.py b/system_test/utils/__init__.py deleted file mode 100644 index 8d1c8b6..0000000 --- a/system_test/utils/__init__.py +++ /dev/null @@ -1 +0,0 @@ -
