szaszm commented on code in PR #2059:
URL: https://github.com/apache/nifi-minifi-cpp/pull/2059#discussion_r2585016416


##########
extensions/kafka/tests/features/steps/kafka_server_container.py:
##########
@@ -0,0 +1,126 @@
+# Licensed to the Apache Software Foundation (ASF) under one or more
+# contributor license agreements.  See the NOTICE file distributed with
+# this work for additional information regarding copyright ownership.
+# The ASF licenses this file to You under the Apache License, Version 2.0
+# (the "License"); you may not use this file except in compliance with
+# the License.  You may obtain a copy of the License at
+#
+#     http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+import logging
+import re
+import jks
+
+from OpenSSL import crypto
+from minifi_test_framework.core.helpers import wait_for_condition
+from minifi_test_framework.core.ssl_utils import make_server_cert
+from minifi_test_framework.containers.container import Container
+from minifi_test_framework.containers.file import File
+from minifi_test_framework.core.minifi_test_context import MinifiTestContext
+
+
+class KafkaServer(Container):
+    def __init__(self, test_context: MinifiTestContext):
+        super().__init__("apache/kafka:4.1.0", 
f"kafka-server-{test_context.scenario_id}", test_context.network)
+        self.environment.append("KAFKA_NODE_ID=1")
+        self.environment.append("KAFKA_PROCESS_ROLES=controller,broker")
+        self.environment.append("KAFKA_INTER_BROKER_LISTENER_NAME=PLAINTEXT")
+        self.environment.append("KAFKA_CONTROLLER_LISTENER_NAMES=CONTROLLER")
+        self.environment.append("KAFKA_OFFSETS_TOPIC_REPLICATION_FACTOR=1")
+        
self.environment.append("KAFKA_TRANSACTION_STATE_LOG_REPLICATION_FACTOR=1")
+        self.environment.append("KAFKA_TRANSACTION_STATE_LOG_MIN_ISR=1")
+        
self.environment.append(f"KAFKA_CONTROLLER_QUORUM_VOTERS=1@kafka-server-{test_context.scenario_id}:9096")
+        
self.environment.append(f"KAFKA_LISTENERS=PLAINTEXT://kafka-server-{test_context.scenario_id}:9092,SASL_PLAINTEXT://kafka-server-{test_context.scenario_id}:9094,SSL://kafka-server-{test_context.scenario_id}:9093,SASL_SSL://kafka-server-{test_context.scenario_id}:9095,CONTROLLER://kafka-server-{test_context.scenario_id}:9096")
+        
self.environment.append(f"KAFKA_ADVERTISED_LISTENERS=PLAINTEXT://kafka-server-{test_context.scenario_id}:9092,SASL_PLAINTEXT://kafka-server-{test_context.scenario_id}:9094,SSL://kafka-server-{test_context.scenario_id}:9093,SASL_SSL://kafka-server-{test_context.scenario_id}:9095,CONTROLLER://kafka-server-{test_context.scenario_id}:9096")
+        
self.environment.append("KAFKA_LISTENER_SECURITY_PROTOCOL_MAP=PLAINTEXT:PLAINTEXT,SASL_PLAINTEXT:SASL_PLAINTEXT,SSL:SSL,SASL_SSL:SASL_SSL,CONTROLLER:PLAINTEXT")
+        
self.environment.append("KAFKA_SASL_MECHANISM_INTER_BROKER_PROTOCOL=PLAIN")
+        self.environment.append("KAFKA_SASL_ENABLED_MECHANISMS=PLAIN")
+        
self.environment.append("KAFKA_OPTS=-Djava.security.auth.login.config=/opt/kafka/config/kafka_jaas.conf
 -Dlog4j2.rootLogger.level=DEBUG 
-Dlog4j2.logger.org.apache.kafka.controller.level=DEBUG")
+        self.environment.append("KAFKA_SSL_PROTOCOL=TLS")
+        self.environment.append("KAFKA_SSL_ENABLED_PROTOCOLS=TLSv1.2")
+        self.environment.append("KAFKA_SSL_KEYSTORE_TYPE=JKS")
+        
self.environment.append("KAFKA_SSL_KEYSTORE_FILENAME=kafka.keystore.jks")
+        
self.environment.append("KAFKA_SSL_KEYSTORE_CREDENTIALS=credentials.conf")
+        self.environment.append("KAFKA_SSL_KEY_CREDENTIALS=credentials.conf")
+        
self.environment.append("KAFKA_SSL_TRUSTSTORE_CREDENTIALS=credentials.conf")
+        self.environment.append("KAFKA_SSL_TRUSTSTORE_TYPE=JKS")
+        
self.environment.append("KAFKA_SSL_TRUSTSTORE_FILENAME=kafka.truststore.jks")
+        self.environment.append("KAFKA_SSL_CLIENT_AUTH=none")
+
+        kafka_cert, kafka_key = make_server_cert(self.container_name, 
test_context.root_ca_cert, test_context.root_ca_key)
+
+        pke = jks.PrivateKeyEntry.new('kafka-broker-cert', 
[crypto.dump_certificate(crypto.FILETYPE_ASN1, kafka_cert)], 
crypto.dump_privatekey(crypto.FILETYPE_ASN1, kafka_key), 'rsa_raw')
+        server_keystore = jks.KeyStore.new('jks', [pke])
+        server_keystore_content = server_keystore.saves('abcdefgh')
+        self.files.append(File("/etc/kafka/secrets/kafka.keystore.jks", 
server_keystore_content, permissions=0o644))
+        self.files.append(File("/etc/kafka/secrets/credentials.conf", 
b'abcdefgh', permissions=0o644))
+
+        trusted_cert = jks.TrustedCertEntry.new(
+            'root-ca',  # Alias for the certificate
+            crypto.dump_certificate(crypto.FILETYPE_ASN1, 
test_context.root_ca_cert)
+        )
+
+        # Create a JKS keystore that will serve as a truststore with the 
trusted cert entry.
+        truststore = jks.KeyStore.new('jks', [trusted_cert])
+        truststore_content = truststore.saves('abcdefgh')
+        self.files.append(File("/etc/kafka/secrets/kafka.truststore.jks", 
truststore_content, permissions=0o644))
+
+        jaas_config_file_content = """
+KafkaServer {
+  org.apache.kafka.common.security.plain.PlainLoginModule required
+  username="admin"
+  password="admin-secret"
+  user_admin="admin-secret"
+  user_alice="alice-secret";
+};
+
+Client {
+  org.apache.kafka.common.security.plain.PlainLoginModule required
+  username="admin"
+  password="admin-secret";
+};
+"""
+        self.files.append(File("/opt/kafka/config/kafka_jaas.conf", 
jaas_config_file_content, permissions=0o644))
+
+    def deploy(self):
+        super().deploy()
+        finished_str = "Kafka Server started"
+        return wait_for_condition(
+            condition=lambda: finished_str in self.get_logs(),
+            timeout_seconds=60,
+            bail_condition=lambda: self.exited,
+            context=None)
+
+    def create_topic(self, topic_name: str):
+        (code, output) = self.exec_run(["/bin/bash", "-c", 
f"/opt/kafka/bin/kafka-topics.sh --create --topic {topic_name} 
--bootstrap-server {self.container_name}:9092"])

Review Comment:
   optional: sh + quotes to handle more special characters
   ```suggestion
           (code, output) = self.exec_run(["/bin/sh", "-c", 
f"/opt/kafka/bin/kafka-topics.sh --create --topic '{topic_name}' 
--bootstrap-server '{self.container_name}':9092"])
   ```



##########
extensions/kafka/tests/features/consumekafka.feature:
##########
@@ -75,18 +76,19 @@ Feature: Receiving data from using Kafka streaming platform 
using ConsumeKafka
       | Crime and Punishment | Фёдор Михайлович Достоевский  | Hex             
       |
 
   Scenario Outline: ConsumeKafka transactional behaviour is supported
-    Given a ConsumeKafka processor set up in a "kafka-consumer-flow" flow
+    Given a Kafka server is set up
+    And ConsumeKafka processor is set up to communicate with that server
     And the "Topic Names" property of the ConsumeKafka processor is set to 
"ConsumeKafkaTest"
     And the "Honor Transactions" property of the ConsumeKafka processor is set 
to "<honor transactions>"
-    And a PutFile processor with the "Directory" property set to "/tmp/output" 
in the "kafka-consumer-flow" flow
+    And a PutFile processor with the "Directory" property set to "/tmp/output"
+    And PutFile is EVENT_DRIVEN
     And the "success" relationship of the ConsumeKafka processor is connected 
to the PutFile
-
-    And a kafka broker is set up in correspondence with the third-party kafka 
publisher
+    And PutFile's success relationship is auto-terminated
 
     When all instances start up
     And the publisher performs a <transaction type> transaction publishing to 
the "ConsumeKafkaTest" topic these messages: <messages sent>
 
-    Then <number of flowfiles expected> flowfiles are placed in the monitored 
directory in less than 15 seconds
+    Then there are <number of flowfiles expected> files in the "/tmp/output" 
directory in less than 15 seconds

Review Comment:
   @fgerlits comment applies to this too: it's clearer to expect an event than 
a state, so the old version was better.



##########
extensions/kafka/tests/features/consumekafka.feature:
##########
@@ -19,54 +19,55 @@ Feature: Receiving data from using Kafka streaming platform 
using ConsumeKafka
   As a user of MiNiFi
   I need to have ConsumeKafka processor
 
-  Background:
-    Given the content of "/tmp/output" is monitored
-
   Scenario Outline: ConsumeKafka parses and uses kafka topics and topic name 
formats
-    Given a ConsumeKafka processor set up in a "kafka-consumer-flow" flow
+    Given a Kafka server is set up
+    And ConsumeKafka processor is set up to communicate with that server
     And the "Topic Names" property of the ConsumeKafka processor is set to 
"<topic names>"
     And the "Topic Name Format" property of the ConsumeKafka processor is set 
to "<topic name format>"
     And the "Offset Reset" property of the ConsumeKafka processor is set to 
"earliest"
-    And a PutFile processor with the "Directory" property set to "/tmp/output" 
in the "kafka-consumer-flow" flow
+    And a PutFile processor with the "Directory" property set to "/tmp/output"
+    And PutFile is EVENT_DRIVEN
     And the "success" relationship of the ConsumeKafka processor is connected 
to the PutFile
+    And PutFile's success relationship is auto-terminated
 
-    And a kafka broker is set up in correspondence with the third-party kafka 
publisher
-    And the kafka broker is started
+    And the Kafka server is started
     And the topic "ConsumeKafkaTest" is initialized on the kafka broker
 
     When a message with content "<message 1>" is published to the 
"ConsumeKafkaTest" topic
-    And all other processes start up
+    And the MiNiFi instance starts up
     And a message with content "<message 2>" is published to the 
"ConsumeKafkaTest" topic
 
-    Then two flowfiles with the contents "<message 1>" and "<message 2>" are 
placed in the monitored directory in less than 90 seconds
+    Then the contents of "/tmp/output" in less than 30 seconds are: "<message 
1>" and "<message 2>"
 
     Examples: Topic names and formats to test
       | message 1            | message 2           | topic names              
| topic name format |
-      | Ulysses              | James Joyce         | ConsumeKafkaTest         
| (not set)         |
       | The Great Gatsby     | F. Scott Fitzgerald | ConsumeKafkaTest         
| Names             |
       | War and Peace        | Lev Tolstoy         | a,b,c,ConsumeKafkaTest,d 
| Names             |
       | Nineteen Eighty Four | George Orwell       | ConsumeKafkaTest         
| Patterns          |
       | Hamlet               | William Shakespeare | Cons[emu]*KafkaTest      
| Patterns          |
 
   Scenario Outline: ConsumeKafka key attribute is encoded according to the 
"Key Attribute Encoding" property
-    Given a ConsumeKafka processor set up in a "kafka-consumer-flow" flow
+    Given a Kafka server is set up
+    And ConsumeKafka processor is set up to communicate with that server
     And the "Key Attribute Encoding" property of the ConsumeKafka processor is 
set to "<key attribute encoding>"
-    And a RouteOnAttribute processor in the "kafka-consumer-flow" flow
-    And a LogAttribute processor in the "kafka-consumer-flow" flow
-    And a PutFile processor with the "Directory" property set to "/tmp/output" 
in the "kafka-consumer-flow" flow
+    And a RouteOnAttribute processor
+    And RouteOnAttribute is EVENT_DRIVEN
+    And a LogAttribute processor
+    And LogAttribute is EVENT_DRIVEN
+    And a PutFile processor with the "Directory" property set to "/tmp/output"
+    And PutFile is EVENT_DRIVEN

Review Comment:
   Shouldn't it be default in the test framework to have non-source processors 
use event driven scheduling? to reduce boilerplate



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to