lordgamez commented on code in PR #2059:
URL: https://github.com/apache/nifi-minifi-cpp/pull/2059#discussion_r2661354208


##########
behave_framework/src/minifi_test_framework/containers/container.py:
##########
@@ -254,6 +285,59 @@ def verify_file_contents(self, directory_path: str, 
expected_contents: list[str]
 
         return sorted(actual_file_contents) == sorted(expected_contents)
 
+    def _verify_file_contents_in_stopped_container(self, directory_path: str, 
expected_contents: list[str]) -> bool:
+        if not self.container:
+            return False
+
+        with tempfile.TemporaryDirectory() as temp_dir:
+            extracted_dir = 
self._extract_directory_from_container(directory_path, temp_dir)
+            if not extracted_dir:
+                return False
+
+            actual_file_contents = 
self._read_files_from_directory(extracted_dir)
+            if actual_file_contents is None:
+                return False
+
+            return sorted(actual_file_contents) == sorted(expected_contents)
+
+    def _extract_directory_from_container(self, directory_path: str, temp_dir: 
str) -> str | None:
+        try:
+            bits, _ = self.container.get_archive(directory_path)
+            temp_tar_path = os.path.join(temp_dir, "archive.tar")
+
+            with open(temp_tar_path, 'wb') as f:
+                for chunk in bits:
+                    f.write(chunk)
+
+            with tarfile.open(temp_tar_path) as tar:
+                tar.extractall(path=temp_dir)
+
+            return os.path.join(temp_dir, 
os.path.basename(directory_path.strip('/')))
+        except Exception as e:
+            logging.error(f"Error extracting files from container: {e}")

Review Comment:
   Extended log message in 
https://github.com/apache/nifi-minifi-cpp/pull/2059/commits/595d85e0827fe540ca0011ef28dc9c5bb73e659e



##########
behave_framework/src/minifi_test_framework/containers/container.py:
##########
@@ -85,6 +97,22 @@ def deploy(self) -> bool:
             raise
         return True
 
+    def start(self):
+        if self.container:
+            self.container.start()
+
+    def stop(self):
+        if self.container:
+            self.container.stop()
+
+    def kill(self):
+        if self.container:
+            self.container.kill()
+
+    def restart(self):
+        if self.container:
+            self.container.restart()

Review Comment:
   Added logs in 
https://github.com/apache/nifi-minifi-cpp/pull/2059/commits/595d85e0827fe540ca0011ef28dc9c5bb73e659e



##########
extensions/kafka/tests/features/steps/kafka_server_container.py:
##########
@@ -0,0 +1,126 @@
+# Licensed to the Apache Software Foundation (ASF) under one or more
+# contributor license agreements.  See the NOTICE file distributed with
+# this work for additional information regarding copyright ownership.
+# The ASF licenses this file to You under the Apache License, Version 2.0
+# (the "License"); you may not use this file except in compliance with
+# the License.  You may obtain a copy of the License at
+#
+#     http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+import logging
+import re
+import jks
+
+from OpenSSL import crypto
+from minifi_test_framework.core.helpers import wait_for_condition
+from minifi_test_framework.core.ssl_utils import make_server_cert
+from minifi_test_framework.containers.container import Container
+from minifi_test_framework.containers.file import File
+from minifi_test_framework.core.minifi_test_context import MinifiTestContext
+
+
+class KafkaServer(Container):
+    def __init__(self, test_context: MinifiTestContext):
+        super().__init__("apache/kafka:4.1.0", 
f"kafka-server-{test_context.scenario_id}", test_context.network)
+        self.environment.append("KAFKA_NODE_ID=1")
+        self.environment.append("KAFKA_PROCESS_ROLES=controller,broker")
+        self.environment.append("KAFKA_INTER_BROKER_LISTENER_NAME=PLAINTEXT")
+        self.environment.append("KAFKA_CONTROLLER_LISTENER_NAMES=CONTROLLER")
+        self.environment.append("KAFKA_OFFSETS_TOPIC_REPLICATION_FACTOR=1")
+        
self.environment.append("KAFKA_TRANSACTION_STATE_LOG_REPLICATION_FACTOR=1")
+        self.environment.append("KAFKA_TRANSACTION_STATE_LOG_MIN_ISR=1")
+        
self.environment.append(f"KAFKA_CONTROLLER_QUORUM_VOTERS=1@kafka-server-{test_context.scenario_id}:9096")
+        
self.environment.append(f"KAFKA_LISTENERS=PLAINTEXT://kafka-server-{test_context.scenario_id}:9092,SASL_PLAINTEXT://kafka-server-{test_context.scenario_id}:9094,SSL://kafka-server-{test_context.scenario_id}:9093,SASL_SSL://kafka-server-{test_context.scenario_id}:9095,CONTROLLER://kafka-server-{test_context.scenario_id}:9096")
+        
self.environment.append(f"KAFKA_ADVERTISED_LISTENERS=PLAINTEXT://kafka-server-{test_context.scenario_id}:9092,SASL_PLAINTEXT://kafka-server-{test_context.scenario_id}:9094,SSL://kafka-server-{test_context.scenario_id}:9093,SASL_SSL://kafka-server-{test_context.scenario_id}:9095,CONTROLLER://kafka-server-{test_context.scenario_id}:9096")
+        
self.environment.append("KAFKA_LISTENER_SECURITY_PROTOCOL_MAP=PLAINTEXT:PLAINTEXT,SASL_PLAINTEXT:SASL_PLAINTEXT,SSL:SSL,SASL_SSL:SASL_SSL,CONTROLLER:PLAINTEXT")
+        
self.environment.append("KAFKA_SASL_MECHANISM_INTER_BROKER_PROTOCOL=PLAIN")
+        self.environment.append("KAFKA_SASL_ENABLED_MECHANISMS=PLAIN")
+        
self.environment.append("KAFKA_OPTS=-Djava.security.auth.login.config=/opt/kafka/config/kafka_jaas.conf
 -Dlog4j2.rootLogger.level=DEBUG 
-Dlog4j2.logger.org.apache.kafka.controller.level=DEBUG")
+        self.environment.append("KAFKA_SSL_PROTOCOL=TLS")
+        self.environment.append("KAFKA_SSL_ENABLED_PROTOCOLS=TLSv1.2")
+        self.environment.append("KAFKA_SSL_KEYSTORE_TYPE=JKS")
+        
self.environment.append("KAFKA_SSL_KEYSTORE_FILENAME=kafka.keystore.jks")
+        
self.environment.append("KAFKA_SSL_KEYSTORE_CREDENTIALS=credentials.conf")
+        self.environment.append("KAFKA_SSL_KEY_CREDENTIALS=credentials.conf")
+        
self.environment.append("KAFKA_SSL_TRUSTSTORE_CREDENTIALS=credentials.conf")
+        self.environment.append("KAFKA_SSL_TRUSTSTORE_TYPE=JKS")
+        
self.environment.append("KAFKA_SSL_TRUSTSTORE_FILENAME=kafka.truststore.jks")
+        self.environment.append("KAFKA_SSL_CLIENT_AUTH=none")
+
+        kafka_cert, kafka_key = make_server_cert(self.container_name, 
test_context.root_ca_cert, test_context.root_ca_key)
+
+        pke = jks.PrivateKeyEntry.new('kafka-broker-cert', 
[crypto.dump_certificate(crypto.FILETYPE_ASN1, kafka_cert)], 
crypto.dump_privatekey(crypto.FILETYPE_ASN1, kafka_key), 'rsa_raw')
+        server_keystore = jks.KeyStore.new('jks', [pke])
+        server_keystore_content = server_keystore.saves('abcdefgh')
+        self.files.append(File("/etc/kafka/secrets/kafka.keystore.jks", 
server_keystore_content, permissions=0o644))
+        self.files.append(File("/etc/kafka/secrets/credentials.conf", 
b'abcdefgh', permissions=0o644))
+
+        trusted_cert = jks.TrustedCertEntry.new(
+            'root-ca',  # Alias for the certificate
+            crypto.dump_certificate(crypto.FILETYPE_ASN1, 
test_context.root_ca_cert)
+        )
+
+        # Create a JKS keystore that will serve as a truststore with the 
trusted cert entry.
+        truststore = jks.KeyStore.new('jks', [trusted_cert])
+        truststore_content = truststore.saves('abcdefgh')
+        self.files.append(File("/etc/kafka/secrets/kafka.truststore.jks", 
truststore_content, permissions=0o644))
+
+        jaas_config_file_content = """
+KafkaServer {
+  org.apache.kafka.common.security.plain.PlainLoginModule required
+  username="admin"
+  password="admin-secret"
+  user_admin="admin-secret"
+  user_alice="alice-secret";
+};
+
+Client {
+  org.apache.kafka.common.security.plain.PlainLoginModule required
+  username="admin"
+  password="admin-secret";
+};
+"""
+        self.files.append(File("/opt/kafka/config/kafka_jaas.conf", 
jaas_config_file_content, permissions=0o644))
+
+    def deploy(self):
+        super().deploy()
+        finished_str = "Kafka Server started"
+        return wait_for_condition(
+            condition=lambda: finished_str in self.get_logs(),
+            timeout_seconds=60,
+            bail_condition=lambda: self.exited,
+            context=None)
+
+    def create_topic(self, topic_name: str):
+        (code, output) = self.exec_run(["/bin/bash", "-c", 
f"/opt/kafka/bin/kafka-topics.sh --create --topic {topic_name} 
--bootstrap-server {self.container_name}:9092"])
+        logging.info("Create topic output: %s", output)
+        return code == 0
+
+    def produce_message(self, topic_name: str, message: str):
+        (code, output) = self.exec_run(["/bin/bash", "-c", 
f"/opt/kafka/bin/kafka-console-producer.sh --topic {topic_name} 
--bootstrap-server {self.container_name}:9092 <<< '{message}'"])
+        logging.info("Produce message output: %s", output)
+        return code == 0
+
+    def produce_message_with_key(self, topic_name: str, message: str, 
message_key: str):
+        (code, output) = self.exec_run(["/bin/bash", "-c", 
f"/opt/kafka/bin/kafka-console-producer.sh --property 'key.separator=:' 
--property 'parse.key=true' --topic {topic_name} --bootstrap-server 
{self.container_name}:9092 <<< '{message_key}:{message}'"])
+        logging.info("Produce message with key output: %s", output)
+        return code == 0

Review Comment:
   Added `code` to logs in 
https://github.com/apache/nifi-minifi-cpp/pull/2059/commits/595d85e0827fe540ca0011ef28dc9c5bb73e659e



##########
extensions/kafka/tests/features/consumekafka.feature:
##########
@@ -19,54 +19,55 @@ Feature: Receiving data from using Kafka streaming platform 
using ConsumeKafka
   As a user of MiNiFi
   I need to have ConsumeKafka processor
 
-  Background:
-    Given the content of "/tmp/output" is monitored
-
   Scenario Outline: ConsumeKafka parses and uses kafka topics and topic name 
formats
-    Given a ConsumeKafka processor set up in a "kafka-consumer-flow" flow
+    Given a Kafka server is set up
+    And ConsumeKafka processor is set up to communicate with that server
     And the "Topic Names" property of the ConsumeKafka processor is set to 
"<topic names>"
     And the "Topic Name Format" property of the ConsumeKafka processor is set 
to "<topic name format>"
     And the "Offset Reset" property of the ConsumeKafka processor is set to 
"earliest"
-    And a PutFile processor with the "Directory" property set to "/tmp/output" 
in the "kafka-consumer-flow" flow
+    And a PutFile processor with the "Directory" property set to "/tmp/output"
+    And PutFile is EVENT_DRIVEN
     And the "success" relationship of the ConsumeKafka processor is connected 
to the PutFile
+    And PutFile's success relationship is auto-terminated
 
-    And a kafka broker is set up in correspondence with the third-party kafka 
publisher
-    And the kafka broker is started
+    And the Kafka server is started
     And the topic "ConsumeKafkaTest" is initialized on the kafka broker
 
     When a message with content "<message 1>" is published to the 
"ConsumeKafkaTest" topic
-    And all other processes start up
+    And the MiNiFi instance starts up
     And a message with content "<message 2>" is published to the 
"ConsumeKafkaTest" topic
 
-    Then two flowfiles with the contents "<message 1>" and "<message 2>" are 
placed in the monitored directory in less than 90 seconds
+    Then the contents of "/tmp/output" in less than 30 seconds are: "<message 
1>" and "<message 2>"
 
     Examples: Topic names and formats to test
       | message 1            | message 2           | topic names              
| topic name format |
-      | Ulysses              | James Joyce         | ConsumeKafkaTest         
| (not set)         |
       | The Great Gatsby     | F. Scott Fitzgerald | ConsumeKafkaTest         
| Names             |
       | War and Peace        | Lev Tolstoy         | a,b,c,ConsumeKafkaTest,d 
| Names             |
       | Nineteen Eighty Four | George Orwell       | ConsumeKafkaTest         
| Patterns          |
       | Hamlet               | William Shakespeare | Cons[emu]*KafkaTest      
| Patterns          |
 
   Scenario Outline: ConsumeKafka key attribute is encoded according to the 
"Key Attribute Encoding" property
-    Given a ConsumeKafka processor set up in a "kafka-consumer-flow" flow
+    Given a Kafka server is set up
+    And ConsumeKafka processor is set up to communicate with that server
     And the "Key Attribute Encoding" property of the ConsumeKafka processor is 
set to "<key attribute encoding>"
-    And a RouteOnAttribute processor in the "kafka-consumer-flow" flow
-    And a LogAttribute processor in the "kafka-consumer-flow" flow
-    And a PutFile processor with the "Directory" property set to "/tmp/output" 
in the "kafka-consumer-flow" flow
+    And a RouteOnAttribute processor
+    And RouteOnAttribute is EVENT_DRIVEN
+    And a LogAttribute processor
+    And LogAttribute is EVENT_DRIVEN
+    And a PutFile processor with the "Directory" property set to "/tmp/output"
+    And PutFile is EVENT_DRIVEN
     And the "success" property of the RouteOnAttribute processor is set to 
match <key attribute encoding> encoded kafka message key 
"consume_kafka_test_key"
 
     And the "success" relationship of the ConsumeKafka processor is connected 
to the LogAttribute
     And the "success" relationship of the LogAttribute processor is connected 
to the RouteOnAttribute
     And the "success" relationship of the RouteOnAttribute processor is 
connected to the PutFile
-
-    And a kafka broker is set up in correspondence with the third-party kafka 
publisher
+    And PutFile's success relationship is auto-terminated
 
     When all instances start up
     And a message with content "<message 1>" is published to the 
"ConsumeKafkaTest" topic with key "consume_kafka_test_key"
     And a message with content "<message 2>" is published to the 
"ConsumeKafkaTest" topic with key "consume_kafka_test_key"
 
-    Then two flowfiles with the contents "<message 1>" and "<message 2>" are 
placed in the monitored directory in less than 45 seconds
+    Then the contents of "/tmp/output" in less than 30 seconds are: "<message 
1>" and "<message 2>"

Review Comment:
   Updated in 
https://github.com/apache/nifi-minifi-cpp/pull/2059/commits/595d85e0827fe540ca0011ef28dc9c5bb73e659e



##########
extensions/kafka/tests/features/consumekafka.feature:
##########
@@ -19,54 +19,55 @@ Feature: Receiving data from using Kafka streaming platform 
using ConsumeKafka
   As a user of MiNiFi
   I need to have ConsumeKafka processor
 
-  Background:
-    Given the content of "/tmp/output" is monitored
-
   Scenario Outline: ConsumeKafka parses and uses kafka topics and topic name 
formats
-    Given a ConsumeKafka processor set up in a "kafka-consumer-flow" flow
+    Given a Kafka server is set up
+    And ConsumeKafka processor is set up to communicate with that server
     And the "Topic Names" property of the ConsumeKafka processor is set to 
"<topic names>"
     And the "Topic Name Format" property of the ConsumeKafka processor is set 
to "<topic name format>"
     And the "Offset Reset" property of the ConsumeKafka processor is set to 
"earliest"
-    And a PutFile processor with the "Directory" property set to "/tmp/output" 
in the "kafka-consumer-flow" flow
+    And a PutFile processor with the "Directory" property set to "/tmp/output"
+    And PutFile is EVENT_DRIVEN
     And the "success" relationship of the ConsumeKafka processor is connected 
to the PutFile
+    And PutFile's success relationship is auto-terminated
 
-    And a kafka broker is set up in correspondence with the third-party kafka 
publisher
-    And the kafka broker is started
+    And the Kafka server is started
     And the topic "ConsumeKafkaTest" is initialized on the kafka broker
 
     When a message with content "<message 1>" is published to the 
"ConsumeKafkaTest" topic
-    And all other processes start up
+    And the MiNiFi instance starts up
     And a message with content "<message 2>" is published to the 
"ConsumeKafkaTest" topic
 
-    Then two flowfiles with the contents "<message 1>" and "<message 2>" are 
placed in the monitored directory in less than 90 seconds
+    Then the contents of "/tmp/output" in less than 30 seconds are: "<message 
1>" and "<message 2>"
 
     Examples: Topic names and formats to test
       | message 1            | message 2           | topic names              
| topic name format |
-      | Ulysses              | James Joyce         | ConsumeKafkaTest         
| (not set)         |
       | The Great Gatsby     | F. Scott Fitzgerald | ConsumeKafkaTest         
| Names             |

Review Comment:
   Good catch, it was a mistake, restored in 
https://github.com/apache/nifi-minifi-cpp/pull/2059/commits/595d85e0827fe540ca0011ef28dc9c5bb73e659e



##########
extensions/kafka/tests/features/steps/kafka_server_container.py:
##########
@@ -0,0 +1,126 @@
+# Licensed to the Apache Software Foundation (ASF) under one or more
+# contributor license agreements.  See the NOTICE file distributed with
+# this work for additional information regarding copyright ownership.
+# The ASF licenses this file to You under the Apache License, Version 2.0
+# (the "License"); you may not use this file except in compliance with
+# the License.  You may obtain a copy of the License at
+#
+#     http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+import logging
+import re
+import jks
+
+from OpenSSL import crypto
+from minifi_test_framework.core.helpers import wait_for_condition
+from minifi_test_framework.core.ssl_utils import make_server_cert
+from minifi_test_framework.containers.container import Container
+from minifi_test_framework.containers.file import File
+from minifi_test_framework.core.minifi_test_context import MinifiTestContext
+
+
+class KafkaServer(Container):
+    def __init__(self, test_context: MinifiTestContext):
+        super().__init__("apache/kafka:4.1.0", 
f"kafka-server-{test_context.scenario_id}", test_context.network)
+        self.environment.append("KAFKA_NODE_ID=1")
+        self.environment.append("KAFKA_PROCESS_ROLES=controller,broker")
+        self.environment.append("KAFKA_INTER_BROKER_LISTENER_NAME=PLAINTEXT")
+        self.environment.append("KAFKA_CONTROLLER_LISTENER_NAMES=CONTROLLER")
+        self.environment.append("KAFKA_OFFSETS_TOPIC_REPLICATION_FACTOR=1")
+        
self.environment.append("KAFKA_TRANSACTION_STATE_LOG_REPLICATION_FACTOR=1")
+        self.environment.append("KAFKA_TRANSACTION_STATE_LOG_MIN_ISR=1")
+        
self.environment.append(f"KAFKA_CONTROLLER_QUORUM_VOTERS=1@kafka-server-{test_context.scenario_id}:9096")
+        
self.environment.append(f"KAFKA_LISTENERS=PLAINTEXT://kafka-server-{test_context.scenario_id}:9092,SASL_PLAINTEXT://kafka-server-{test_context.scenario_id}:9094,SSL://kafka-server-{test_context.scenario_id}:9093,SASL_SSL://kafka-server-{test_context.scenario_id}:9095,CONTROLLER://kafka-server-{test_context.scenario_id}:9096")
+        
self.environment.append(f"KAFKA_ADVERTISED_LISTENERS=PLAINTEXT://kafka-server-{test_context.scenario_id}:9092,SASL_PLAINTEXT://kafka-server-{test_context.scenario_id}:9094,SSL://kafka-server-{test_context.scenario_id}:9093,SASL_SSL://kafka-server-{test_context.scenario_id}:9095,CONTROLLER://kafka-server-{test_context.scenario_id}:9096")
+        
self.environment.append("KAFKA_LISTENER_SECURITY_PROTOCOL_MAP=PLAINTEXT:PLAINTEXT,SASL_PLAINTEXT:SASL_PLAINTEXT,SSL:SSL,SASL_SSL:SASL_SSL,CONTROLLER:PLAINTEXT")
+        
self.environment.append("KAFKA_SASL_MECHANISM_INTER_BROKER_PROTOCOL=PLAIN")
+        self.environment.append("KAFKA_SASL_ENABLED_MECHANISMS=PLAIN")
+        
self.environment.append("KAFKA_OPTS=-Djava.security.auth.login.config=/opt/kafka/config/kafka_jaas.conf
 -Dlog4j2.rootLogger.level=DEBUG 
-Dlog4j2.logger.org.apache.kafka.controller.level=DEBUG")
+        self.environment.append("KAFKA_SSL_PROTOCOL=TLS")
+        self.environment.append("KAFKA_SSL_ENABLED_PROTOCOLS=TLSv1.2")
+        self.environment.append("KAFKA_SSL_KEYSTORE_TYPE=JKS")
+        
self.environment.append("KAFKA_SSL_KEYSTORE_FILENAME=kafka.keystore.jks")
+        
self.environment.append("KAFKA_SSL_KEYSTORE_CREDENTIALS=credentials.conf")
+        self.environment.append("KAFKA_SSL_KEY_CREDENTIALS=credentials.conf")
+        
self.environment.append("KAFKA_SSL_TRUSTSTORE_CREDENTIALS=credentials.conf")
+        self.environment.append("KAFKA_SSL_TRUSTSTORE_TYPE=JKS")
+        
self.environment.append("KAFKA_SSL_TRUSTSTORE_FILENAME=kafka.truststore.jks")
+        self.environment.append("KAFKA_SSL_CLIENT_AUTH=none")
+
+        kafka_cert, kafka_key = make_server_cert(self.container_name, 
test_context.root_ca_cert, test_context.root_ca_key)
+
+        pke = jks.PrivateKeyEntry.new('kafka-broker-cert', 
[crypto.dump_certificate(crypto.FILETYPE_ASN1, kafka_cert)], 
crypto.dump_privatekey(crypto.FILETYPE_ASN1, kafka_key), 'rsa_raw')
+        server_keystore = jks.KeyStore.new('jks', [pke])
+        server_keystore_content = server_keystore.saves('abcdefgh')
+        self.files.append(File("/etc/kafka/secrets/kafka.keystore.jks", 
server_keystore_content, permissions=0o644))
+        self.files.append(File("/etc/kafka/secrets/credentials.conf", 
b'abcdefgh', permissions=0o644))
+
+        trusted_cert = jks.TrustedCertEntry.new(
+            'root-ca',  # Alias for the certificate
+            crypto.dump_certificate(crypto.FILETYPE_ASN1, 
test_context.root_ca_cert)
+        )
+
+        # Create a JKS keystore that will serve as a truststore with the 
trusted cert entry.
+        truststore = jks.KeyStore.new('jks', [trusted_cert])
+        truststore_content = truststore.saves('abcdefgh')
+        self.files.append(File("/etc/kafka/secrets/kafka.truststore.jks", 
truststore_content, permissions=0o644))
+
+        jaas_config_file_content = """
+KafkaServer {
+  org.apache.kafka.common.security.plain.PlainLoginModule required
+  username="admin"
+  password="admin-secret"
+  user_admin="admin-secret"
+  user_alice="alice-secret";
+};
+
+Client {
+  org.apache.kafka.common.security.plain.PlainLoginModule required
+  username="admin"
+  password="admin-secret";
+};
+"""
+        self.files.append(File("/opt/kafka/config/kafka_jaas.conf", 
jaas_config_file_content, permissions=0o644))
+
+    def deploy(self):
+        super().deploy()
+        finished_str = "Kafka Server started"
+        return wait_for_condition(
+            condition=lambda: finished_str in self.get_logs(),
+            timeout_seconds=60,
+            bail_condition=lambda: self.exited,
+            context=None)
+
+    def create_topic(self, topic_name: str):
+        (code, output) = self.exec_run(["/bin/bash", "-c", 
f"/opt/kafka/bin/kafka-topics.sh --create --topic {topic_name} 
--bootstrap-server {self.container_name}:9092"])
+        logging.info("Create topic output: %s", output)
+        return code == 0
+
+    def produce_message(self, topic_name: str, message: str):
+        (code, output) = self.exec_run(["/bin/bash", "-c", 
f"/opt/kafka/bin/kafka-console-producer.sh --topic {topic_name} 
--bootstrap-server {self.container_name}:9092 <<< '{message}'"])

Review Comment:
   Updated in 
https://github.com/apache/nifi-minifi-cpp/pull/2059/commits/595d85e0827fe540ca0011ef28dc9c5bb73e659e



##########
extensions/kafka/tests/features/steps/kafka_server_container.py:
##########
@@ -0,0 +1,126 @@
+# Licensed to the Apache Software Foundation (ASF) under one or more
+# contributor license agreements.  See the NOTICE file distributed with
+# this work for additional information regarding copyright ownership.
+# The ASF licenses this file to You under the Apache License, Version 2.0
+# (the "License"); you may not use this file except in compliance with
+# the License.  You may obtain a copy of the License at
+#
+#     http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+import logging
+import re
+import jks
+
+from OpenSSL import crypto
+from minifi_test_framework.core.helpers import wait_for_condition
+from minifi_test_framework.core.ssl_utils import make_server_cert
+from minifi_test_framework.containers.container import Container
+from minifi_test_framework.containers.file import File
+from minifi_test_framework.core.minifi_test_context import MinifiTestContext
+
+
+class KafkaServer(Container):
+    def __init__(self, test_context: MinifiTestContext):
+        super().__init__("apache/kafka:4.1.0", 
f"kafka-server-{test_context.scenario_id}", test_context.network)
+        self.environment.append("KAFKA_NODE_ID=1")
+        self.environment.append("KAFKA_PROCESS_ROLES=controller,broker")
+        self.environment.append("KAFKA_INTER_BROKER_LISTENER_NAME=PLAINTEXT")
+        self.environment.append("KAFKA_CONTROLLER_LISTENER_NAMES=CONTROLLER")
+        self.environment.append("KAFKA_OFFSETS_TOPIC_REPLICATION_FACTOR=1")
+        
self.environment.append("KAFKA_TRANSACTION_STATE_LOG_REPLICATION_FACTOR=1")
+        self.environment.append("KAFKA_TRANSACTION_STATE_LOG_MIN_ISR=1")
+        
self.environment.append(f"KAFKA_CONTROLLER_QUORUM_VOTERS=1@kafka-server-{test_context.scenario_id}:9096")
+        
self.environment.append(f"KAFKA_LISTENERS=PLAINTEXT://kafka-server-{test_context.scenario_id}:9092,SASL_PLAINTEXT://kafka-server-{test_context.scenario_id}:9094,SSL://kafka-server-{test_context.scenario_id}:9093,SASL_SSL://kafka-server-{test_context.scenario_id}:9095,CONTROLLER://kafka-server-{test_context.scenario_id}:9096")
+        
self.environment.append(f"KAFKA_ADVERTISED_LISTENERS=PLAINTEXT://kafka-server-{test_context.scenario_id}:9092,SASL_PLAINTEXT://kafka-server-{test_context.scenario_id}:9094,SSL://kafka-server-{test_context.scenario_id}:9093,SASL_SSL://kafka-server-{test_context.scenario_id}:9095,CONTROLLER://kafka-server-{test_context.scenario_id}:9096")
+        
self.environment.append("KAFKA_LISTENER_SECURITY_PROTOCOL_MAP=PLAINTEXT:PLAINTEXT,SASL_PLAINTEXT:SASL_PLAINTEXT,SSL:SSL,SASL_SSL:SASL_SSL,CONTROLLER:PLAINTEXT")
+        
self.environment.append("KAFKA_SASL_MECHANISM_INTER_BROKER_PROTOCOL=PLAIN")
+        self.environment.append("KAFKA_SASL_ENABLED_MECHANISMS=PLAIN")
+        
self.environment.append("KAFKA_OPTS=-Djava.security.auth.login.config=/opt/kafka/config/kafka_jaas.conf
 -Dlog4j2.rootLogger.level=DEBUG 
-Dlog4j2.logger.org.apache.kafka.controller.level=DEBUG")
+        self.environment.append("KAFKA_SSL_PROTOCOL=TLS")
+        self.environment.append("KAFKA_SSL_ENABLED_PROTOCOLS=TLSv1.2")
+        self.environment.append("KAFKA_SSL_KEYSTORE_TYPE=JKS")
+        
self.environment.append("KAFKA_SSL_KEYSTORE_FILENAME=kafka.keystore.jks")
+        
self.environment.append("KAFKA_SSL_KEYSTORE_CREDENTIALS=credentials.conf")
+        self.environment.append("KAFKA_SSL_KEY_CREDENTIALS=credentials.conf")
+        
self.environment.append("KAFKA_SSL_TRUSTSTORE_CREDENTIALS=credentials.conf")
+        self.environment.append("KAFKA_SSL_TRUSTSTORE_TYPE=JKS")
+        
self.environment.append("KAFKA_SSL_TRUSTSTORE_FILENAME=kafka.truststore.jks")
+        self.environment.append("KAFKA_SSL_CLIENT_AUTH=none")
+
+        kafka_cert, kafka_key = make_server_cert(self.container_name, 
test_context.root_ca_cert, test_context.root_ca_key)
+
+        pke = jks.PrivateKeyEntry.new('kafka-broker-cert', 
[crypto.dump_certificate(crypto.FILETYPE_ASN1, kafka_cert)], 
crypto.dump_privatekey(crypto.FILETYPE_ASN1, kafka_key), 'rsa_raw')
+        server_keystore = jks.KeyStore.new('jks', [pke])
+        server_keystore_content = server_keystore.saves('abcdefgh')
+        self.files.append(File("/etc/kafka/secrets/kafka.keystore.jks", 
server_keystore_content, permissions=0o644))
+        self.files.append(File("/etc/kafka/secrets/credentials.conf", 
b'abcdefgh', permissions=0o644))
+
+        trusted_cert = jks.TrustedCertEntry.new(
+            'root-ca',  # Alias for the certificate
+            crypto.dump_certificate(crypto.FILETYPE_ASN1, 
test_context.root_ca_cert)
+        )
+
+        # Create a JKS keystore that will serve as a truststore with the 
trusted cert entry.
+        truststore = jks.KeyStore.new('jks', [trusted_cert])
+        truststore_content = truststore.saves('abcdefgh')
+        self.files.append(File("/etc/kafka/secrets/kafka.truststore.jks", 
truststore_content, permissions=0o644))
+
+        jaas_config_file_content = """
+KafkaServer {
+  org.apache.kafka.common.security.plain.PlainLoginModule required
+  username="admin"
+  password="admin-secret"
+  user_admin="admin-secret"
+  user_alice="alice-secret";
+};
+
+Client {
+  org.apache.kafka.common.security.plain.PlainLoginModule required
+  username="admin"
+  password="admin-secret";
+};
+"""
+        self.files.append(File("/opt/kafka/config/kafka_jaas.conf", 
jaas_config_file_content, permissions=0o644))
+
+    def deploy(self):
+        super().deploy()
+        finished_str = "Kafka Server started"
+        return wait_for_condition(
+            condition=lambda: finished_str in self.get_logs(),
+            timeout_seconds=60,
+            bail_condition=lambda: self.exited,
+            context=None)
+
+    def create_topic(self, topic_name: str):
+        (code, output) = self.exec_run(["/bin/bash", "-c", 
f"/opt/kafka/bin/kafka-topics.sh --create --topic {topic_name} 
--bootstrap-server {self.container_name}:9092"])
+        logging.info("Create topic output: %s", output)
+        return code == 0
+
+    def produce_message(self, topic_name: str, message: str):
+        (code, output) = self.exec_run(["/bin/bash", "-c", 
f"/opt/kafka/bin/kafka-console-producer.sh --topic {topic_name} 
--bootstrap-server {self.container_name}:9092 <<< '{message}'"])
+        logging.info("Produce message output: %s", output)
+        return code == 0
+
+    def produce_message_with_key(self, topic_name: str, message: str, 
message_key: str):
+        (code, output) = self.exec_run(["/bin/bash", "-c", 
f"/opt/kafka/bin/kafka-console-producer.sh --property 'key.separator=:' 
--property 'parse.key=true' --topic {topic_name} --bootstrap-server 
{self.container_name}:9092 <<< '{message_key}:{message}'"])

Review Comment:
   Updated in 
https://github.com/apache/nifi-minifi-cpp/pull/2059/commits/595d85e0827fe540ca0011ef28dc9c5bb73e659e



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to