This is an automated email from the ASF dual-hosted git repository.

manikumar pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/kafka-site.git


The following commit(s) were added to refs/heads/asf-site by this push:
     new c6c253e  MINOR: remove the indent in security doc (#330)
c6c253e is described below

commit c6c253e6bd0dcf0e56d4266f4b7679bd241836ce
Author: Luke Chen <[email protected]>
AuthorDate: Mon Feb 8 20:48:35 2021 +0800

    MINOR: remove the indent in security doc (#330)
---
 27/security.html | 466 +++++++++++++++++++++++++++----------------------------
 1 file changed, 233 insertions(+), 233 deletions(-)

diff --git a/27/security.html b/27/security.html
index ed88776..c4d7088 100644
--- a/27/security.html
+++ b/27/security.html
@@ -47,7 +47,7 @@
             The tool supports two different keystore formats, the Java 
specific jks format which has been deprecated by now, as well as PKCS12.
             PKCS12 is the default format as of Java version 9, to ensure this 
format is being used regardless of the Java version in use all following
             commands explicitly specify the PKCS12 format.
-            <pre class="line-numbers"><code class="language-bash">             
   keytool -keystore {keystorefile} -alias localhost -validity {validity} 
-genkey -keyalg RSA -storetype pkcs12</code></pre>
+            <pre class="line-numbers"><code class="language-bash">keytool 
-keystore {keystorefile} -alias localhost -validity {validity} -genkey -keyalg 
RSA -storetype pkcs12</code></pre>
             You need to specify two parameters in the above command:
             <ol>
                 <li>keystorefile: the keystore file that stores the keys (and 
later the certificate) for this broker. The keystore file contains the private
@@ -63,7 +63,7 @@
             authentication purposes.<br>
             To generate certificate signing requests run the following command 
for all server keystores created so far.
 
-            <pre class="line-numbers"><code class="language-bash">             
   keytool -keystore server.keystore.jks -alias localhost -validity {validity} 
-genkey -keyalg RSA -destkeystoretype pkcs12 -ext 
SAN=DNS:{FQDN},IP:{IPADDRESS1}</code></pre>
+            <pre class="line-numbers"><code class="language-bash">keytool 
-keystore server.keystore.jks -alias localhost -validity {validity} -genkey 
-keyalg RSA -destkeystoretype pkcs12 -ext 
SAN=DNS:{FQDN},IP:{IPADDRESS1}</code></pre>
             This command assumes that you want to add hostname information to 
the certificate, if this is not the case, you can omit the extension parameter 
<code>-ext SAN=DNS:{FQDN},IP:{IPADDRESS1}</code>. Please see below for more 
information on this.
 
             <h5>Host Name Verification</h5>
@@ -76,7 +76,7 @@
             Server host name verification may be disabled by setting 
<code>ssl.endpoint.identification.algorithm</code> to an empty string.<br>
             For dynamically configured broker listeners, hostname verification 
may be disabled using <code>kafka-configs.sh</code>:<br>
 
-            <pre class="line-numbers"><code class="language-text">             
   bin/kafka-configs.sh --bootstrap-server localhost:9093 --entity-type brokers 
--entity-name 0 --alter --add-config 
"listener.name.internal.ssl.endpoint.identification.algorithm="</code></pre>
+            <pre class="line-numbers"><code 
class="language-text">bin/kafka-configs.sh --bootstrap-server localhost:9093 
--entity-type brokers --entity-name 0 --alter --add-config 
"listener.name.internal.ssl.endpoint.identification.algorithm="</code></pre>
 
             <p><b>Note:</b></p>
             Normally there is no good reason to disable hostname verification 
apart from being the quickest way to "just get it to work" followed
@@ -99,7 +99,7 @@
 
 
             To add a SAN field append the following argument <code> -ext 
SAN=DNS:{FQDN},IP:{IPADDRESS} </code> to the keytool command:
-            <pre class="line-numbers"><code class="language-bash">             
   keytool -keystore server.keystore.jks -alias localhost -validity {validity} 
-genkey -keyalg RSA -destkeystoretype pkcs12 -ext 
SAN=DNS:{FQDN},IP:{IPADDRESS1}</code></pre>
+            <pre class="line-numbers"><code class="language-bash">keytool 
-keystore server.keystore.jks -alias localhost -validity {validity} -genkey 
-keyalg RSA -destkeystoretype pkcs12 -ext 
SAN=DNS:{FQDN},IP:{IPADDRESS1}</code></pre>
         </li>
 
         <li><h4 class="anchor-heading"><a id="security_ssl_ca" 
class="anchor-link"></a><a href="#security_ssl_ca">Creating your own CA</a></h4>
@@ -208,25 +208,25 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
             Then create a database and serial number file, these will be used 
to keep track of which certificates were signed with this CA. Both of
             these are simply text files that reside in the same directory as 
your CA keys.
 
-            <pre class="line-numbers"><code class="language-bash">             
   echo 01 > serial.txt
-                touch index.txt</code></pre>
+            <pre class="line-numbers"><code class="language-bash">echo 01 > 
serial.txt
+touch index.txt</code></pre>
 
             With these steps done you are now ready to generate your CA that 
will be used to sign certificates later.
 
-            <pre class="line-numbers"><code class="language-bash">            
openssl req -x509 -config openssl-ca.cnf -newkey rsa:4096 -sha256 -nodes -out 
cacert.pem -outform PEM</code></pre>
+            <pre class="line-numbers"><code class="language-bash">openssl req 
-x509 -config openssl-ca.cnf -newkey rsa:4096 -sha256 -nodes -out cacert.pem 
-outform PEM</code></pre>
 
             The CA is simply a public/private key pair and certificate that is 
signed by itself, and is only intended to sign other certificates.<br>
             This keypair should be kept very safe, if someone gains access to 
it, they can create and sign certificates that will be trusted by your
             infrastructure, which means they will be able to impersonate 
anybody when connecting to any service that trusts this CA.<br>
 
             The next step is to add the generated CA to the **clients' 
truststore** so that the clients can trust this CA:
-            <pre class="line-numbers"><code class="language-bash">             
   keytool -keystore client.truststore.jks -alias CARoot -import -file 
ca-cert</code></pre>
+            <pre class="line-numbers"><code class="language-bash">keytool 
-keystore client.truststore.jks -alias CARoot -import -file ca-cert</code></pre>
 
             <b>Note:</b>
             If you configure the Kafka brokers to require client 
authentication by setting ssl.client.auth to be "requested" or "required" in the
             <a href="#brokerconfigs">Kafka brokers config</a> then you must 
provide a truststore for the Kafka brokers as well and it should have
             all the CA certificates that clients' keys were signed by.
-            <pre class="line-numbers"><code class="language-bash">             
   keytool -keystore server.truststore.jks -alias CARoot -import -file 
ca-cert</code></pre>
+            <pre class="line-numbers"><code class="language-bash">keytool 
-keystore server.truststore.jks -alias CARoot -import -file ca-cert</code></pre>
 
             In contrast to the keystore in step 1 that stores each machine's 
own identity, the truststore of a client stores all the certificates
             that the client should trust. Importing a certificate into one's 
truststore also means trusting all certificates that are signed by that
@@ -237,11 +237,11 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
         </li>
         <li><h4 class="anchor-heading"><a id="security_ssl_signing" 
class="anchor-link"></a><a href="#security_ssl_signing">Signing the 
certificate</a></h4>
             Then sign it with the CA:
-            <pre class="line-numbers"><code class="language-bash">             
   openssl ca -config openssl-ca.cnf -policy signing_policy -extensions 
signing_req -out {server certificate} -infiles {certificate signing 
request}</code></pre>
+            <pre class="line-numbers"><code class="language-bash">openssl ca 
-config openssl-ca.cnf -policy signing_policy -extensions signing_req -out 
{server certificate} -infiles {certificate signing request}</code></pre>
 
             Finally, you need to import both the certificate of the CA and the 
signed certificate into the keystore:
-            <pre class="line-numbers"><code class="language-bash">             
   keytool -keystore {keystore} -alias CARoot -import -file {CA certificate}
-                keytool -keystore {keystore} -alias localhost -import -file 
cert-signed</code></pre>
+            <pre class="line-numbers"><code class="language-bash">keytool 
-keystore {keystore} -alias CARoot -import -file {CA certificate}
+keytool -keystore {keystore} -alias localhost -import -file 
cert-signed</code></pre>
 
             The definitions of the parameters are the following:
             <ol>
@@ -310,7 +310,7 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
                     harder for a malicious party to obtain certificates with 
potentially misleading or fraudulent values.
                     It is adviseable to double check signed certificates, 
whether these contain all requested SAN fields to enable proper hostname 
verification.
                     The following command can be used to print certificate 
details to the console, which should be compared with what was originally 
requested:
-                    <pre class="line-numbers"><code class="language-bash">     
                   openssl x509 -in certificate.crt -text -noout</code></pre>
+                    <pre class="line-numbers"><code 
class="language-bash">openssl x509 -in certificate.crt -text -noout</code></pre>
                 </li>
             </ol>
         </li>
@@ -320,14 +320,14 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
             <pre>listeners</code></pre>
 
             If SSL is not enabled for inter-broker communication (see below 
for how to enable it), both PLAINTEXT and SSL ports will be necessary.
-            <pre class="line-numbers"><code class="language-text">            
listeners=PLAINTEXT://host.name:port,SSL://host.name:port</code></pre>
+            <pre class="line-numbers"><code 
class="language-text">listeners=PLAINTEXT://host.name:port,SSL://host.name:port</code></pre>
 
             Following SSL configs are needed on the broker side
-            <pre class="line-numbers"><code class="language-text">            
ssl.keystore.location=/var/private/ssl/server.keystore.jks
-            ssl.keystore.password=test1234
-            ssl.key.password=test1234
-            ssl.truststore.location=/var/private/ssl/server.truststore.jks
-            ssl.truststore.password=test1234</code></pre>
+            <pre class="line-numbers"><code 
class="language-text">ssl.keystore.location=/var/private/ssl/server.keystore.jks
+ssl.keystore.password=test1234
+ssl.key.password=test1234
+ssl.truststore.location=/var/private/ssl/server.truststore.jks
+ssl.truststore.password=test1234</code></pre>
 
             Note: ssl.truststore.password is technically optional but highly 
recommended. If a password is not set access to the truststore is still 
available, but integrity checking is disabled.
 
@@ -341,7 +341,7 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
                 <li>ssl.secure.random.implementation=SHA1PRNG</li>
             </ol>
             If you want to enable SSL for inter-broker communication, add the 
following to the server.properties file (it defaults to PLAINTEXT)
-            <pre class="line-numbers"><code class="language-text">            
security.inter.broker.protocol=SSL</code></pre>
+            <pre class="line-numbers"><code 
class="language-text">security.inter.broker.protocol=SSL</code></pre>
 
             <p>
             Due to import regulations in some countries, the Oracle 
implementation limits the strength of cryptographic algorithms available by 
default. If stronger algorithms are needed (for example, AES with 256-bit 
keys), the <a 
href="http://www.oracle.com/technetwork/java/javase/downloads/index.html";>JCE 
Unlimited Strength Jurisdiction Policy Files</a> must be obtained and installed 
in the JDK/JRE. See the
@@ -357,31 +357,31 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
             </p>
 
             Once you start the broker you should be able to see in the 
server.log
-            <pre class="line-numbers"><code class="language-text">            
with addresses: PLAINTEXT -> EndPoint(192.168.64.1,9092,PLAINTEXT),SSL -> 
EndPoint(192.168.64.1,9093,SSL)</code></pre>
+            <pre class="line-numbers"><code class="language-text">with 
addresses: PLAINTEXT -> EndPoint(192.168.64.1,9092,PLAINTEXT),SSL -> 
EndPoint(192.168.64.1,9093,SSL)</code></pre>
 
             To check quickly if  the server keystore and truststore are setup 
properly you can run the following command
             <pre>openssl s_client -debug -connect localhost:9093 
-tls1</code></pre> (Note: TLSv1 should be listed under 
ssl.enabled.protocols)<br>
             In the output of this command you should see server's certificate:
-            <pre class="line-numbers"><code class="language-text">            
-----BEGIN CERTIFICATE-----
-            {variable sized random bytes}
-            -----END CERTIFICATE-----
-            subject=/C=US/ST=CA/L=Santa Clara/O=org/OU=org/CN=Sriharsha 
Chintalapani
-            issuer=/C=US/ST=CA/L=Santa 
Clara/O=org/OU=org/CN=kafka/[email protected]</code></pre>
+            <pre class="line-numbers"><code class="language-text">-----BEGIN 
CERTIFICATE-----
+{variable sized random bytes}
+-----END CERTIFICATE-----
+subject=/C=US/ST=CA/L=Santa Clara/O=org/OU=org/CN=Sriharsha Chintalapani
+issuer=/C=US/ST=CA/L=Santa 
Clara/O=org/OU=org/CN=kafka/[email protected]</code></pre>
             If the certificate does not show up or if there are any other 
error messages then your keystore is not setup properly.</li>
 
         <li><h4 class="anchor-heading"><a id="security_configclients" 
class="anchor-link"></a><a href="#security_configclients">Configuring Kafka 
Clients</a></h4>
             SSL is supported only for the new Kafka Producer and Consumer, the 
older API is not supported. The configs for SSL will be the same for both 
producer and consumer.<br>
             If client authentication is not required in the broker, then the 
following is a minimal configuration example:
-            <pre class="line-numbers"><code class="language-text">            
security.protocol=SSL
-            ssl.truststore.location=/var/private/ssl/client.truststore.jks
-            ssl.truststore.password=test1234</code></pre>
+            <pre class="line-numbers"><code 
class="language-text">security.protocol=SSL
+ssl.truststore.location=/var/private/ssl/client.truststore.jks
+ssl.truststore.password=test1234</code></pre>
 
             Note: ssl.truststore.password is technically optional but highly 
recommended. If a password is not set access to the truststore is still 
available, but integrity checking is disabled.
 
             If client authentication is required, then a keystore must be 
created like in step 1 and the following must also be configured:
-            <pre class="line-numbers"><code class="language-text">            
ssl.keystore.location=/var/private/ssl/client.keystore.jks
-            ssl.keystore.password=test1234
-            ssl.key.password=test1234</code></pre>
+            <pre class="line-numbers"><code 
class="language-text">ssl.keystore.location=/var/private/ssl/client.keystore.jks
+ssl.keystore.password=test1234
+ssl.key.password=test1234</code></pre>
 
             Other configuration settings that may also be needed depending on 
our requirements and the broker configuration:
                 <ol>
@@ -393,8 +393,8 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
                 </ol>
     <br>
             Examples using console-producer and console-consumer:
-            <pre class="line-numbers"><code class="language-bash">            
kafka-console-producer.sh --bootstrap-server localhost:9093 --topic test 
--producer.config client-ssl.properties
-            kafka-console-consumer.sh --bootstrap-server localhost:9093 
--topic test --consumer.config client-ssl.properties</code></pre>
+            <pre class="line-numbers"><code 
class="language-bash">kafka-console-producer.sh --bootstrap-server 
localhost:9093 --topic test --producer.config client-ssl.properties
+kafka-console-consumer.sh --bootstrap-server localhost:9093 --topic test 
--consumer.config client-ssl.properties</code></pre>
         </li>
     </ol>
     <h3 class="anchor-heading"><a id="security_sasl" 
class="anchor-link"></a><a href="#security_sasl">7.3 Authentication using 
SASL</a></h3>
@@ -434,14 +434,14 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
             login module may be specified in the config value. If multiple 
mechanisms are configured on a
             listener, configs must be provided for each mechanism using the 
listener and mechanism prefix.
             For example,
-                    <pre class="line-numbers"><code class="language-text">     
   
listener.name.sasl_ssl.scram-sha-256.sasl.jaas.config=org.apache.kafka.common.security.scram.ScramLoginModule
 required \
-            username="admin" \
-            password="admin-secret";
-        
listener.name.sasl_ssl.plain.sasl.jaas.config=org.apache.kafka.common.security.plain.PlainLoginModule
 required \
-            username="admin" \
-            password="admin-secret" \
-            user_admin="admin-secret" \
-            user_alice="alice-secret";</code></pre>
+                    <pre class="line-numbers"><code 
class="language-text">listener.name.sasl_ssl.scram-sha-256.sasl.jaas.config=org.apache.kafka.common.security.scram.ScramLoginModule
 required \
+    username="admin" \
+    password="admin-secret";
+listener.name.sasl_ssl.plain.sasl.jaas.config=org.apache.kafka.common.security.plain.PlainLoginModule
 required \
+    username="admin" \
+    password="admin-secret" \
+    user_admin="admin-secret" \
+    user_alice="alice-secret";</code></pre>
 
             If JAAS configuration is defined at different levels, the order of 
precedence used is:
             <ul>
@@ -491,13 +491,13 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
                     <a 
href="#security_sasl_oauthbearer_clientconfig">OAUTHBEARER</a>.
                     For example, <a 
href="#security_sasl_gssapi_clientconfig">GSSAPI</a>
                     credentials may be configured as:
-                    <pre class="line-numbers"><code class="language-text">     
   KafkaClient {
-        com.sun.security.auth.module.Krb5LoginModule required
-        useKeyTab=true
-        storeKey=true
-        keyTab="/etc/security/keytabs/kafka_client.keytab"
-        principal="[email protected]";
-    };</code></pre>
+                    <pre class="line-numbers"><code 
class="language-text">KafkaClient {
+    com.sun.security.auth.module.Krb5LoginModule required
+    useKeyTab=true
+    storeKey=true
+    keyTab="/etc/security/keytabs/kafka_client.keytab"
+    principal="[email protected]";
+};</code></pre>
                 </li>
                 <li>Pass the JAAS config file location as JVM parameter to 
each client JVM. For example:
                     <pre class="language-bash">    
-Djava.security.auth.login.config=/etc/kafka/kafka_client_jaas.conf</code></pre></li>
@@ -532,11 +532,11 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
             <li>Configure a SASL port in server.properties, by adding at least 
one of
                 SASL_PLAINTEXT or SASL_SSL to the <i>listeners</i> parameter, 
which
                 contains one or more comma-separated values:
-                <pre>    listeners=SASL_PLAINTEXT://host.name:port</code></pre>
+                <pre>listeners=SASL_PLAINTEXT://host.name:port</code></pre>
                 If you are only configuring a SASL port (or if you want
                 the Kafka brokers to authenticate each other using SASL) then 
make sure
                 you set the same SASL protocol for inter-broker communication:
-                <pre>    security.inter.broker.protocol=SASL_PLAINTEXT (or 
SASL_SSL)</code></pre></li>
+                <pre>security.inter.broker.protocol=SASL_PLAINTEXT (or 
SASL_SSL)</code></pre></li>
             <li>Select one or more  <a 
href="#security_sasl_mechanism">supported mechanisms</a>
                 to enable in the broker and follow the steps to configure SASL 
for the mechanism.
                 To enable multiple mechanisms in the broker, follow the steps
@@ -563,45 +563,45 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
             <li><b>Create Kerberos Principals</b><br>
             If you are using the organization's Kerberos or Active Directory 
server, ask your Kerberos administrator for a principal for each Kafka broker 
in your cluster and for every operating system user that will access Kafka with 
Kerberos authentication (via clients and tools).</br>
             If you have installed your own Kerberos, you will need to create 
these principals yourself using the following commands:
-                <pre class="line-numbers"><code class="language-bash">        
sudo /usr/sbin/kadmin.local -q 'addprinc -randkey kafka/{hostname}@{REALM}'
-        sudo /usr/sbin/kadmin.local -q "ktadd -k 
/etc/security/keytabs/{keytabname}.keytab 
kafka/{hostname}@{REALM}"</code></pre></li>
+                <pre class="line-numbers"><code class="language-bash">sudo 
/usr/sbin/kadmin.local -q 'addprinc -randkey kafka/{hostname}@{REALM}'
+sudo /usr/sbin/kadmin.local -q "ktadd -k 
/etc/security/keytabs/{keytabname}.keytab 
kafka/{hostname}@{REALM}"</code></pre></li>
             <li><b>Make sure all hosts can be reachable using hostnames</b> - 
it is a Kerberos requirement that all your hosts can be resolved with their 
FQDNs.</li>
         </ol>
         <li><h5 class="anchor-heading"><a 
id="security_sasl_kerberos_brokerconfig" class="anchor-link"></a><a 
href="#security_sasl_kerberos_brokerconfig">Configuring Kafka Brokers</a></h5>
         <ol>
             <li>Add a suitably modified JAAS file similar to the one below to 
each Kafka broker's config directory, let's call it kafka_server_jaas.conf for 
this example (note that each broker should have its own keytab):
-            <pre class="line-numbers"><code class="language-text">        
KafkaServer {
-            com.sun.security.auth.module.Krb5LoginModule required
-            useKeyTab=true
-            storeKey=true
-            keyTab="/etc/security/keytabs/kafka_server.keytab"
-            principal="kafka/[email protected]";
-        };
-
-        // Zookeeper client authentication
-        Client {
-        com.sun.security.auth.module.Krb5LoginModule required
-        useKeyTab=true
-        storeKey=true
-        keyTab="/etc/security/keytabs/kafka_server.keytab"
-        principal="kafka/[email protected]";
-        };</code></pre>
+            <pre class="line-numbers"><code class="language-text">KafkaServer {
+    com.sun.security.auth.module.Krb5LoginModule required
+    useKeyTab=true
+    storeKey=true
+    keyTab="/etc/security/keytabs/kafka_server.keytab"
+    principal="kafka/[email protected]";
+};
+
+// Zookeeper client authentication
+Client {
+    com.sun.security.auth.module.Krb5LoginModule required
+    useKeyTab=true
+    storeKey=true
+    keyTab="/etc/security/keytabs/kafka_server.keytab"
+    principal="kafka/[email protected]";
+};</code></pre>
 
             </li>
             <tt>KafkaServer</tt> section in the JAAS file tells the broker 
which principal to use and the location of the keytab where this principal is 
stored. It
             allows the broker to login using the keytab specified in this 
section. See <a href="#security_jaas_broker">notes</a> for more details on 
Zookeeper SASL configuration.
             <li>Pass the JAAS and optionally the krb5 file locations as JVM 
parameters to each Kafka broker (see <a 
href="https://docs.oracle.com/javase/8/docs/technotes/guides/security/jgss/tutorials/KerberosReq.html";>here</a>
 for more details):
-                <pre>    -Djava.security.krb5.conf=/etc/kafka/krb5.conf
-        
-Djava.security.auth.login.config=/etc/kafka/kafka_server_jaas.conf</code></pre>
+                <pre>-Djava.security.krb5.conf=/etc/kafka/krb5.conf
+-Djava.security.auth.login.config=/etc/kafka/kafka_server_jaas.conf</code></pre>
             </li>
             <li>Make sure the keytabs configured in the JAAS file are readable 
by the operating system user who is starting kafka broker.</li>
             <li>Configure SASL port and SASL mechanisms in server.properties 
as described <a href="#security_sasl_brokerconfig">here</a>. For example:
-            <pre>    listeners=SASL_PLAINTEXT://host.name:port
-        security.inter.broker.protocol=SASL_PLAINTEXT
-        sasl.mechanism.inter.broker.protocol=GSSAPI
-        sasl.enabled.mechanisms=GSSAPI</code></pre>
+            <pre>listeners=SASL_PLAINTEXT://host.name:port
+security.inter.broker.protocol=SASL_PLAINTEXT
+sasl.mechanism.inter.broker.protocol=GSSAPI
+sasl.enabled.mechanisms=GSSAPI</code></pre>
             </li>We must also configure the service name in server.properties, 
which should match the principal name of the kafka brokers. In the above 
example, principal is "kafka/[email protected]", so:
-            <pre>    sasl.kerberos.service.name=kafka</code></pre>
+            <pre>sasl.kerberos.service.name=kafka</code></pre>
 
         </ol></li>
         <li><h5 class="anchor-heading"><a 
id="security_sasl_kerberos_clientconfig" class="anchor-link"></a><a 
href="#security_sasl_kerberos_clientconfig">Configuring Kafka Clients</a></h5>
@@ -615,27 +615,27 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
                     The property <code>sasl.jaas.config</code> in 
producer.properties or consumer.properties describes
                     how clients like producer and consumer can connect to the 
Kafka Broker. The following is an example
                     configuration for a client using a keytab (recommended for 
long-running processes):
-                <pre class="line-numbers"><code class="language-text">    
sasl.jaas.config=com.sun.security.auth.module.Krb5LoginModule required \
-        useKeyTab=true \
-        storeKey=true  \
-        keyTab="/etc/security/keytabs/kafka_client.keytab" \
-        principal="[email protected]";</code></pre>
+                <pre class="line-numbers"><code 
class="language-text">sasl.jaas.config=com.sun.security.auth.module.Krb5LoginModule
 required \
+    useKeyTab=true \
+    storeKey=true  \
+    keyTab="/etc/security/keytabs/kafka_client.keytab" \
+    principal="[email protected]";</code></pre>
 
                    For command-line utilities like kafka-console-consumer or 
kafka-console-producer, kinit can be used
                    along with "useTicketCache=true" as in:
-                <pre class="line-numbers"><code class="language-text">    
sasl.jaas.config=com.sun.security.auth.module.Krb5LoginModule required \
-        useTicketCache=true;</code></pre>
+                <pre class="line-numbers"><code 
class="language-text">sasl.jaas.config=com.sun.security.auth.module.Krb5LoginModule
 required \
+    useTicketCache=true;</code></pre>
 
                    JAAS configuration for clients may alternatively be 
specified as a JVM parameter similar to brokers
                    as described <a 
href="#security_client_staticjaas">here</a>. Clients use the login section named
                    <tt>KafkaClient</tt>. This option allows only one user for 
all client connections from a JVM.</li>
                 <li>Make sure the keytabs configured in the JAAS configuration 
are readable by the operating system user who is starting kafka client.</li>
                 <li>Optionally pass the krb5 file locations as JVM parameters 
to each client JVM (see <a 
href="https://docs.oracle.com/javase/8/docs/technotes/guides/security/jgss/tutorials/KerberosReq.html";>here</a>
 for more details):
-                <pre>    
-Djava.security.krb5.conf=/etc/kafka/krb5.conf</code></pre></li>
+                
<pre>-Djava.security.krb5.conf=/etc/kafka/krb5.conf</code></pre></li>
                 <li>Configure the following properties in producer.properties 
or consumer.properties:
-                <pre class="line-numbers"><code class="language-text">    
security.protocol=SASL_PLAINTEXT (or SASL_SSL)
-    sasl.mechanism=GSSAPI
-    sasl.kerberos.service.name=kafka</code></pre></li>
+                <pre class="line-numbers"><code 
class="language-text">security.protocol=SASL_PLAINTEXT (or SASL_SSL)
+sasl.mechanism=GSSAPI
+sasl.kerberos.service.name=kafka</code></pre></li>
             </ol>
         </li>
         </ol>
@@ -649,25 +649,25 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
         <li><h5 class="anchor-heading"><a 
id="security_sasl_plain_brokerconfig" class="anchor-link"></a><a 
href="#security_sasl_plain_brokerconfig">Configuring Kafka Brokers</a></h5>
             <ol>
             <li>Add a suitably modified JAAS file similar to the one below to 
each Kafka broker's config directory, let's call it kafka_server_jaas.conf for 
this example:
-                <pre class="line-numbers"><code class="language-text">        
KafkaServer {
-            org.apache.kafka.common.security.plain.PlainLoginModule required
-            username="admin"
-            password="admin-secret"
-            user_admin="admin-secret"
-            user_alice="alice-secret";
-        };</code></pre>
+                <pre class="line-numbers"><code 
class="language-text">KafkaServer {
+    org.apache.kafka.common.security.plain.PlainLoginModule required
+    username="admin"
+    password="admin-secret"
+    user_admin="admin-secret"
+    user_alice="alice-secret";
+};</code></pre>
                 This configuration defines two users (<i>admin</i> and 
<i>alice</i>). The properties <tt>username</tt> and <tt>password</tt>
                 in the <tt>KafkaServer</tt> section are used by the broker to 
initiate connections to other brokers. In this example,
                 <i>admin</i> is the user for inter-broker communication. The 
set of properties <tt>user_<i>userName</i></tt> defines
                 the passwords for all users that connect to the broker and the 
broker validates all client connections including
                 those from other brokers using these properties.</li>
             <li>Pass the JAAS config file location as JVM parameter to each 
Kafka broker:
-                <pre>    
-Djava.security.auth.login.config=/etc/kafka/kafka_server_jaas.conf</code></pre></li>
+                
<pre>-Djava.security.auth.login.config=/etc/kafka/kafka_server_jaas.conf</code></pre></li>
             <li>Configure SASL port and SASL mechanisms in server.properties 
as described <a href="#security_sasl_brokerconfig">here</a>. For example:
-                <pre>    listeners=SASL_SSL://host.name:port
-        security.inter.broker.protocol=SASL_SSL
-        sasl.mechanism.inter.broker.protocol=PLAIN
-        sasl.enabled.mechanisms=PLAIN</code></pre></li>
+                <pre>listeners=SASL_SSL://host.name:port
+security.inter.broker.protocol=SASL_SSL
+sasl.mechanism.inter.broker.protocol=PLAIN
+sasl.enabled.mechanisms=PLAIN</code></pre></li>
             </ol>
         </li>
 
@@ -677,9 +677,9 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
             <li>Configure the JAAS configuration property for each client in 
producer.properties or consumer.properties.
                 The login module describes how the clients like producer and 
consumer can connect to the Kafka Broker.
                 The following is an example configuration for a client for the 
PLAIN mechanism:
-                <pre class="line-numbers"><code class="language-text">    
sasl.jaas.config=org.apache.kafka.common.security.plain.PlainLoginModule 
required \
-        username="alice" \
-        password="alice-secret";</code></pre>
+                <pre class="line-numbers"><code 
class="language-text">sasl.jaas.config=org.apache.kafka.common.security.plain.PlainLoginModule
 required \
+    username="alice" \
+    password="alice-secret";</code></pre>
                 <p>The options <tt>username</tt> and <tt>password</tt> are 
used by clients to configure
                 the user for client connections. In this example, clients 
connect to the broker as user <i>alice</i>.
                 Different clients within a JVM may connect as different users 
by specifying different user names
@@ -689,8 +689,8 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
                 as described <a href="#security_client_staticjaas">here</a>. 
Clients use the login section named
                 <tt>KafkaClient</tt>. This option allows only one user for all 
client connections from a JVM.</p></li>
             <li>Configure the following properties in producer.properties or 
consumer.properties:
-                <pre class="line-numbers"><code class="language-text">    
security.protocol=SASL_SSL
-    sasl.mechanism=PLAIN</code></pre></li>
+                <pre class="line-numbers"><code 
class="language-text">security.protocol=SASL_SSL
+sasl.mechanism=PLAIN</code></pre></li>
             </ol>
         </li>
         <li><h5><a id="security_sasl_plain_production" 
href="#security_sasl_plain_production">Use of SASL/PLAIN in production</a></h5>
@@ -726,35 +726,35 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
         before Kafka brokers are started. Client credentials may be created 
and updated dynamically and updated
         credentials will be used to authenticate new connections.</p>
         <p>Create SCRAM credentials for user <i>alice</i> with password 
<i>alice-secret</i>:
-        <pre class="line-numbers"><code class="language-bash">    > 
bin/kafka-configs.sh --zookeeper localhost:2182 --zk-tls-config-file 
zk_tls_config.properties --alter --add-config 
'SCRAM-SHA-256=[iterations=8192,password=alice-secret],SCRAM-SHA-512=[password=alice-secret]'
 --entity-type users --entity-name alice</code></pre>
+        <pre class="line-numbers"><code class="language-bash">> 
bin/kafka-configs.sh --zookeeper localhost:2182 --zk-tls-config-file 
zk_tls_config.properties --alter --add-config 
'SCRAM-SHA-256=[iterations=8192,password=alice-secret],SCRAM-SHA-512=[password=alice-secret]'
 --entity-type users --entity-name alice</code></pre>
         <p>The default iteration count of 4096 is used if iterations are not 
specified. A random salt is created
         and the SCRAM identity consisting of salt, iterations, StoredKey and 
ServerKey are stored in Zookeeper.
         See <a href="https://tools.ietf.org/html/rfc5802";>RFC 5802</a> for 
details on SCRAM identity and the individual fields.
         <p>The following examples also require a user <i>admin</i> for 
inter-broker communication which can be created using:
-        <pre class="line-numbers"><code class="language-bash">    > 
bin/kafka-configs.sh --zookeeper localhost:2182 --zk-tls-config-file 
zk_tls_config.properties --alter --add-config 
'SCRAM-SHA-256=[password=admin-secret],SCRAM-SHA-512=[password=admin-secret]' 
--entity-type users --entity-name admin</code></pre>
+        <pre class="line-numbers"><code class="language-bash">> 
bin/kafka-configs.sh --zookeeper localhost:2182 --zk-tls-config-file 
zk_tls_config.properties --alter --add-config 
'SCRAM-SHA-256=[password=admin-secret],SCRAM-SHA-512=[password=admin-secret]' 
--entity-type users --entity-name admin</code></pre>
         <p>Existing credentials may be listed using the <i>--describe</i> 
option:
-        <pre class="line-numbers"><code class="language-bash">    > 
bin/kafka-configs.sh --zookeeper localhost:2182 --zk-tls-config-file 
zk_tls_config.properties --describe --entity-type users --entity-name 
alice</code></pre>
+        <pre class="line-numbers"><code class="language-bash">> 
bin/kafka-configs.sh --zookeeper localhost:2182 --zk-tls-config-file 
zk_tls_config.properties --describe --entity-type users --entity-name 
alice</code></pre>
         <p>Credentials may be deleted for one or more SCRAM mechanisms using 
the <i>--alter --delete-config</i> option:
-        <pre class="line-numbers"><code class="language-bash">    > 
bin/kafka-configs.sh --zookeeper localhost:2182 --zk-tls-config-file 
zk_tls_config.properties --alter --delete-config 'SCRAM-SHA-512' --entity-type 
users --entity-name alice</code></pre>
+        <pre class="line-numbers"><code class="language-bash">> 
bin/kafka-configs.sh --zookeeper localhost:2182 --zk-tls-config-file 
zk_tls_config.properties --alter --delete-config 'SCRAM-SHA-512' --entity-type 
users --entity-name alice</code></pre>
         </li>
         <li><h5 class="anchor-heading"><a 
id="security_sasl_scram_brokerconfig" class="anchor-link"></a><a 
href="#security_sasl_scram_brokerconfig">Configuring Kafka Brokers</a></h5>
             <ol>
             <li>Add a suitably modified JAAS file similar to the one below to 
each Kafka broker's config directory, let's call it kafka_server_jaas.conf for 
this example:
-                <pre class="line-numbers"><code class="language-text">    
KafkaServer {
-        org.apache.kafka.common.security.scram.ScramLoginModule required
-        username="admin"
-        password="admin-secret";
-    };</code></pre>
+                <pre class="line-numbers"><code 
class="language-text">KafkaServer {
+    org.apache.kafka.common.security.scram.ScramLoginModule required
+    username="admin"
+    password="admin-secret";
+};</code></pre>
                 The properties <tt>username</tt> and <tt>password</tt> in the 
<tt>KafkaServer</tt> section are used by
                 the broker to initiate connections to other brokers. In this 
example, <i>admin</i> is the user for
                 inter-broker communication.</li>
             <li>Pass the JAAS config file location as JVM parameter to each 
Kafka broker:
-                <pre>    
-Djava.security.auth.login.config=/etc/kafka/kafka_server_jaas.conf</code></pre></li>
+                
<pre>-Djava.security.auth.login.config=/etc/kafka/kafka_server_jaas.conf</code></pre></li>
             <li>Configure SASL port and SASL mechanisms in server.properties 
as described <a href="#security_sasl_brokerconfig">here</a>.</code></pre> For 
example:
-                <pre class="line-numbers"><code class="language-text">    
listeners=SASL_SSL://host.name:port
-    security.inter.broker.protocol=SASL_SSL
-    sasl.mechanism.inter.broker.protocol=SCRAM-SHA-256 (or SCRAM-SHA-512)
-    sasl.enabled.mechanisms=SCRAM-SHA-256 (or SCRAM-SHA-512)</code></pre></li>
+                <pre class="line-numbers"><code 
class="language-text">listeners=SASL_SSL://host.name:port
+security.inter.broker.protocol=SASL_SSL
+sasl.mechanism.inter.broker.protocol=SCRAM-SHA-256 (or SCRAM-SHA-512)
+sasl.enabled.mechanisms=SCRAM-SHA-256 (or SCRAM-SHA-512)</code></pre></li>
             </ol>
         </li>
 
@@ -764,9 +764,9 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
             <li>Configure the JAAS configuration property for each client in 
producer.properties or consumer.properties.
                 The login module describes how the clients like producer and 
consumer can connect to the Kafka Broker.
                 The following is an example configuration for a client for the 
SCRAM mechanisms:
-                <pre class="line-numbers"><code class="language-text">   
sasl.jaas.config=org.apache.kafka.common.security.scram.ScramLoginModule 
required \
-        username="alice" \
-        password="alice-secret";</code></pre>
+                <pre class="line-numbers"><code 
class="language-text">sasl.jaas.config=org.apache.kafka.common.security.scram.ScramLoginModule
 required \
+    username="alice" \
+    password="alice-secret";</code></pre>
 
                 <p>The options <tt>username</tt> and <tt>password</tt> are 
used by clients to configure
                 the user for client connections. In this example, clients 
connect to the broker as user <i>alice</i>.
@@ -777,8 +777,8 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
                 as described <a href="#security_client_staticjaas">here</a>. 
Clients use the login section named
                 <tt>KafkaClient</tt>. This option allows only one user for all 
client connections from a JVM.</p></li>
             <li>Configure the following properties in producer.properties or 
consumer.properties:
-                <pre class="line-numbers"><code class="language-text">    
security.protocol=SASL_SSL
-    sasl.mechanism=SCRAM-SHA-256 (or SCRAM-SHA-512)</code></pre></li>
+                <pre class="line-numbers"><code 
class="language-text">security.protocol=SASL_SSL
+sasl.mechanism=SCRAM-SHA-256 (or SCRAM-SHA-512)</code></pre></li>
             </ol>
         </li>
         <li><h5><a id="security_sasl_scram_security" 
href="#security_sasl_scram_security">Security Considerations for 
SASL/SCRAM</a></h5>
@@ -811,20 +811,20 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
         <li><h5 class="anchor-heading"><a 
id="security_sasl_oauthbearer_brokerconfig" class="anchor-link"></a><a 
href="#security_sasl_oauthbearer_brokerconfig">Configuring Kafka 
Brokers</a></h5>
             <ol>
             <li>Add a suitably modified JAAS file similar to the one below to 
each Kafka broker's config directory, let's call it kafka_server_jaas.conf for 
this example:
-                <pre class="line-numbers"><code class="language-text">    
KafkaServer {
-        org.apache.kafka.common.security.oauthbearer.OAuthBearerLoginModule 
required
-        unsecuredLoginStringClaim_sub="admin";
-    };</code></pre>
+                <pre class="line-numbers"><code 
class="language-text">KafkaServer {
+    org.apache.kafka.common.security.oauthbearer.OAuthBearerLoginModule 
required
+    unsecuredLoginStringClaim_sub="admin";
+};</code></pre>
                 The property <tt>unsecuredLoginStringClaim_sub</tt> in the 
<tt>KafkaServer</tt> section is used by
                 the broker when it initiates connections to other brokers. In 
this example, <i>admin</i> will appear in the
                 subject (<tt>sub</tt>) claim and will be the user for 
inter-broker communication.</li>
             <li>Pass the JAAS config file location as JVM parameter to each 
Kafka broker:
-                <pre>    
-Djava.security.auth.login.config=/etc/kafka/kafka_server_jaas.conf</code></pre></li>
+                
<pre>-Djava.security.auth.login.config=/etc/kafka/kafka_server_jaas.conf</code></pre></li>
             <li>Configure SASL port and SASL mechanisms in server.properties 
as described <a href="#security_sasl_brokerconfig">here</a>.</code></pre> For 
example:
-                <pre class="line-numbers"><code class="language-text">    
listeners=SASL_SSL://host.name:port (or SASL_PLAINTEXT if non-production)
-    security.inter.broker.protocol=SASL_SSL (or SASL_PLAINTEXT if 
non-production)
-    sasl.mechanism.inter.broker.protocol=OAUTHBEARER
-    sasl.enabled.mechanisms=OAUTHBEARER</code></pre></li>
+                <pre class="line-numbers"><code 
class="language-text">listeners=SASL_SSL://host.name:port (or SASL_PLAINTEXT if 
non-production)
+security.inter.broker.protocol=SASL_SSL (or SASL_PLAINTEXT if non-production)
+sasl.mechanism.inter.broker.protocol=OAUTHBEARER
+sasl.enabled.mechanisms=OAUTHBEARER</code></pre></li>
             </ol>
         </li>
 
@@ -834,8 +834,8 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
            <li>Configure the JAAS configuration property for each client in 
producer.properties or consumer.properties.
                 The login module describes how the clients like producer and 
consumer can connect to the Kafka Broker.
                The following is an example configuration for a client for the 
OAUTHBEARER mechanisms:
-                <pre class="line-numbers"><code class="language-text">   
sasl.jaas.config=org.apache.kafka.common.security.oauthbearer.OAuthBearerLoginModule
 required \
-        unsecuredLoginStringClaim_sub="alice";</code></pre>
+                <pre class="line-numbers"><code 
class="language-text">sasl.jaas.config=org.apache.kafka.common.security.oauthbearer.OAuthBearerLoginModule
 required \
+    unsecuredLoginStringClaim_sub="alice";</code></pre>
 
                 <p>The option <tt>unsecuredLoginStringClaim_sub</tt> is used 
by clients to configure
                 the subject (<tt>sub</tt>) claim, which determines the user 
for client connections.
@@ -847,8 +847,8 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
                 as described <a href="#security_client_staticjaas">here</a>. 
Clients use the login section named
                 <tt>KafkaClient</tt>. This option allows only one user for all 
client connections from a JVM.</p></li>
             <li>Configure the following properties in producer.properties or 
consumer.properties:
-                <pre class="line-numbers"><code class="language-text">    
security.protocol=SASL_SSL (or SASL_PLAINTEXT if non-production)
-    sasl.mechanism=OAUTHBEARER</code></pre></li>
+                <pre class="line-numbers"><code 
class="language-text">security.protocol=SASL_SSL (or SASL_PLAINTEXT if 
non-production)
+sasl.mechanism=OAUTHBEARER</code></pre></li>
              <li>The default implementation of SASL/OAUTHBEARER depends on the 
jackson-databind library.
                  Since it's an optional dependency, users have to configure it 
as a dependency via their build tool.</li>
             </ol>
@@ -1006,23 +1006,23 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
     <li><h4 class="anchor-heading"><a id="security_sasl_multimechanism" 
class="anchor-link"></a><a href="#security_sasl_multimechanism">Enabling 
multiple SASL mechanisms in a broker</a></h4>
         <ol>
         <li>Specify configuration for the login modules of all enabled 
mechanisms in the <tt>KafkaServer</tt> section of the JAAS config file. For 
example:
-            <pre class="line-numbers"><code class="language-text">        
KafkaServer {
-            com.sun.security.auth.module.Krb5LoginModule required
-            useKeyTab=true
-            storeKey=true
-            keyTab="/etc/security/keytabs/kafka_server.keytab"
-            principal="kafka/[email protected]";
-
-            org.apache.kafka.common.security.plain.PlainLoginModule required
-            username="admin"
-            password="admin-secret"
-            user_admin="admin-secret"
-            user_alice="alice-secret";
-        };</code></pre></li>
-        <li>Enable the SASL mechanisms in server.properties: <pre>    
sasl.enabled.mechanisms=GSSAPI,PLAIN,SCRAM-SHA-256,SCRAM-SHA-512,OAUTHBEARER</code></pre></li>
+            <pre class="line-numbers"><code class="language-text">KafkaServer {
+    com.sun.security.auth.module.Krb5LoginModule required
+    useKeyTab=true
+    storeKey=true
+    keyTab="/etc/security/keytabs/kafka_server.keytab"
+    principal="kafka/[email protected]";
+
+    org.apache.kafka.common.security.plain.PlainLoginModule required
+    username="admin"
+    password="admin-secret"
+    user_admin="admin-secret"
+    user_alice="alice-secret";
+};</code></pre></li>
+        <li>Enable the SASL mechanisms in server.properties: 
<pre>sasl.enabled.mechanisms=GSSAPI,PLAIN,SCRAM-SHA-256,SCRAM-SHA-512,OAUTHBEARER</code></pre></li>
         <li>Specify the SASL security protocol and mechanism for inter-broker 
communication in server.properties if required:
-            <pre class="line-numbers"><code class="language-text">    
security.inter.broker.protocol=SASL_PLAINTEXT (or SASL_SSL)
-    sasl.mechanism.inter.broker.protocol=GSSAPI (or one of the other enabled 
mechanisms)</code></pre></li>
+            <pre class="line-numbers"><code 
class="language-text">security.inter.broker.protocol=SASL_PLAINTEXT (or 
SASL_SSL)
+sasl.mechanism.inter.broker.protocol=GSSAPI (or one of the other enabled 
mechanisms)</code></pre></li>
         <li>Follow the mechanism-specific steps in <a 
href="#security_sasl_kerberos_brokerconfig">GSSAPI (Kerberos)</a>,
             <a href="#security_sasl_plain_brokerconfig">PLAIN</a>,
             <a href="#security_sasl_scram_brokerconfig">SCRAM</a> and <a 
href="#security_sasl_oauthbearer_brokerconfig">OAUTHBEARER</a>
@@ -1081,13 +1081,13 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
             Tokens can not be requests if the initial authentication is done 
through delegation token.
             <tt>kafka-delegation-tokens.sh</tt> script examples are given 
below.</p>
         <p>Create a delegation token:
-        <pre class="line-numbers"><code class="language-bash">    > 
bin/kafka-delegation-tokens.sh --bootstrap-server localhost:9092 --create   
--max-life-time-period -1 --command-config client.properties 
--renewer-principal User:user1</code></pre>
+        <pre class="line-numbers"><code class="language-bash">> 
bin/kafka-delegation-tokens.sh --bootstrap-server localhost:9092 --create   
--max-life-time-period -1 --command-config client.properties 
--renewer-principal User:user1</code></pre>
         <p>Renew a delegation token:
-        <pre class="line-numbers"><code class="language-bash">    > 
bin/kafka-delegation-tokens.sh --bootstrap-server localhost:9092 --renew    
--renew-time-period -1 --command-config client.properties --hmac 
ABCDEFGHIJK</code></pre>
+        <pre class="line-numbers"><code class="language-bash">> 
bin/kafka-delegation-tokens.sh --bootstrap-server localhost:9092 --renew    
--renew-time-period -1 --command-config client.properties --hmac 
ABCDEFGHIJK</code></pre>
         <p>Expire a delegation token:
-        <pre class="line-numbers"><code class="language-bash">    > 
bin/kafka-delegation-tokens.sh --bootstrap-server localhost:9092 --expire   
--expiry-time-period -1   --command-config client.properties  --hmac 
ABCDEFGHIJK</code></pre>
+        <pre class="line-numbers"><code class="language-bash">> 
bin/kafka-delegation-tokens.sh --bootstrap-server localhost:9092 --expire   
--expiry-time-period -1   --command-config client.properties  --hmac 
ABCDEFGHIJK</code></pre>
         <p>Existing tokens can be described using the --describe option:
-        <pre class="line-numbers"><code class="language-bash">    > 
bin/kafka-delegation-tokens.sh --bootstrap-server localhost:9092 --describe 
--command-config client.properties  --owner-principal User:user1</code></pre>
+        <pre class="line-numbers"><code class="language-bash">> 
bin/kafka-delegation-tokens.sh --bootstrap-server localhost:9092 --describe 
--command-config client.properties  --owner-principal User:user1</code></pre>
         </li>
         <li><h5 class="anchor-heading"><a id="security_token_authentication" 
class="anchor-link"></a><a href="#security_token_authentication">Token 
Authentication</a></h5>
             <p>Delegation token authentication piggybacks on the current 
SASL/SCRAM authentication mechanism. We must enable
@@ -1098,10 +1098,10 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
                     <li>Configure the JAAS configuration property for each 
client in producer.properties or consumer.properties.
                 The login module describes how the clients like producer and 
consumer can connect to the Kafka Broker.
                 The following is an example configuration for a client for the 
token authentication:
-                <pre class="line-numbers"><code class="language-text">   
sasl.jaas.config=org.apache.kafka.common.security.scram.ScramLoginModule 
required \
-        username="tokenID123" \
-        password="lAYYSFmLs4bTjf+lTZ1LCHR/ZZFNA==" \
-        tokenauth="true";</code></pre>
+                <pre class="line-numbers"><code 
class="language-text">sasl.jaas.config=org.apache.kafka.common.security.scram.ScramLoginModule
 required \
+    username="tokenID123" \
+    password="lAYYSFmLs4bTjf+lTZ1LCHR/ZZFNA==" \
+    tokenauth="true";</code></pre>
 
                 <p>The options <tt>username</tt> and <tt>password</tt> are 
used by clients to configure the token id and
                     token HMAC. And the option <tt>tokenauth</tt> is used to 
indicate the server about token authentication.
@@ -1153,14 +1153,14 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
     string representation of the X.500 certificate distinguished name. If the 
distinguished name matches the pattern, then the replacement command will be 
run over the name.
     This also supports lowercase/uppercase options, to force the translated 
result to be all lower/uppercase case. This is done by adding a "/L" or "/U' to 
the end of the rule.
 
-    <pre class="line-numbers"><code class="language-text">        
RULE:pattern/replacement/
-        RULE:pattern/replacement/[LU]</code></pre>
+    <pre class="line-numbers"><code 
class="language-text">RULE:pattern/replacement/
+RULE:pattern/replacement/[LU]</code></pre>
 
     Example <code>ssl.principal.mapping.rules</code> values are:
-    <pre class="line-numbers"><code class="language-text">        
RULE:^CN=(.*?),OU=ServiceUsers.*$/$1/,
-        RULE:^CN=(.*?),OU=(.*?),O=(.*?),L=(.*?),ST=(.*?),C=(.*?)$/$1@$2/L,
-        RULE:^.*[Cc][Nn]=([a-zA-Z0-9.]*).*$/$1/L,
-        DEFAULT</code></pre>
+    <pre class="line-numbers"><code 
class="language-text">RULE:^CN=(.*?),OU=ServiceUsers.*$/$1/,
+RULE:^CN=(.*?),OU=(.*?),O=(.*?),L=(.*?),ST=(.*?),C=(.*?)$/$1@$2/L,
+RULE:^.*[Cc][Nn]=([a-zA-Z0-9.]*).*$/$1/L,
+DEFAULT</code></pre>
 
     Above rules translate distinguished name 
"CN=serviceuser,OU=ServiceUsers,O=Unknown,L=Unknown,ST=Unknown,C=Unknown" to 
"serviceuser"
     and "CN=adminUser,OU=Admin,O=Unknown,L=Unknown,ST=Unknown,C=Unknown" to 
"adminuser@admin".
@@ -1173,12 +1173,12 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
     By default, the SASL user name will be the primary part of the Kerberos 
principal. One can change that by setting 
<code>sasl.kerberos.principal.to.local.rules</code> to a customized rule in 
server.properties.
     The format of <code>sasl.kerberos.principal.to.local.rules</code> is a 
list where each rule works in the same way as the auth_to_local in <a 
href="http://web.mit.edu/Kerberos/krb5-latest/doc/admin/conf_files/krb5_conf.html";>Kerberos
 configuration file (krb5.conf)</a>. This also support additional 
lowercase/uppercase rule, to force the translated result to be all 
lowercase/uppercase. This is done by adding a "/L" or "/U" to the end of the 
rule. check below formats for syntax.
     Each rules starts with RULE: and contains an expression as the following 
formats. See the kerberos documentation for more details.
-    <pre class="line-numbers"><code class="language-text">        
RULE:[n:string](regexp)s/pattern/replacement/
-        RULE:[n:string](regexp)s/pattern/replacement/g
-        RULE:[n:string](regexp)s/pattern/replacement//L
-        RULE:[n:string](regexp)s/pattern/replacement/g/L
-        RULE:[n:string](regexp)s/pattern/replacement//U
-        RULE:[n:string](regexp)s/pattern/replacement/g/U</code></pre>
+    <pre class="line-numbers"><code 
class="language-text">RULE:[n:string](regexp)s/pattern/replacement/
+RULE:[n:string](regexp)s/pattern/replacement/g
+RULE:[n:string](regexp)s/pattern/replacement//L
+RULE:[n:string](regexp)s/pattern/replacement/g/L
+RULE:[n:string](regexp)s/pattern/replacement//U
+RULE:[n:string](regexp)s/pattern/replacement/g/U</code></pre>
 
     An example of adding a rule to properly translate [email protected] to 
user while also keeping the default rule in place is:
     
<pre>sasl.kerberos.principal.to.local.rules=RULE:[1:$1@$0](.*@MYDOMAIN.COM)s/@.*//,DEFAULT</code></pre>
@@ -1417,9 +1417,9 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
             Users having Alter permission on ClusterResource can use Admin API 
for ACL management. kafka-acls.sh script supports AdminClient API to manage 
ACLs without interacting with zookeeper/authorizer directly.
             All the above examples can be executed by using 
<b>--bootstrap-server</b> option. For example:
 
-            <pre class="line-numbers"><code class="language-bash">            
bin/kafka-acls.sh --bootstrap-server localhost:9092 --command-config 
/tmp/adminclient-configs.conf --add --allow-principal User:Bob --producer 
--topic Test-topic
-            bin/kafka-acls.sh --bootstrap-server localhost:9092 
--command-config /tmp/adminclient-configs.conf --add --allow-principal User:Bob 
--consumer --topic Test-topic --group Group-1
-            bin/kafka-acls.sh --bootstrap-server localhost:9092 
--command-config /tmp/adminclient-configs.conf --list --topic 
Test-topic</code></pre></li>
+            <pre class="line-numbers"><code 
class="language-bash">bin/kafka-acls.sh --bootstrap-server localhost:9092 
--command-config /tmp/adminclient-configs.conf --add --allow-principal User:Bob 
--producer --topic Test-topic
+bin/kafka-acls.sh --bootstrap-server localhost:9092 --command-config 
/tmp/adminclient-configs.conf --add --allow-principal User:Bob --consumer 
--topic Test-topic --group Group-1
+bin/kafka-acls.sh --bootstrap-server localhost:9092 --command-config 
/tmp/adminclient-configs.conf --list --topic Test-topic</code></pre></li>
 
     </ul>
 
@@ -1929,43 +1929,43 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
         When performing an incremental bounce stop the brokers cleanly via a 
SIGTERM. It's also good practice to wait for restarted replicas to return to 
the ISR list before moving onto the next node.
         <p></p>
         As an example, say we wish to encrypt both broker-client and 
broker-broker communication with SSL. In the first incremental bounce, an SSL 
port is opened on each node:
-            <pre class="line-numbers"><code class="language-text">            
listeners=PLAINTEXT://broker1:9091,SSL://broker1:9092</code></pre>
+            <pre class="line-numbers"><code 
class="language-text">listeners=PLAINTEXT://broker1:9091,SSL://broker1:9092</code></pre>
 
         We then restart the clients, changing their config to point at the 
newly opened, secured port:
 
-            <pre class="line-numbers"><code class="language-text">            
bootstrap.servers = [broker1:9092,...]
-            security.protocol = SSL
-            ...etc</code></pre>
+            <pre class="line-numbers"><code 
class="language-text">bootstrap.servers = [broker1:9092,...]
+security.protocol = SSL
+...etc</code></pre>
 
         In the second incremental server bounce we instruct Kafka to use SSL 
as the broker-broker protocol (which will use the same SSL port):
 
-            <pre class="line-numbers"><code class="language-text">            
listeners=PLAINTEXT://broker1:9091,SSL://broker1:9092
-            security.inter.broker.protocol=SSL</code></pre>
+            <pre class="line-numbers"><code 
class="language-text">listeners=PLAINTEXT://broker1:9091,SSL://broker1:9092
+security.inter.broker.protocol=SSL</code></pre>
 
         In the final bounce we secure the cluster by closing the PLAINTEXT 
port:
 
-            <pre class="line-numbers"><code class="language-text">            
listeners=SSL://broker1:9092
-            security.inter.broker.protocol=SSL</code></pre>
+            <pre class="line-numbers"><code 
class="language-text">listeners=SSL://broker1:9092
+security.inter.broker.protocol=SSL</code></pre>
 
         Alternatively we might choose to open multiple ports so that different 
protocols can be used for broker-broker and broker-client communication. Say we 
wished to use SSL encryption throughout (i.e. for broker-broker and 
broker-client communication) but we'd like to add SASL authentication to the 
broker-client connection also. We would achieve this by opening two additional 
ports during the first bounce:
 
-            <pre class="line-numbers"><code class="language-text">            
listeners=PLAINTEXT://broker1:9091,SSL://broker1:9092,SASL_SSL://broker1:9093</code></pre>
+            <pre class="line-numbers"><code 
class="language-text">listeners=PLAINTEXT://broker1:9091,SSL://broker1:9092,SASL_SSL://broker1:9093</code></pre>
 
         We would then restart the clients, changing their config to point at 
the newly opened, SASL & SSL secured port:
 
-            <pre class="line-numbers"><code class="language-text">            
bootstrap.servers = [broker1:9093,...]
-            security.protocol = SASL_SSL
-            ...etc</code></pre>
+            <pre class="line-numbers"><code 
class="language-text">bootstrap.servers = [broker1:9093,...]
+security.protocol = SASL_SSL
+...etc</code></pre>
 
         The second server bounce would switch the cluster to use encrypted 
broker-broker communication via the SSL port we previously opened on port 9092:
 
-            <pre class="line-numbers"><code class="language-text">            
listeners=PLAINTEXT://broker1:9091,SSL://broker1:9092,SASL_SSL://broker1:9093
-            security.inter.broker.protocol=SSL</code></pre>
+            <pre class="line-numbers"><code 
class="language-text">listeners=PLAINTEXT://broker1:9091,SSL://broker1:9092,SASL_SSL://broker1:9093
+security.inter.broker.protocol=SSL</code></pre>
 
         The final bounce secures the cluster by closing the PLAINTEXT port.
 
-            <pre class="line-numbers"><code class="language-text">        
listeners=SSL://broker1:9092,SASL_SSL://broker1:9093
-        security.inter.broker.protocol=SSL</code></pre>
+            <pre class="line-numbers"><code 
class="language-text">listeners=SSL://broker1:9092,SASL_SSL://broker1:9093
+security.inter.broker.protocol=SSL</code></pre>
 
         ZooKeeper can be secured independently of the Kafka cluster. The steps 
for doing this are covered in section <a href="#zk_authz_migration">7.6.2</a>.
 
@@ -2028,13 +2028,13 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
     Here is a sample (partial) ZooKeeper configuration for enabling TLS 
authentication.
     These configurations are described in the
     <a 
href="https://zookeeper.apache.org/doc/r3.5.7/zookeeperAdmin.html#sc_authOptions";>ZooKeeper
 Admin Guide</a>.
-    <pre class="line-numbers"><code class="language-text">        
secureClientPort=2182
-        serverCnxnFactory=org.apache.zookeeper.server.NettyServerCnxnFactory
-        
authProvider.x509=org.apache.zookeeper.server.auth.X509AuthenticationProvider
-        ssl.keyStore.location=/path/to/zk/keystore.jks
-        ssl.keyStore.password=zk-ks-passwd
-        ssl.trustStore.location=/path/to/zk/truststore.jks
-        ssl.trustStore.password=zk-ts-passwd</code></pre>
+    <pre class="line-numbers"><code class="language-text">secureClientPort=2182
+serverCnxnFactory=org.apache.zookeeper.server.NettyServerCnxnFactory
+authProvider.x509=org.apache.zookeeper.server.auth.X509AuthenticationProvider
+ssl.keyStore.location=/path/to/zk/keystore.jks
+ssl.keyStore.password=zk-ks-passwd
+ssl.trustStore.location=/path/to/zk/truststore.jks
+ssl.trustStore.password=zk-ts-passwd</code></pre>
     <strong>IMPORTANT</strong>: ZooKeeper does not support setting the key 
password in the ZooKeeper server keystore
     to a value different from the keystore password itself.
     Be sure to set the key password to be the same as the keystore password.
@@ -2042,19 +2042,19 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
     <p>Here is a sample (partial) Kafka Broker configuration for connecting to 
ZooKeeper with mTLS authentication.
     These configurations are described above in <a 
href="#brokerconfigs">Broker Configs</a>.
     </p>
-    <pre class="line-numbers"><code class="language-text">        # connect to 
the ZooKeeper port configured for TLS
-        zookeeper.connect=zk1:2182,zk2:2182,zk3:2182
-        # required to use TLS to ZooKeeper (default is false)
-        zookeeper.ssl.client.enable=true
-        # required to use TLS to ZooKeeper
-        zookeeper.clientCnxnSocket=org.apache.zookeeper.ClientCnxnSocketNetty
-        # define key/trust stores to use TLS to ZooKeeper; ignored unless 
zookeeper.ssl.client.enable=true
-        zookeeper.ssl.keystore.location=/path/to/kafka/keystore.jks
-        zookeeper.ssl.keystore.password=kafka-ks-passwd
-        zookeeper.ssl.truststore.location=/path/to/kafka/truststore.jks
-        zookeeper.ssl.truststore.password=kafka-ts-passwd
-        # tell broker to create ACLs on znodes
-        zookeeper.set.acl=true</code></pre>
+    <pre class="line-numbers"><code class="language-text"># connect to the 
ZooKeeper port configured for TLS
+zookeeper.connect=zk1:2182,zk2:2182,zk3:2182
+# required to use TLS to ZooKeeper (default is false)
+zookeeper.ssl.client.enable=true
+# required to use TLS to ZooKeeper
+zookeeper.clientCnxnSocket=org.apache.zookeeper.ClientCnxnSocketNetty
+# define key/trust stores to use TLS to ZooKeeper; ignored unless 
zookeeper.ssl.client.enable=true
+zookeeper.ssl.keystore.location=/path/to/kafka/keystore.jks
+zookeeper.ssl.keystore.password=kafka-ks-passwd
+zookeeper.ssl.truststore.location=/path/to/kafka/truststore.jks
+zookeeper.ssl.truststore.password=kafka-ts-passwd
+# tell broker to create ACLs on znodes
+zookeeper.set.acl=true</code></pre>
     <strong>IMPORTANT</strong>: ZooKeeper does not support setting the key 
password in the ZooKeeper client (i.e. broker) keystore
     to a value different from the keystore password itself.
     Be sure to set the key password to be the same as the keystore password.
@@ -2063,14 +2063,14 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
     If you are running a version of Kafka that does not support security or 
simply with security disabled, and you want to make the cluster secure, then 
you need to execute the following steps to enable ZooKeeper authentication with 
minimal disruption to your operations:
     <ol>
         <li>Enable SASL and/or mTLS authentication on ZooKeeper.  If enabling 
mTLS, you would now have both a non-TLS port and a TLS port, like this:
-            <pre class="line-numbers"><code class="language-text">    
clientPort=2181
-    secureClientPort=2182
-    serverCnxnFactory=org.apache.zookeeper.server.NettyServerCnxnFactory
-    
authProvider.x509=org.apache.zookeeper.server.auth.X509AuthenticationProvider
-    ssl.keyStore.location=/path/to/zk/keystore.jks
-    ssl.keyStore.password=zk-ks-passwd
-    ssl.trustStore.location=/path/to/zk/truststore.jks
-    ssl.trustStore.password=zk-ts-passwd</code></pre>
+            <pre class="line-numbers"><code 
class="language-text">clientPort=2181
+secureClientPort=2182
+serverCnxnFactory=org.apache.zookeeper.server.NettyServerCnxnFactory
+authProvider.x509=org.apache.zookeeper.server.auth.X509AuthenticationProvider
+ssl.keyStore.location=/path/to/zk/keystore.jks
+ssl.keyStore.password=zk-ks-passwd
+ssl.trustStore.location=/path/to/zk/truststore.jks
+ssl.trustStore.password=zk-ts-passwd</code></pre>
         </li>
         <li>Perform a rolling restart of brokers setting the JAAS login file 
and/or defining ZooKeeper mutual TLS configurations (including connecting to 
the TLS-enabled ZooKeeper port) as required, which enables brokers to 
authenticate to ZooKeeper. At the end of the rolling restart, brokers are able 
to manipulate znodes with strict ACLs, but they will not create znodes with 
those ACLs</li>
         <li>If you enabled mTLS, disable the non-TLS port in ZooKeeper</li>
@@ -2086,9 +2086,9 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
         <li>If you are disabling mTLS, disable the TLS port in ZooKeeper</li>
     </ol>
     Here is an example of how to run the migration tool:
-    <pre class="line-numbers"><code class="language-bash">    
bin/zookeeper-security-migration.sh --zookeeper.acl=secure 
--zookeeper.connect=localhost:2181</code></pre>
+    <pre class="line-numbers"><code 
class="language-bash">bin/zookeeper-security-migration.sh 
--zookeeper.acl=secure --zookeeper.connect=localhost:2181</code></pre>
     <p>Run this to see the full list of parameters:</p>
-    <pre class="line-numbers"><code class="language-bash">    
bin/zookeeper-security-migration.sh --help</code></pre>
+    <pre class="line-numbers"><code 
class="language-bash">bin/zookeeper-security-migration.sh --help</code></pre>
     <h4 class="anchor-heading"><a id="zk_authz_ensemble" 
class="anchor-link"></a><a href="#zk_authz_ensemble">7.6.3 Migrating the 
ZooKeeper ensemble</a></h4>
     It is also necessary to enable SASL and/or mTLS authentication on the 
ZooKeeper ensemble. To do it, we need to perform a rolling restart of the 
server and set a few properties. See above for mTLS information.  Please refer 
to the ZooKeeper documentation for more detail:
     <ol>
@@ -2106,18 +2106,18 @@ keyUsage               = digitalSignature, 
keyEncipherment</code></pre>
     and setting this value to <tt>none</tt> in ZooKeeper allows clients to 
connect via a TLS-encrypted connection
     without presenting their own certificate.  Here is a sample (partial) 
Kafka Broker configuration for connecting to ZooKeeper with just TLS encryption.
     These configurations are described above in <a 
href="#brokerconfigs">Broker Configs</a>.
-    <pre class="line-numbers"><code class="language-text">        # connect to 
the ZooKeeper port configured for TLS
-        zookeeper.connect=zk1:2182,zk2:2182,zk3:2182
-        # required to use TLS to ZooKeeper (default is false)
-        zookeeper.ssl.client.enable=true
-        # required to use TLS to ZooKeeper
-        zookeeper.clientCnxnSocket=org.apache.zookeeper.ClientCnxnSocketNetty
-        # define trust stores to use TLS to ZooKeeper; ignored unless 
zookeeper.ssl.client.enable=true
-        # no need to set keystore information assuming ssl.clientAuth=none on 
ZooKeeper
-        zookeeper.ssl.truststore.location=/path/to/kafka/truststore.jks
-        zookeeper.ssl.truststore.password=kafka-ts-passwd
-        # tell broker to create ACLs on znodes (if using SASL authentication, 
otherwise do not set this)
-        zookeeper.set.acl=true</code></pre>
+    <pre class="line-numbers"><code class="language-text"># connect to the 
ZooKeeper port configured for TLS
+zookeeper.connect=zk1:2182,zk2:2182,zk3:2182
+# required to use TLS to ZooKeeper (default is false)
+zookeeper.ssl.client.enable=true
+# required to use TLS to ZooKeeper
+zookeeper.clientCnxnSocket=org.apache.zookeeper.ClientCnxnSocketNetty
+# define trust stores to use TLS to ZooKeeper; ignored unless 
zookeeper.ssl.client.enable=true
+# no need to set keystore information assuming ssl.clientAuth=none on ZooKeeper
+zookeeper.ssl.truststore.location=/path/to/kafka/truststore.jks
+zookeeper.ssl.truststore.password=kafka-ts-passwd
+# tell broker to create ACLs on znodes (if using SASL authentication, 
otherwise do not set this)
+zookeeper.set.acl=true</code></pre>
 </script>
 
 <div class="p-security"></div>

Reply via email to