Not sure if this is the case for you, but if the KDC was never set up to issue renewable tickets, but the principals were already created, you'll have to edit them to issue renewable tickets (KDC change doesn't affect existing principals).
See point 3: https://github.com/apache/metron/blob/master/metron-deployment/Kerberos-manual-setup.md#storm-authorization On Tue, May 16, 2017 at 9:50 AM, Nick Allen <[email protected]> wrote: > Hi Mohan - In my experience you need to setup your KDC so that it can > issue renewable tickets. I would not move beyond that step until you > figure out that issue. > > On Mon, May 15, 2017 at 12:32 PM, Mohan Venkateshaiah < > [email protected]> wrote: > >> Hi All, >> >> >> >> I am enabling kerberos on a 12 node cluster, I did successful >> installation of KDC and set all the required properties in the conf files >> and added the required principals. While enabling Kerberos through wizard >> from Amabari it fails at ‘Start and Test Services “. The task at the >> failure happened is ‘Metron Enrichment Start’. Below is the Trace for the >> same, the exception is >> >> >> >> Caused by: java.lang.RuntimeException: The TGT found is not renewable >> >> >> >> I have set the ‘max_renewable_life = 7d’ In >> /var/kerberos/krb5kdc/kdc.conf in in the realm section, If the KDC cannot >> issue renewable tickets should I remove this property and proceed ? >> >> >> >> >> >> stderr: >> >> Traceback (most recent call last): >> >> File "/var/lib/ambari-agent/cache/common-services/METRON/0.4.0.1. >> 1.0.0/package/scripts/enrichment_master.py", line 113, in <module> >> >> Enrichment().execute() >> >> File >> "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", >> line 280, in execute >> >> method(env) >> >> File "/var/lib/ambari-agent/cache/common-services/METRON/0.4.0.1. >> 1.0.0/package/scripts/enrichment_master.py", line 74, in start >> >> commands.start_enrichment_topology() >> >> File "/var/lib/ambari-agent/cache/common-services/METRON/0.4.0.1. >> 1.0.0/package/scripts/enrichment_commands.py", line 146, in >> start_enrichment_topology >> >> user=self.__params.metron_user) >> >> File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", >> line 155, in __init__ >> >> self.env.run() >> >> File >> "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", >> line 160, in run >> >> self.run_action(resource, action) >> >> File >> "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", >> line 124, in run_action >> >> provider_action() >> >> File >> "/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py", >> line 273, in action_run >> >> tries=self.resource.tries, try_sleep=self.resource.try_sleep) >> >> File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", >> line 70, in inner >> >> result = function(command, **kwargs) >> >> File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", >> line 92, in checked_call >> >> tries=tries, try_sleep=try_sleep) >> >> File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", >> line 140, in _call_wrapper >> >> result = _call(command, **kwargs_copy) >> >> File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", >> line 293, in _call >> >> raise ExecutionFailed(err_msg, code, out, err) >> >> resource_management.core.exceptions.ExecutionFailed: Execution of >> '/usr/hcp/1.1.0.0-71/metron/bin/start_enrichment_topology.sh >> -s enrichment >> -z >> hcpa-11.openstacklocal:2181,hcpa-12.openstacklocal:2181,hcpa-10.openstacklocal:2181' >> returned 1. Running: /usr/jdk64/jdk1.8.0_77/bin/java -server >> -Ddaemon.name= -Dstorm.options= -Dstorm.home=/grid/0/hdp/2.5.3.0-37/storm >> -Dstorm.log.dir=/var/log/storm >> -Djava.library.path=/usr/local/lib:/opt/local/lib:/usr/lib >> -Dstorm.conf.file= -cp /grid/0/hdp/2.5.3.0-37/storm/l >> ib/zookeeper.jar:/grid/0/hdp/2.5.3.0-37/storm/lib/storm-core >> -1.0.1.2.5.3.0-37.jar:/grid/0/hdp/2.5.3.0-37/storm/lib/kryo- >> 3.0.3.jar:/grid/0/hdp/2.5.3.0-37/storm/lib/log4j-slf4j-impl- >> 2.1.jar:/grid/0/hdp/2.5.3.0-37/storm/lib/log4j-core-2.1. >> jar:/grid/0/hdp/2.5.3.0-37/storm/lib/ring-cors-0.1.5.jar: >> /grid/0/hdp/2.5.3.0-37/storm/lib/log4j-api-2.1.jar:/grid/0/ >> hdp/2.5.3.0-37/storm/lib/servlet-api-2.5.jar:/grid/0/hdp/2.5.3. >> 0-37/storm/lib/minlog-1.3.0.jar:/grid/0/hdp/2.5.3.0-37/storm/lib/log4j- >> over-slf4j-1.6.6.jar:/grid/0/hdp/2.5.3.0-37/storm/lib/ >> objenesis-2.1.jar:/grid/0/hdp/2.5.3.0-37/storm/lib/asm-5.0. >> 3.jar:/grid/0/hdp/2.5.3.0-37/storm/lib/clojure-1.7.0.jar:/ >> grid/0/hdp/2.5.3.0-37/storm/lib/storm-rename-hack-1.0.1.2. >> 5.3.0-37.jar:/grid/0/hdp/2.5.3.0-37/storm/lib/disruptor-3. >> 3.2.jar:/grid/0/hdp/2.5.3.0-37/storm/lib/slf4j-api-1.7.7. >> jar:/grid/0/hdp/2.5.3.0-37/storm/lib/reflectasm-1.10.1.jar >> org.apache.storm.daemon.ClientJarTransformerRunner >> org.apache.storm.hack.StormShadeTransformer >> /usr/hcp/1.1.0.0-71/metron/lib/metron-enrichment-0.4.0.1.1.0.0-71-uber.jar >> /tmp/e11b7abc396311e7b0c2fa163e0f2645.jar >> >> Running: /usr/jdk64/jdk1.8.0_77/bin/java -client -Ddaemon.name= >> -Dstorm.options= -Dstorm.home=/grid/0/hdp/2.5.3.0-37/storm >> -Dstorm.log.dir=/var/log/storm >> -Djava.library.path=/usr/local/lib:/opt/local/lib:/usr/lib >> -Dstorm.conf.file= -cp /grid/0/hdp/2.5.3.0-37/storm/l >> ib/zookeeper.jar:/grid/0/hdp/2.5.3.0-37/storm/lib/storm-core >> -1.0.1.2.5.3.0-37.jar:/grid/0/hdp/2.5.3.0-37/storm/lib/kryo- >> 3.0.3.jar:/grid/0/hdp/2.5.3.0-37/storm/lib/log4j-slf4j-impl- >> 2.1.jar:/grid/0/hdp/2.5.3.0-37/storm/lib/log4j-core-2.1. >> jar:/grid/0/hdp/2.5.3.0-37/storm/lib/ring-cors-0.1.5.jar: >> /grid/0/hdp/2.5.3.0-37/storm/lib/log4j-api-2.1.jar:/grid/0/ >> hdp/2.5.3.0-37/storm/lib/servlet-api-2.5.jar:/grid/0/hdp/2.5.3. >> 0-37/storm/lib/minlog-1.3.0.jar:/grid/0/hdp/2.5.3.0-37/storm/lib/log4j- >> over-slf4j-1.6.6.jar:/grid/0/hdp/2.5.3.0-37/storm/lib/ >> objenesis-2.1.jar:/grid/0/hdp/2.5.3.0-37/storm/lib/asm-5.0. >> 3.jar:/grid/0/hdp/2.5.3.0-37/storm/lib/clojure-1.7.0.jar:/ >> grid/0/hdp/2.5.3.0-37/storm/lib/storm-rename-hack-1.0.1.2. >> 5.3.0-37.jar:/grid/0/hdp/2.5.3.0-37/storm/lib/disruptor-3. >> 3.2.jar:/grid/0/hdp/2.5.3.0-37/storm/lib/slf4j-api-1.7.7. >> jar:/grid/0/hdp/2.5.3.0-37/storm/lib/reflectasm-1.10.1. >> jar:/tmp/e11b7abc396311e7b0c2fa163e0f2645.jar:/home/metron/. >> storm:/grid/0/hdp/2.5.3.0-37/storm/bin >> -Dstorm.jar=/tmp/e11b7abc396311e7b0c2fa163e0f2645.jar >> org.apache.storm.flux.Flux --remote >> /usr/hcp/1.1.0.0-71/metron/flux/enrichment/remote.yaml >> --filter /usr/hcp/1.1.0.0-71/metron/config/enrichment.properties >> >> ███████╗██╗ ██╗ ██╗██╗ ██╗ >> >> ██╔════╝██║ ██║ ██║╚██╗██╔╝ >> >> █████╗ ██║ ██║ ██║ ╚███╔╝ >> >> ██╔══╝ ██║ ██║ ██║ ██╔██╗ >> >> ██║ ███████╗╚██████╔╝██╔╝ ██╗ >> >> ╚═╝ ╚══════╝ ╚═════╝ ╚═╝ ╚═╝ >> >> +- Apache Storm -+ >> >> +- data FLow User eXperience -+ >> >> Version: 1.0.1 >> >> Parsing file: /usr/hcp/1.1.0.0-71/metron/flux/enrichment/remote.yaml >> >> 629 [main] INFO o.a.s.f.p.FluxParser - loading YAML from input stream... >> >> 640 [main] INFO o.a.s.f.p.FluxParser - Performing property substitution. >> >> 658 [main] INFO o.a.s.f.p.FluxParser - Not performing environment >> variable substitution. >> >> 982 [main] INFO o.a.c.f.i.CuratorFrameworkImpl - Starting >> >> 1097 [main-EventThread] INFO o.a.c.f.s.ConnectionStateManager - State >> change: CONNECTED >> >> 1439 [main] INFO o.a.s.f.FluxBuilder - Detected DSL topology... >> >> 1728 [main] INFO o.a.s.k.s.KafkaSpoutStream - Declared [streamId = >> default], [outputFields = [value]] for [topic = enrichments] >> >> ---------- TOPOLOGY DETAILS ---------- >> >> Topology Name: enrichment >> >> --------------- SPOUTS --------------- >> >> kafkaSpout [1] (org.apache.metron.storm.kafka.flux.StormKafkaSpout) >> >> ---------------- BOLTS --------------- >> >> enrichmentSplitBolt [1] (org.apache.metron.enrichment. >> bolt.EnrichmentSplitterBolt) >> >> geoEnrichmentBolt [1] (org.apache.metron.enrichment. >> bolt.GenericEnrichmentBolt) >> >> stellarEnrichmentBolt [1] (org.apache.metron.enrichment. >> bolt.GenericEnrichmentBolt) >> >> hostEnrichmentBolt [1] (org.apache.metron.enrichment. >> bolt.GenericEnrichmentBolt) >> >> simpleHBaseEnrichmentBolt [1] (org.apache.metron.enrichment. >> bolt.GenericEnrichmentBolt) >> >> enrichmentJoinBolt [1] (org.apache.metron.enrichment. >> bolt.EnrichmentJoinBolt) >> >> enrichmentErrorOutputBolt [1] (org.apache.metron.writer.bolt >> .BulkMessageWriterBolt) >> >> threatIntelSplitBolt [1] (org.apache.metron.enrichment. >> bolt.ThreatIntelSplitterBolt) >> >> simpleHBaseThreatIntelBolt [1] (org.apache.metron.enrichment. >> bolt.GenericEnrichmentBolt) >> >> stellarThreatIntelBolt [1] (org.apache.metron.enrichment. >> bolt.GenericEnrichmentBolt) >> >> threatIntelJoinBolt [1] (org.apache.metron.enrichment. >> bolt.ThreatIntelJoinBolt) >> >> threatIntelErrorOutputBolt [1] (org.apache.metron.writer.bolt >> .BulkMessageWriterBolt) >> >> outputBolt [1] (org.apache.metron.writer.bolt.BulkMessageWriterBolt) >> >> --------------- STREAMS --------------- >> >> kafkaSpout --SHUFFLE--> enrichmentSplitBolt >> >> enrichmentSplitBolt --FIELDS--> hostEnrichmentBolt >> >> enrichmentSplitBolt --FIELDS--> geoEnrichmentBolt >> >> enrichmentSplitBolt --FIELDS--> stellarEnrichmentBolt >> >> enrichmentSplitBolt --FIELDS--> simpleHBaseEnrichmentBolt >> >> enrichmentSplitBolt --FIELDS--> enrichmentJoinBolt >> >> geoEnrichmentBolt --FIELDS--> enrichmentJoinBolt >> >> stellarEnrichmentBolt --FIELDS--> enrichmentJoinBolt >> >> simpleHBaseEnrichmentBolt --FIELDS--> enrichmentJoinBolt >> >> hostEnrichmentBolt --FIELDS--> enrichmentJoinBolt >> >> geoEnrichmentBolt --FIELDS--> enrichmentErrorOutputBolt >> >> stellarEnrichmentBolt --FIELDS--> enrichmentErrorOutputBolt >> >> hostEnrichmentBolt --FIELDS--> enrichmentErrorOutputBolt >> >> simpleHBaseEnrichmentBolt --FIELDS--> enrichmentErrorOutputBolt >> >> enrichmentJoinBolt --FIELDS--> threatIntelSplitBolt >> >> threatIntelSplitBolt --FIELDS--> simpleHBaseThreatIntelBolt >> >> threatIntelSplitBolt --FIELDS--> stellarThreatIntelBolt >> >> simpleHBaseThreatIntelBolt --FIELDS--> threatIntelJoinBolt >> >> stellarThreatIntelBolt --FIELDS--> threatIntelJoinBolt >> >> threatIntelSplitBolt --FIELDS--> threatIntelJoinBolt >> >> threatIntelJoinBolt --FIELDS--> outputBolt >> >> simpleHBaseThreatIntelBolt --FIELDS--> threatIntelErrorOutputBolt >> >> stellarThreatIntelBolt --FIELDS--> threatIntelErrorOutputBolt >> >> -------------------------------------- >> >> 1752 [main] INFO o.a.s.f.Flux - Running remotely... >> >> 1752 [main] INFO o.a.s.f.Flux - Deploying topology in an ACTIVE state... >> >> 1768 [main] INFO o.a.s.StormSubmitter - Generated ZooKeeper secret >> payload for MD5-digest: -7795983889088274963:-7645329918562802951 >> >> 1851 [main] INFO o.a.s.s.a.AuthUtils - Got AutoCreds >> [org.apache.storm.security.auth.kerberos.AutoTGT@798256c5] >> >> 1851 [main] INFO o.a.s.StormSubmitter - Running >> org.apache.storm.security.auth.kerberos.AutoTGT@798256c5 >> >> Exception in thread "main" java.lang.RuntimeException: >> java.lang.RuntimeException: The TGT found is not renewable >> >> at org.apache.storm.security.auth >> .kerberos.AutoTGT.populateCredentials(AutoTGT.java:103) >> >> at org.apache.storm.StormSubmitte >> r.populateCredentials(StormSubmitter.java:94) >> >> at org.apache.storm.StormSubmitte >> r.submitTopologyAs(StormSubmitter.java:214) >> >> at org.apache.storm.StormSubmitte >> r.submitTopology(StormSubmitter.java:310) >> >> at org.apache.storm.flux.Flux.runCli(Flux.java:171) >> >> at org.apache.storm.flux.Flux.main(Flux.java:98) >> >> Caused by: java.lang.RuntimeException: The TGT found is not renewable >> >> at org.apache.storm.security.auth >> .kerberos.AutoTGT.populateCredentials(AutoTGT.java:94) >> >> ... 5 more >> >> stdout: >> >> 2017-05-15 11:44:35,171 - Using hadoop conf dir: >> /usr/hdp/current/hadoop-client/conf >> >> 2017-05-15 11:44:35,314 - Using hadoop conf dir: >> /usr/hdp/current/hadoop-client/conf >> >> 2017-05-15 11:44:35,316 - Group['metron'] {} >> >> 2017-05-15 11:44:35,317 - Group['livy'] {} >> >> 2017-05-15 11:44:35,317 - Group['elasticsearch'] {} >> >> 2017-05-15 11:44:35,317 - Group['spark'] {} >> >> 2017-05-15 11:44:35,317 - Group['zeppelin'] {} >> >> 2017-05-15 11:44:35,317 - Group['hadoop'] {} >> >> 2017-05-15 11:44:35,318 - Group['kibana'] {} >> >> 2017-05-15 11:44:35,318 - Group['users'] {} >> >> 2017-05-15 11:44:35,318 - User['hive'] {'gid': 'hadoop', >> 'fetch_nonlocal_groups': True, 'groups': [u'hadoop']} >> >> 2017-05-15 11:44:35,319 - User['storm'] {'gid': 'hadoop', >> 'fetch_nonlocal_groups': True, 'groups': [u'hadoop']} >> >> 2017-05-15 11:44:35,319 - User['zookeeper'] {'gid': 'hadoop', >> 'fetch_nonlocal_groups': True, 'groups': [u'hadoop']} >> >> 2017-05-15 11:44:35,320 - User['tez'] {'gid': 'hadoop', >> 'fetch_nonlocal_groups': True, 'groups': [u'users']} >> >> 2017-05-15 11:44:35,320 - User['zeppelin'] {'gid': 'hadoop', >> 'fetch_nonlocal_groups': True, 'groups': [u'hadoop']} >> >> 2017-05-15 11:44:35,321 - User['metron'] {'gid': 'hadoop', >> 'fetch_nonlocal_groups': True, 'groups': [u'hadoop']} >> >> 2017-05-15 11:44:35,321 - User['livy'] {'gid': 'hadoop', >> 'fetch_nonlocal_groups': True, 'groups': [u'hadoop']} >> >> 2017-05-15 11:44:35,322 - User['elasticsearch'] {'gid': 'hadoop', >> 'fetch_nonlocal_groups': True, 'groups': [u'hadoop']} >> >> 2017-05-15 11:44:35,323 - User['spark'] {'gid': 'hadoop', >> 'fetch_nonlocal_groups': True, 'groups': [u'hadoop']} >> >> 2017-05-15 11:44:35,323 - User['ambari-qa'] {'gid': 'hadoop', >> 'fetch_nonlocal_groups': True, 'groups': [u'users']} >> >> 2017-05-15 11:44:35,324 - User['kafka'] {'gid': 'hadoop', >> 'fetch_nonlocal_groups': True, 'groups': [u'hadoop']} >> >> 2017-05-15 11:44:35,324 - User['hdfs'] {'gid': 'hadoop', >> 'fetch_nonlocal_groups': True, 'groups': [u'hadoop']} >> >> 2017-05-15 11:44:35,325 - User['yarn'] {'gid': 'hadoop', >> 'fetch_nonlocal_groups': True, 'groups': [u'hadoop']} >> >> 2017-05-15 11:44:35,325 - User['kibana'] {'gid': 'hadoop', >> 'fetch_nonlocal_groups': True, 'groups': [u'hadoop']} >> >> 2017-05-15 11:44:35,326 - User['mapred'] {'gid': 'hadoop', >> 'fetch_nonlocal_groups': True, 'groups': [u'hadoop']} >> >> 2017-05-15 11:44:35,326 - User['hbase'] {'gid': 'hadoop', >> 'fetch_nonlocal_groups': True, 'groups': [u'hadoop']} >> >> 2017-05-15 11:44:35,327 - User['hcat'] {'gid': 'hadoop', >> 'fetch_nonlocal_groups': True, 'groups': [u'hadoop']} >> >> 2017-05-15 11:44:35,328 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] >> {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555} >> >> 2017-05-15 11:44:35,329 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh >> ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari >> -qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa'] {'not_if': '(test $(id -u >> ambari-qa) -gt 1000) || (false)'} >> >> 2017-05-15 11:44:35,344 - Skipping >> Execute['/var/lib/ambari-agent/tmp/changeUid.sh >> ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari >> -qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa'] due to not_if >> >> 2017-05-15 11:44:35,344 - Directory['/tmp/hbase-hbase'] {'owner': >> 'hbase', 'create_parents': True, 'mode': 0775, 'cd_access': 'a'} >> >> 2017-05-15 11:44:35,345 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] >> {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555} >> >> 2017-05-15 11:44:35,346 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh >> hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase'] >> {'not_if': '(test $(id -u hbase) -gt 1000) || (false)'} >> >> 2017-05-15 11:44:35,363 - Skipping >> Execute['/var/lib/ambari-agent/tmp/changeUid.sh >> hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase'] >> due to not_if >> >> 2017-05-15 11:44:35,364 - Group['hdfs'] {} >> >> 2017-05-15 11:44:35,364 - User['hdfs'] {'fetch_nonlocal_groups': True, >> 'groups': [u'hadoop', u'hdfs']} >> >> 2017-05-15 11:44:35,364 - FS Type: >> >> 2017-05-15 11:44:35,365 - Directory['/etc/hadoop'] {'mode': 0755} >> >> 2017-05-15 11:44:35,379 - >> File['/usr/hdp/current/hadoop-client/conf/hadoop-env.sh'] >> {'content': InlineTemplate(...), 'owner': 'root', 'group': 'hadoop'} >> >> 2017-05-15 11:44:35,379 - >> Directory['/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir'] >> {'owner': 'hdfs', 'group': 'hadoop', 'mode': 01777} >> >> 2017-05-15 11:44:35,396 - Execute[('setenforce', '0')] {'not_if': '(! >> which getenforce ) || (which getenforce && getenforce | grep -q Disabled)', >> 'sudo': True, 'only_if': 'test -f /selinux/enforce'} >> >> 2017-05-15 11:44:35,429 - Skipping Execute[('setenforce', '0')] due to >> only_if >> >> 2017-05-15 11:44:35,430 - Directory['/var/log/hadoop'] {'owner': 'root', >> 'create_parents': True, 'group': 'hadoop', 'mode': 0775, 'cd_access': 'a'} >> >> 2017-05-15 11:44:35,433 - Directory['/var/run/hadoop'] {'owner': 'root', >> 'create_parents': True, 'group': 'root', 'cd_access': 'a'} >> >> 2017-05-15 11:44:35,434 - Directory['/tmp/hadoop-hdfs'] {'owner': 'hdfs', >> 'create_parents': True, 'cd_access': 'a'} >> >> 2017-05-15 11:44:35,441 - File['/usr/hdp/current/hadoop-client/conf/ >> commons-logging.properties'] {'content': >> Template('commons-logging.properties.j2'), >> 'owner': 'root'} >> >> 2017-05-15 11:44:35,442 - >> File['/usr/hdp/current/hadoop-client/conf/health_check'] >> {'content': Template('health_check.j2'), 'owner': 'root'} >> >> 2017-05-15 11:44:35,443 - >> File['/usr/hdp/current/hadoop-client/conf/log4j.properties'] >> {'content': ..., 'owner': 'hdfs', 'group': 'hadoop', 'mode': 0644} >> >> 2017-05-15 11:44:35,454 - File['/usr/hdp/current/hadoop-client/conf/ >> hadoop-metrics2.properties'] {'content': >> Template('hadoop-metrics2.properties.j2'), >> 'owner': 'hdfs', 'group': 'hadoop'} >> >> 2017-05-15 11:44:35,455 - File['/usr/hdp/current/hadoop- >> client/conf/task-log4j.properties'] {'content': >> StaticFile('task-log4j.properties'), 'mode': 0755} >> >> 2017-05-15 11:44:35,456 - File['/usr/hdp/current/hadoop- >> client/conf/configuration.xsl'] {'owner': 'hdfs', 'group': 'hadoop'} >> >> 2017-05-15 11:44:35,460 - File['/etc/hadoop/conf/topology_mappings.data'] >> {'owner': 'hdfs', 'content': Template('topology_mappings.data.j2'), >> 'only_if': 'test -d /etc/hadoop/conf', 'group': 'hadoop'} >> >> 2017-05-15 11:44:35,477 - File['/etc/hadoop/conf/topology_script.py'] >> {'content': StaticFile('topology_script.py'), 'only_if': 'test -d >> /etc/hadoop/conf', 'mode': 0755} >> >> 2017-05-15 11:44:35,709 - Using hadoop conf dir: >> /usr/hdp/current/hadoop-client/conf >> >> 2017-05-15 11:44:35,711 - Running enrichment configure >> >> 2017-05-15 11:44:35,717 - File['/usr/hcp/1.1.0.0-71/metr >> on/config/enrichment.properties'] {'owner': 'metron', 'content': >> Template('enrichment.properties.j2'), 'group': 'metron'} >> >> 2017-05-15 11:44:35,719 - Calling security setup >> >> 2017-05-15 11:44:35,720 - Directory['/usr/hcp/1.1.0.0-71/metron'] >> {'owner': 'metron', 'group': 'metron', 'create_parents': True, 'mode': 0755} >> >> 2017-05-15 11:44:35,720 - Directory['/home/metron/.storm'] {'owner': >> 'metron', 'group': 'metron', 'mode': 0755} >> >> 2017-05-15 11:44:35,722 - File['/usr/hcp/1.1.0.0-71/metron/client_jaas.conf'] >> {'owner': 'metron', 'content': Template('client_jaas.conf.j2'), 'group': >> 'metron', 'mode': 0755} >> >> 2017-05-15 11:44:35,724 - File['/home/metron/.storm/storm.yaml'] >> {'owner': 'metron', 'content': Template('storm.yaml.j2'), 'group': >> 'metron', 'mode': 0755} >> >> 2017-05-15 11:44:35,725 - File['/home/metron/.storm/storm.config'] >> {'owner': 'metron', 'content': Template('storm.config.j2'), 'group': >> 'metron', 'mode': 0755} >> >> 2017-05-15 11:44:35,726 - kinit command: /usr/bin/kinit -kt >> /etc/security/keytabs/metron.headless.keytab [email protected]; as >> user: metron >> >> 2017-05-15 11:44:35,726 - Execute['/usr/bin/kinit -kt >> /etc/security/keytabs/metron.headless.keytab [email protected]; '] >> {'user': 'metron'} >> >> 2017-05-15 11:44:35,807 - Create Metron Local Config Directory >> >> 2017-05-15 11:44:35,807 - Configure Metron global.json >> >> 2017-05-15 11:44:35,807 - >> Directory['/usr/hcp/1.1.0.0-71/metron/config/zookeeper'] >> {'owner': 'metron', 'group': 'metron', 'mode': 0755} >> >> 2017-05-15 11:44:35,810 - File['/usr/hcp/1.1.0.0-71/metr >> on/config/zookeeper/global.json'] {'content': InlineTemplate(...), >> 'owner': 'metron'} >> >> 2017-05-15 11:44:35,813 - File['/usr/hcp/1.1.0.0-71/metr >> on/config/zookeeper/../elasticsearch.properties'] {'content': >> InlineTemplate(...), 'owner': 'metron'} >> >> 2017-05-15 11:44:35,814 - Loading config into ZooKeeper >> >> 2017-05-15 11:44:35,814 - >> Execute['/usr/hcp/1.1.0.0-71/metron/bin/zk_load_configs.sh >> --mode PUSH -i /usr/hcp/1.1.0.0-71/metron/config/zookeeper -z >> hcpa-11.openstacklocal:2181,hcpa-12.openstacklocal:2181,hcpa-10.openstacklocal:2181'] >> {'path': [u'/usr/jdk64/jdk1.8.0_77/bin']} >> >> 2017-05-15 11:44:37,486 - Starting Metron enrichment topology: enrichment >> >> 2017-05-15 11:44:37,486 - Starting enrichment >> >> 2017-05-15 11:44:37,486 - Execute['/usr/hcp/1.1.0.0-71/m >> etron/bin/start_enrichment_topology.sh >> -s enrichment -z >> hcpa-11.openstacklocal:2181,hcpa-12.openstacklocal:2181,hcpa-10.openstacklocal:2181'] >> {'user': 'metron'} >> >> >> >> Command failed after 1 tries >> >> >> >> >> >> Thanks >> >> Mohan DV >> >> >> > >
