Nick Allen created METRON-894:
---------------------------------
Summary: Ambari "Restart Metron Parsers" Fails If YAF Parser Not
Running
Key: METRON-894
URL: https://issues.apache.org/jira/browse/METRON-894
Project: Metron
Issue Type: Bug
Affects Versions: 0.3.1
Reporter: Nick Allen
Priority: Minor
The "Restart Metron Parsers" action failed in Ambari. It failed because the
"stop" portion of the "restart" failed because the YAF topology was not
running. This should not be treated as an error condition.
I was able to work around this by simply using a "start" operation instead of a
"restart".
{code}
stderr: /var/lib/ambari-agent/data/errors-966.txt
Traceback (most recent call last):
File
"/var/lib/ambari-agent/cache/common-services/METRON/0.4.0/package/scripts/parser_master.py",
line 93, in <module>
ParserMaster().execute()
File
"/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py",
line 280, in execute
method(env)
File
"/var/lib/ambari-agent/cache/common-services/METRON/0.4.0/package/scripts/parser_master.py",
line 81, in restart
commands.restart_parser_topologies(env)
File
"/var/lib/ambari-agent/cache/common-services/METRON/0.4.0/package/scripts/parser_commands.py",
line 146, in restart_parser_topologies
self.stop_parser_topologies()
File
"/var/lib/ambari-agent/cache/common-services/METRON/0.4.0/package/scripts/parser_commands.py",
line 141, in stop_parser_topologies
Execute(stop_cmd, user=self.__params.metron_user)
File "/usr/lib/python2.6/site-packages/resource_management/core/base.py",
line 155, in __init__
self.env.run()
File
"/usr/lib/python2.6/site-packages/resource_management/core/environment.py",
line 160, in run
self.run_action(resource, action)
File
"/usr/lib/python2.6/site-packages/resource_management/core/environment.py",
line 124, in run_action
provider_action()
File
"/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py",
line 273, in action_run
tries=self.resource.tries, try_sleep=self.resource.try_sleep)
File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py",
line 70, in inner
result = function(command, **kwargs)
File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py",
line 92, in checked_call
tries=tries, try_sleep=try_sleep)
File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py",
line 140, in _call_wrapper
result = _call(command, **kwargs_copy)
File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py",
line 293, in _call
raise ExecutionFailed(err_msg, code, out, err)
resource_management.core.exceptions.ExecutionFailed: Execution of 'storm kill
yaf' returned 1. Running: /usr/jdk64/jdk1.8.0_77/bin/java -client
-Ddaemon.name= -Dstorm.options= -Dstorm.home=/usr/hdp/2.5.3.0-37/storm
-Dstorm.log.dir=/var/log/storm
-Djava.library.path=/usr/local/lib:/opt/local/lib:/usr/lib:/usr/hdp/current/storm-client/lib
-Dstorm.conf.file= -cp
/usr/hdp/2.5.3.0-37/storm/lib/clojure-1.7.0.jar:/usr/hdp/2.5.3.0-37/storm/lib/disruptor-3.3.2.jar:/usr/hdp/2.5.3.0-37/storm/lib/log4j-slf4j-impl-2.1.jar:/usr/hdp/2.5.3.0-37/storm/lib/storm-rename-hack-1.0.1.2.5.3.0-37.jar:/usr/hdp/2.5.3.0-37/storm/lib/log4j-api-2.1.jar:/usr/hdp/2.5.3.0-37/storm/lib/ring-cors-0.1.5.jar:/usr/hdp/2.5.3.0-37/storm/lib/log4j-core-2.1.jar:/usr/hdp/2.5.3.0-37/storm/lib/asm-5.0.3.jar:/usr/hdp/2.5.3.0-37/storm/lib/log4j-over-slf4j-1.6.6.jar:/usr/hdp/2.5.3.0-37/storm/lib/slf4j-api-1.7.7.jar:/usr/hdp/2.5.3.0-37/storm/lib/servlet-api-2.5.jar:/usr/hdp/2.5.3.0-37/storm/lib/zookeeper.jar:/usr/hdp/2.5.3.0-37/storm/lib/minlog-1.3.0.jar:/usr/hdp/2.5.3.0-37/storm/lib/kryo-3.0.3.jar:/usr/hdp/2.5.3.0-37/storm/lib/storm-core-1.0.1.2.5.3.0-37.jar:/usr/hdp/2.5.3.0-37/storm/lib/reflectasm-1.10.1.jar:/usr/hdp/2.5.3.0-37/storm/lib/objenesis-2.1.jar:/usr/hdp/2.5.3.0-37/storm/lib/ambari-metrics-storm-sink.jar:/usr/hdp/2.5.3.0-37/storm/extlib-daemon/ranger-storm-plugin-shim-0.6.0.2.5.3.0-37.jar:/usr/hdp/2.5.3.0-37/storm/extlib-daemon/ojdbc6.jar:/usr/hdp/2.5.3.0-37/storm/extlib-daemon/ranger-plugin-classloader-0.6.0.2.5.3.0-37.jar:/usr/hdp/current/storm-supervisor/conf:/usr/hdp/2.5.3.0-37/storm/bin
org.apache.storm.command.kill_topology yaf
Exception in thread "main" NotAliveException(msg:yaf is not alive)
at
org.apache.storm.generated.Nimbus$killTopologyWithOpts_result$killTopologyWithOpts_resultStandardScheme.read(Nimbus.java:10748)
at
org.apache.storm.generated.Nimbus$killTopologyWithOpts_result$killTopologyWithOpts_resultStandardScheme.read(Nimbus.java:10734)
at
org.apache.storm.generated.Nimbus$killTopologyWithOpts_result.read(Nimbus.java:10676)
at
org.apache.storm.thrift.TServiceClient.receiveBase(TServiceClient.java:86)
at
org.apache.storm.generated.Nimbus$Client.recv_killTopologyWithOpts(Nimbus.java:383)
at
org.apache.storm.generated.Nimbus$Client.killTopologyWithOpts(Nimbus.java:369)
at
org.apache.storm.command.kill_topology$_main.doInvoke(kill_topology.clj:27)
at clojure.lang.RestFn.applyTo(RestFn.java:137)
at org.apache.storm.command.kill_topology.main(Unknown Source)
stdout: /var/lib/ambari-agent/data/output-966.txt
2017-04-26 18:21:46,880 - The hadoop conf dir
/usr/hdp/current/hadoop-client/conf exists, will call conf-select on it for
version 2.5.3.0-37
2017-04-26 18:21:46,882 - Checking if need to create versioned conf dir
/etc/hadoop/2.5.3.0-37/0
2017-04-26 18:21:46,884 - call[('ambari-python-wrap', u'/usr/bin/conf-select',
'create-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.3.0-37',
'--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False,
'stderr': -1}
2017-04-26 18:21:46,921 - call returned (1, '/etc/hadoop/2.5.3.0-37/0 exist
already', '')
2017-04-26 18:21:46,922 - checked_call[('ambari-python-wrap',
u'/usr/bin/conf-select', 'set-conf-dir', '--package', 'hadoop',
'--stack-version', '2.5.3.0-37', '--conf-version', '0')] {'logoutput': False,
'sudo': True, 'quiet': False}
2017-04-26 18:21:46,960 - checked_call returned (0, '')
2017-04-26 18:21:46,962 - Ensuring that hadoop has the correct symlink structure
2017-04-26 18:21:46,962 - Using hadoop conf dir:
/usr/hdp/current/hadoop-client/conf
2017-04-26 18:21:47,150 - The hadoop conf dir
/usr/hdp/current/hadoop-client/conf exists, will call conf-select on it for
version 2.5.3.0-37
2017-04-26 18:21:47,152 - Checking if need to create versioned conf dir
/etc/hadoop/2.5.3.0-37/0
2017-04-26 18:21:47,155 - call[('ambari-python-wrap', u'/usr/bin/conf-select',
'create-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.3.0-37',
'--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False,
'stderr': -1}
2017-04-26 18:21:47,193 - call returned (1, '/etc/hadoop/2.5.3.0-37/0 exist
already', '')
2017-04-26 18:21:47,194 - checked_call[('ambari-python-wrap',
u'/usr/bin/conf-select', 'set-conf-dir', '--package', 'hadoop',
'--stack-version', '2.5.3.0-37', '--conf-version', '0')] {'logoutput': False,
'sudo': True, 'quiet': False}
2017-04-26 18:21:47,232 - checked_call returned (0, '')
2017-04-26 18:21:47,233 - Ensuring that hadoop has the correct symlink structure
2017-04-26 18:21:47,233 - Using hadoop conf dir:
/usr/hdp/current/hadoop-client/conf
2017-04-26 18:21:47,235 - Group['metron'] {}
2017-04-26 18:21:47,238 - Group['livy'] {}
2017-04-26 18:21:47,238 - Group['elasticsearch'] {}
2017-04-26 18:21:47,238 - Group['spark'] {}
2017-04-26 18:21:47,239 - Group['zeppelin'] {}
2017-04-26 18:21:47,239 - Group['hadoop'] {}
2017-04-26 18:21:47,239 - Group['kibana'] {}
2017-04-26 18:21:47,240 - Group['users'] {}
2017-04-26 18:21:47,240 - User['hive'] {'gid': 'hadoop',
'fetch_nonlocal_groups': True, 'groups': [u'hadoop']}
2017-04-26 18:21:47,242 - User['storm'] {'gid': 'hadoop',
'fetch_nonlocal_groups': True, 'groups': [u'hadoop']}
2017-04-26 18:21:47,243 - User['zookeeper'] {'gid': 'hadoop',
'fetch_nonlocal_groups': True, 'groups': [u'hadoop']}
2017-04-26 18:21:47,244 - User['ams'] {'gid': 'hadoop',
'fetch_nonlocal_groups': True, 'groups': [u'hadoop']}
2017-04-26 18:21:47,245 - User['tez'] {'gid': 'hadoop',
'fetch_nonlocal_groups': True, 'groups': [u'users']}
2017-04-26 18:21:47,246 - User['zeppelin'] {'gid': 'hadoop',
'fetch_nonlocal_groups': True, 'groups': [u'hadoop']}
2017-04-26 18:21:47,247 - User['metron'] {'gid': 'hadoop',
'fetch_nonlocal_groups': True, 'groups': [u'hadoop']}
2017-04-26 18:21:47,248 - User['livy'] {'gid': 'hadoop',
'fetch_nonlocal_groups': True, 'groups': [u'hadoop']}
2017-04-26 18:21:47,248 - User['elasticsearch'] {'gid': 'hadoop',
'fetch_nonlocal_groups': True, 'groups': [u'hadoop']}
2017-04-26 18:21:47,249 - User['spark'] {'gid': 'hadoop',
'fetch_nonlocal_groups': True, 'groups': [u'hadoop']}
2017-04-26 18:21:47,250 - User['ambari-qa'] {'gid': 'hadoop',
'fetch_nonlocal_groups': True, 'groups': [u'users']}
2017-04-26 18:21:47,251 - User['kafka'] {'gid': 'hadoop',
'fetch_nonlocal_groups': True, 'groups': [u'hadoop']}
2017-04-26 18:21:47,252 - User['hdfs'] {'gid': 'hadoop',
'fetch_nonlocal_groups': True, 'groups': [u'hadoop']}
2017-04-26 18:21:47,253 - User['yarn'] {'gid': 'hadoop',
'fetch_nonlocal_groups': True, 'groups': [u'hadoop']}
2017-04-26 18:21:47,254 - User['kibana'] {'gid': 'hadoop',
'fetch_nonlocal_groups': True, 'groups': [u'hadoop']}
2017-04-26 18:21:47,255 - User['mapred'] {'gid': 'hadoop',
'fetch_nonlocal_groups': True, 'groups': [u'hadoop']}
2017-04-26 18:21:47,256 - User['hbase'] {'gid': 'hadoop',
'fetch_nonlocal_groups': True, 'groups': [u'hadoop']}
2017-04-26 18:21:47,257 - User['hcat'] {'gid': 'hadoop',
'fetch_nonlocal_groups': True, 'groups': [u'hadoop']}
2017-04-26 18:21:47,258 - File['/var/lib/ambari-agent/tmp/changeUid.sh']
{'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
2017-04-26 18:21:47,261 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh
ambari-qa
/tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa']
{'not_if': '(test $(id -u ambari-qa) -gt 1000) || (false)'}
2017-04-26 18:21:47,269 - Skipping
Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa
/tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa']
due to not_if
2017-04-26 18:21:47,270 - Directory['/tmp/hbase-hbase'] {'owner': 'hbase',
'create_parents': True, 'mode': 0775, 'cd_access': 'a'}
2017-04-26 18:21:47,272 - File['/var/lib/ambari-agent/tmp/changeUid.sh']
{'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
2017-04-26 18:21:47,274 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh hbase
/home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase']
{'not_if': '(test $(id -u hbase) -gt 1000) || (false)'}
2017-04-26 18:21:47,281 - Skipping
Execute['/var/lib/ambari-agent/tmp/changeUid.sh hbase
/home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase'] due to
not_if
2017-04-26 18:21:47,282 - Group['hdfs'] {}
2017-04-26 18:21:47,283 - User['hdfs'] {'fetch_nonlocal_groups': True,
'groups': [u'hadoop', u'hdfs']}
2017-04-26 18:21:47,284 - FS Type:
2017-04-26 18:21:47,284 - Directory['/etc/hadoop'] {'mode': 0755}
2017-04-26 18:21:47,308 -
File['/usr/hdp/current/hadoop-client/conf/hadoop-env.sh'] {'content':
InlineTemplate(...), 'owner': 'hdfs', 'group': 'hadoop'}
2017-04-26 18:21:47,310 -
Directory['/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir'] {'owner': 'hdfs',
'group': 'hadoop', 'mode': 01777}
2017-04-26 18:21:47,330 - Execute[('setenforce', '0')] {'not_if': '(! which
getenforce ) || (which getenforce && getenforce | grep -q Disabled)', 'sudo':
True, 'only_if': 'test -f /selinux/enforce'}
2017-04-26 18:21:47,341 - Skipping Execute[('setenforce', '0')] due to not_if
2017-04-26 18:21:47,342 - Directory['/var/log/hadoop'] {'owner': 'root',
'create_parents': True, 'group': 'hadoop', 'mode': 0775, 'cd_access': 'a'}
2017-04-26 18:21:47,346 - Directory['/var/run/hadoop'] {'owner': 'root',
'create_parents': True, 'group': 'root', 'cd_access': 'a'}
2017-04-26 18:21:47,346 - Directory['/tmp/hadoop-hdfs'] {'owner': 'hdfs',
'create_parents': True, 'cd_access': 'a'}
2017-04-26 18:21:47,354 -
File['/usr/hdp/current/hadoop-client/conf/commons-logging.properties']
{'content': Template('commons-logging.properties.j2'), 'owner': 'hdfs'}
2017-04-26 18:21:47,357 -
File['/usr/hdp/current/hadoop-client/conf/health_check'] {'content':
Template('health_check.j2'), 'owner': 'hdfs'}
2017-04-26 18:21:47,358 -
File['/usr/hdp/current/hadoop-client/conf/log4j.properties'] {'content': ...,
'owner': 'hdfs', 'group': 'hadoop', 'mode': 0644}
2017-04-26 18:21:47,377 -
File['/usr/hdp/current/hadoop-client/conf/hadoop-metrics2.properties']
{'content': Template('hadoop-metrics2.properties.j2'), 'owner': 'hdfs',
'group': 'hadoop'}
2017-04-26 18:21:47,378 -
File['/usr/hdp/current/hadoop-client/conf/task-log4j.properties'] {'content':
StaticFile('task-log4j.properties'), 'mode': 0755}
2017-04-26 18:21:47,379 -
File['/usr/hdp/current/hadoop-client/conf/configuration.xsl'] {'owner': 'hdfs',
'group': 'hadoop'}
2017-04-26 18:21:47,386 - File['/etc/hadoop/conf/topology_mappings.data']
{'owner': 'hdfs', 'content': Template('topology_mappings.data.j2'), 'only_if':
'test -d /etc/hadoop/conf', 'group': 'hadoop'}
2017-04-26 18:21:47,391 - File['/etc/hadoop/conf/topology_script.py']
{'content': StaticFile('topology_script.py'), 'only_if': 'test -d
/etc/hadoop/conf', 'mode': 0755}
2017-04-26 18:21:47,682 - The hadoop conf dir
/usr/hdp/current/hadoop-client/conf exists, will call conf-select on it for
version 2.5.3.0-37
2017-04-26 18:21:47,684 - Checking if need to create versioned conf dir
/etc/hadoop/2.5.3.0-37/0
2017-04-26 18:21:47,687 - call[('ambari-python-wrap', u'/usr/bin/conf-select',
'create-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.3.0-37',
'--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False,
'stderr': -1}
2017-04-26 18:21:47,726 - call returned (1, '/etc/hadoop/2.5.3.0-37/0 exist
already', '')
2017-04-26 18:21:47,727 - checked_call[('ambari-python-wrap',
u'/usr/bin/conf-select', 'set-conf-dir', '--package', 'hadoop',
'--stack-version', '2.5.3.0-37', '--conf-version', '0')] {'logoutput': False,
'sudo': True, 'quiet': False}
2017-04-26 18:21:47,766 - checked_call returned (0, '')
2017-04-26 18:21:47,767 - Ensuring that hadoop has the correct symlink structure
2017-04-26 18:21:47,767 - Using hadoop conf dir:
/usr/hdp/current/hadoop-client/conf
2017-04-26 18:21:47,771 - Create Metron Local Config Directory
2017-04-26 18:21:47,771 - Configure Metron global.json
2017-04-26 18:21:47,771 - Directory['/usr/metron/0.4.0/config/zookeeper']
{'owner': 'metron', 'group': 'metron', 'mode': 0755}
2017-04-26 18:21:47,781 -
File['/usr/metron/0.4.0/config/zookeeper/global.json'] {'content':
InlineTemplate(...), 'owner': 'metron'}
2017-04-26 18:21:47,786 -
File['/usr/metron/0.4.0/config/zookeeper/../elasticsearch.properties']
{'content': InlineTemplate(...), 'owner': 'metron'}
2017-04-26 18:21:47,787 - Loading config into ZooKeeper
2017-04-26 18:21:47,787 - Execute['/usr/metron/0.4.0/bin/zk_load_configs.sh
--mode PUSH -i /usr/metron/0.4.0/config/zookeeper -z
y113.l42scl.hortonworks.com:2181,y114.l42scl.hortonworks.com:2181,y115.l42scl.hortonworks.com:2181']
{'path': [u'/usr/jdk64/jdk1.8.0_77/bin']}
2017-04-26 18:21:49,396 - Calling security setup
2017-04-26 18:21:49,397 - Restarting the parser topologies
2017-04-26 18:21:49,397 - Stopping parsers
2017-04-26 18:21:49,397 - Stopping bro
2017-04-26 18:21:49,397 - Execute['storm kill bro'] {'user': 'metron'}
2017-04-26 18:21:55,400 - Stopping snort
2017-04-26 18:21:55,401 - Execute['storm kill snort'] {'user': 'metron'}
2017-04-26 18:22:01,016 - Stopping yaf
2017-04-26 18:22:01,017 - Execute['storm kill yaf'] {'user': 'metron'}
Command failed after 1 tries
{code}
--
This message was sent by Atlassian JIRA
(v6.3.15#6346)