----------------------------------------------------------- This is an automatically generated e-mail. To reply, visit: https://reviews.apache.org/r/28649/#review63678 -----------------------------------------------------------
Ship it! Ship It! - Dmitro Lisnichenko On Dec. 3, 2014, 4:10 p.m., Andrew Onischuk wrote: > > ----------------------------------------------------------- > This is an automatically generated e-mail. To reply, visit: > https://reviews.apache.org/r/28649/ > ----------------------------------------------------------- > > (Updated Dec. 3, 2014, 4:10 p.m.) > > > Review request for Ambari and Dmitro Lisnichenko. > > > Bugs: AMBARI-8525 > https://issues.apache.org/jira/browse/AMBARI-8525 > > > Repository: ambari > > > Description > ------- > > Created a 3-node cluster (HDFS, YARN, MR2, TEZ, GANGLIA, NAGIOS, ZOOKEEPER). > Added a 4th host through Add Hosts Wizard. > Selected NodeManager only in Assign Slaves page. > NodeManager Install went fine, but Start failed: > > > > > stderr: > 2014-06-25 02:03:54,423 - Error while executing command 'start': > Traceback (most recent call last): > File > "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", > line 111, in execute > method(env) > File > "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/YARN/package/scripts/nodemanager.py", > line 40, in start > self.configure(env) # FOR SECURITY > File > "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/YARN/package/scripts/nodemanager.py", > line 35, in configure > yarn(name="nodemanager") > File > "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/YARN/package/scripts/yarn.py", > line 61, in yarn > params.HdfsDirectory(None, action="create") > File > "/usr/lib/python2.6/site-packages/resource_management/core/base.py", line > 148, in __init__ > self.env.run() > File > "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", > line 149, in run > self.run_action(resource, action) > File > "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", > line 115, in run_action > provider_action() > File > "/usr/lib/python2.6/site-packages/resource_management/libraries/providers/hdfs_directory.py", > line 104, in action_create > not_if=format("su - {hdp_hdfs_user} -c 'hadoop fs -ls > {dir_list_str}'") > File > "/usr/lib/python2.6/site-packages/resource_management/core/base.py", line > 148, in __init__ > self.env.run() > File > "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", > line 149, in run > self.run_action(resource, action) > File > "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", > line 115, in run_action > provider_action() > File > "/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py", > line 239, in action_run > raise ex > Fail: Execution of 'hadoop fs -mkdir `rpm -q hadoop | grep -q "hadoop-1" > || echo "-p"` /app-logs /mapred /mapred/system /mr-history/tmp > /mr-history/done && hadoop fs -chmod -R 777 /app-logs && hadoop fs -chmod > 777 /mr-history/tmp && hadoop fs -chmod 1777 /mr-history/done && hadoop fs > -chown mapred /mapred && hadoop fs -chown hdfs /mapred/system && hadoop fs > -chown yarn:hadoop /app-logs && hadoop fs -chown mapred:hadoop > /mr-history/tmp /mr-history/done' returned 1. mkdir: No FileSystem for > scheme: hdfs > mkdir: No FileSystem for scheme: hdfs > mkdir: No FileSystem for scheme: hdfs > mkdir: No FileSystem for scheme: hdfs > mkdir: No FileSystem for scheme: hdfs > stdout: > 2014-06-25 02:03:50,084 - Directory['/etc/hadoop/conf.empty'] {'owner': > 'root', 'group': 'root', 'recursive': True} > 2014-06-25 02:03:50,088 - Link['/etc/hadoop/conf'] {'not_if': 'ls > /etc/hadoop/conf', 'to': '/etc/hadoop/conf.empty'} > 2014-06-25 02:03:50,113 - Skipping Link['/etc/hadoop/conf'] due to not_if > 2014-06-25 02:03:50,129 - File['/etc/hadoop/conf/hadoop-env.sh'] > {'content': Template('hadoop-env.sh.j2'), 'owner': 'hdfs'} > 2014-06-25 02:03:50,130 - XmlConfig['core-site.xml'] {'owner': 'hdfs', > 'group': 'hadoop', 'conf_dir': '/etc/hadoop/conf', 'configurations': ...} > 2014-06-25 02:03:50,135 - Generating config: > /etc/hadoop/conf/core-site.xml > 2014-06-25 02:03:50,135 - File['/etc/hadoop/conf/core-site.xml'] > {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': > None} > 2014-06-25 02:03:50,136 - Writing File['/etc/hadoop/conf/core-site.xml'] > because contents don't match > 2014-06-25 02:03:50,151 - Execute['/bin/echo 0 > /selinux/enforce'] > {'only_if': 'test -f /selinux/enforce'} > 2014-06-25 02:03:50,162 - Skipping Execute['/bin/echo 0 > > /selinux/enforce'] due to only_if > 2014-06-25 02:03:50,164 - Execute['mkdir -p > /usr/lib/hadoop/lib/native/Linux-i386-32; ln -sf /usr/lib/libsnappy.so > /usr/lib/hadoop/lib/native/Linux-i386-32/libsnappy.so'] {} > 2014-06-25 02:03:50,178 - Execute['mkdir -p > /usr/lib/hadoop/lib/native/Linux-amd64-64; ln -sf /usr/lib64/libsnappy.so > /usr/lib/hadoop/lib/native/Linux-amd64-64/libsnappy.so'] {} > 2014-06-25 02:03:50,192 - Directory['/var/log/hadoop'] {'owner': 'root', > 'group': 'root', 'recursive': True} > 2014-06-25 02:03:50,193 - Directory['/var/run/hadoop'] {'owner': 'root', > 'group': 'root', 'recursive': True} > 2014-06-25 02:03:50,193 - Directory['/tmp/hadoop-hdfs'] {'owner': 'hdfs', > 'recursive': True} > 2014-06-25 02:03:50,198 - > File['/etc/hadoop/conf/commons-logging.properties'] {'content': > Template('commons-logging.properties.j2'), 'owner': 'hdfs'} > 2014-06-25 02:03:50,199 - File['/etc/hadoop/conf/health_check'] > {'content': Template('health_check-v2.j2'), 'owner': 'hdfs'} > 2014-06-25 02:03:50,200 - File['/etc/hadoop/conf/log4j.properties'] > {'content': '...', 'owner': 'hdfs', 'group': 'hadoop', 'mode': 0644} > 2014-06-25 02:03:50,204 - > File['/etc/hadoop/conf/hadoop-metrics2.properties'] {'content': > Template('hadoop-metrics2.properties.j2'), 'owner': 'hdfs'} > 2014-06-25 02:03:50,205 - File['/etc/hadoop/conf/task-log4j.properties'] > {'content': StaticFile('task-log4j.properties'), 'mode': 0755} > 2014-06-25 02:03:50,205 - File['/etc/hadoop/conf/configuration.xsl'] > {'owner': 'hdfs', 'group': 'hadoop'} > 2014-06-25 02:03:50,300 - HdfsDirectory['/app-logs'] {'security_enabled': > False, 'keytab': [EMPTY], 'conf_dir': '/etc/hadoop/conf', 'hdfs_user': > 'hdfs', 'kinit_path_local': '', 'recursive_chmod': True, 'owner': 'yarn', > 'group': 'hadoop', 'action': ['create_delayed'], 'mode': 0777} > 2014-06-25 02:03:50,301 - HdfsDirectory['/mapred'] {'security_enabled': > False, 'keytab': [EMPTY], 'conf_dir': '/etc/hadoop/conf', 'hdfs_user': > 'hdfs', 'kinit_path_local': '', 'owner': 'mapred', 'action': > ['create_delayed']} > 2014-06-25 02:03:50,301 - HdfsDirectory['/mapred/system'] > {'security_enabled': False, 'keytab': [EMPTY], 'conf_dir': > '/etc/hadoop/conf', 'hdfs_user': 'hdfs', 'kinit_path_local': '', 'owner': > 'hdfs', 'action': ['create_delayed']} > 2014-06-25 02:03:50,302 - HdfsDirectory['/mr-history/tmp'] > {'security_enabled': False, 'keytab': [EMPTY], 'conf_dir': > '/etc/hadoop/conf', 'hdfs_user': 'hdfs', 'kinit_path_local': '', 'mode': > 0777, 'owner': 'mapred', 'group': 'hadoop', 'action': ['create_delayed']} > 2014-06-25 02:03:50,302 - HdfsDirectory['/mr-history/done'] > {'security_enabled': False, 'keytab': [EMPTY], 'conf_dir': > '/etc/hadoop/conf', 'hdfs_user': 'hdfs', 'kinit_path_local': '', 'mode': > 01777, 'owner': 'mapred', 'group': 'hadoop', 'action': ['create_delayed']} > 2014-06-25 02:03:50,302 - HdfsDirectory['None'] {'security_enabled': > False, 'keytab': [EMPTY], 'conf_dir': '/etc/hadoop/conf', 'hdfs_user': > 'hdfs', 'kinit_path_local': '', 'action': ['create']} > 2014-06-25 02:03:50,304 - Execute['hadoop fs -mkdir `rpm -q hadoop | grep > -q "hadoop-1" || echo "-p"` /app-logs /mapred /mapred/system /mr-history/tmp > /mr-history/done && hadoop fs -chmod -R 777 /app-logs && hadoop fs -chmod > 777 /mr-history/tmp && hadoop fs -chmod 1777 /mr-history/done && hadoop fs > -chown mapred /mapred && hadoop fs -chown hdfs /mapred/system && hadoop fs > -chown yarn:hadoop /app-logs && hadoop fs -chown mapred:hadoop > /mr-history/tmp /mr-history/done'] {'not_if': "su - hdfs -c 'hadoop fs -ls > /app-logs /mapred /mapred/system /mr-history/tmp /mr-history/done'", 'user': > 'hdfs'} > 2014-06-25 02:03:54,423 - Error while executing command 'start': > Traceback (most recent call last): > File > "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", > line 111, in execute > method(env) > File > "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/YARN/package/scripts/nodemanager.py", > line 40, in start > self.configure(env) # FOR SECURITY > File > "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/YARN/package/scripts/nodemanager.py", > line 35, in configure > yarn(name="nodemanager") > File > "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/YARN/package/scripts/yarn.py", > line 61, in yarn > params.HdfsDirectory(None, action="create") > File > "/usr/lib/python2.6/site-packages/resource_management/core/base.py", line > 148, in __init__ > self.env.run() > File > "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", > line 149, in run > self.run_action(resource, action) > File > "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", > line 115, in run_action > provider_action() > File > "/usr/lib/python2.6/site-packages/resource_management/libraries/providers/hdfs_directory.py", > line 104, in action_create > not_if=format("su - {hdp_hdfs_user} -c 'hadoop fs -ls > {dir_list_str}'") > File > "/usr/lib/python2.6/site-packages/resource_management/core/base.py", line > 148, in __init__ > self.env.run() > File > "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", > line 149, in run > self.run_action(resource, action) > File > "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", > line 115, in run_action > provider_action() > File > "/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py", > line 239, in action_run > raise ex > Fail: Execution of 'hadoop fs -mkdir `rpm -q hadoop | grep -q "hadoop-1" > || echo "-p"` /app-logs /mapred /mapred/system /mr-history/tmp > /mr-history/done && hadoop fs -chmod -R 777 /app-logs && hadoop fs -chmod > 777 /mr-history/tmp && hadoop fs -chmod 1777 /mr-history/done && hadoop fs > -chown mapred /mapred && hadoop fs -chown hdfs /mapred/system && hadoop fs > -chown yarn:hadoop /app-logs && hadoop fs -chown mapred:hadoop > /mr-history/tmp /mr-history/done' returned 1. mkdir: No FileSystem for > scheme: hdfs > mkdir: No FileSystem for scheme: hdfs > mkdir: No FileSystem for scheme: hdfs > mkdir: No FileSystem for scheme: hdfs > mkdir: No FileSystem for scheme: hdfs > > > > Diffs > ----- > > > ambari-server/src/main/resources/stacks/HDP/2.0.6/services/YARN/metainfo.xml > 2778073 > ambari-server/src/main/resources/stacks/HDP/2.1/services/YARN/metainfo.xml > aaacb7e > ambari-server/src/main/resources/stacks/HDP/2.2/services/YARN/metainfo.xml > 0498268 > > Diff: https://reviews.apache.org/r/28649/diff/ > > > Testing > ------- > > mvn clean test > > > Thanks, > > Andrew Onischuk > >
