[
https://issues.apache.org/jira/browse/AMBARI-17062?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15320753#comment-15320753
]
Hudson commented on AMBARI-17062:
---------------------------------
FAILURE: Integrated in Ambari-trunk-Commit #5033 (See
[https://builds.apache.org/job/Ambari-trunk-Commit/5033/])
AMBARI-17062. Namenode failed to start while installing a cluster from
(aonishuk:
[http://git-wip-us.apache.org/repos/asf?p=ambari.git&a=commit&h=05c12d51f8d1ca2716cad60af5701f38321931e3])
*
ambari-server/src/main/resources/stacks/HDP/2.1/services/YARN/configuration/yarn-env.xml
*
ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/configuration/hadoop-env.xml
*
ambari-server/src/main/resources/stacks/HDP/2.3/services/YARN/configuration/yarn-log4j.xml
* ambari-server/src/main/python/ambari_server/serverSetup.py
*
ambari-server/src/main/resources/stacks/HDP/2.3/services/YARN/configuration/yarn-env.xml
*
ambari-server/src/main/resources/common-services/YARN/2.1.0.2.0/configuration/yarn-log4j.xml
*
ambari-server/src/main/resources/common-services/YARN/2.1.0.2.0/configuration/yarn-env.xml
> Namenode failed to start while installing a cluster from UI
> -----------------------------------------------------------
>
> Key: AMBARI-17062
> URL: https://issues.apache.org/jira/browse/AMBARI-17062
> Project: Ambari
> Issue Type: Bug
> Reporter: Andrew Onischuk
> Assignee: Andrew Onischuk
> Fix For: 2.4.0
>
> Attachments: AMBARI-17062.patch
>
>
> STR: Install a cluster from UI with HDFS, ZK and YARN
> Namenode fails to start
> The cluster is live here: <http://172.22.124.223:8080/>
>
>
>
> Traceback (most recent call last):
> File
> "/var/lib/ambari-agent/cache/common-services/HDFS/2.1.0.2.0/package/scripts/namenode.py",
> line 414, in <module>
> NameNode().execute()
> File
> "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py",
> line 257, in execute
> method(env)
> File
> "/var/lib/ambari-agent/cache/common-services/HDFS/2.1.0.2.0/package/scripts/namenode.py",
> line 101, in start
> upgrade_suspended=params.upgrade_suspended, env=env)
> File
> "/usr/lib/python2.6/site-packages/ambari_commons/os_family_impl.py", line 89,
> in thunk
> return fn(*args, **kwargs)
> File
> "/var/lib/ambari-agent/cache/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_namenode.py",
> line 155, in namenode
> create_log_dir=True
> File
> "/var/lib/ambari-agent/cache/common-services/HDFS/2.1.0.2.0/package/scripts/utils.py",
> line 269, in service
> Execute(daemon_cmd, not_if=process_id_exists_command,
> environment=hadoop_env_exports)
> File
> "/usr/lib/python2.6/site-packages/resource_management/core/base.py", line
> 155, in __init__
> self.env.run()
> File
> "/usr/lib/python2.6/site-packages/resource_management/core/environment.py",
> line 160, in run
> self.run_action(resource, action)
> File
> "/usr/lib/python2.6/site-packages/resource_management/core/environment.py",
> line 124, in run_action
> provider_action()
> File
> "/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py",
> line 273, in action_run
> tries=self.resource.tries, try_sleep=self.resource.try_sleep)
> File
> "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line
> 70, in inner
> result = function(command, **kwargs)
> File
> "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line
> 92, in checked_call
> tries=tries, try_sleep=try_sleep)
> File
> "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line
> 140, in _call_wrapper
> result = _call(command, **kwargs_copy)
> File
> "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line
> 293, in _call
> raise Fail(err_msg)
> resource_management.core.exceptions.Fail: Execution of 'ambari-sudo.sh su
> hdfs -l -s /bin/bash -c 'ulimit -c unlimited ;
> /usr/hdp/current/hadoop-client/sbin/hadoop-daemon.sh --config
> /usr/hdp/current/hadoop-client/conf start namenode'' returned 1. ########
> Hortonworks #############
> This is MOTD message, added for testing in qe infra
> starting namenode, logging to
> /grid/0/log/hdfs/hdfs/hadoop-hdfs-namenode-os-r7-vsrfou-ambari-serv-11r-1.out
> log4j:ERROR setFile(null,true) call failed.
> java.io.FileNotFoundException: ./nm-audit.log (Permission denied)
> at java.io.FileOutputStream.open(Native Method)
> at java.io.FileOutputStream.<init>(FileOutputStream.java:221)
> at java.io.FileOutputStream.<init>(FileOutputStream.java:142)
> at org.apache.log4j.FileAppender.setFile(FileAppender.java:294)
> at org.apache.log4j.FileAppender.activateOptions(FileAppender.java:165)
> at
> org.apache.log4j.DailyRollingFileAppender.activateOptions(DailyRollingFileAppender.java:223)
> at
> org.apache.log4j.config.PropertySetter.activate(PropertySetter.java:307)
> at
> org.apache.log4j.config.PropertySetter.setProperties(PropertySetter.java:172)
>
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)