anubhav singh created AMBARI-5566:
-------------------------------------
Summary: Namenode fails to startup
Key: AMBARI-5566
URL: https://issues.apache.org/jira/browse/AMBARI-5566
Project: Ambari
Issue Type: Bug
Components: Ambari-SCOM
Affects Versions: 1.5.0
Reporter: anubhav singh
Ambari server fails to startup even after successful installation.
2014-04-24 12:53:32,513 - Skipping Execute['sh /tmp/checkForFormat.sh hdfs
/etc/hadoop/conf /var/run/hadoop/hdfs/namenode/formatted/
/data/hadoop/hdfs/namenode'] due to not_if
2014-04-24 12:53:32,513 - Execute['mkdir -p
/var/run/hadoop/hdfs/namenode/formatted/'] {}
2014-04-24 12:53:32,535 - File['/etc/hadoop/conf/dfs.exclude'] {'owner':
'hdfs', 'content': Template('exclude_hosts_list.j2'), 'group': 'hadoop'}
2014-04-24 12:53:32,539 - Directory['/var/run/hadoop/hdfs'] {'owner': 'hdfs',
'recursive': True}
2014-04-24 12:53:32,539 - Directory['/var/log/hadoop/hdfs'] {'owner': 'hdfs',
'recursive': True}
2014-04-24 12:53:32,540 - File['/var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid']
{'action': ['delete'], 'not_if': 'ls
/var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps `cat
/var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
'ignore_failures': True}
2014-04-24 12:53:32,567 - Deleting
File['/var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid']
2014-04-24 12:53:32,567 - Execute['ulimit -c unlimited; export
HADOOP_LIBEXEC_DIR=/usr/lib/hadoop/libexec &&
/usr/lib/hadoop/sbin/hadoop-daemon.sh --config /etc/hadoop/conf start
namenode'] {'not_if': 'ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid
>/dev/null 2>&1 && ps `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid`
>/dev/null 2>&1', 'user': 'hdfs'}
2014-04-24 12:53:36,689 - Error while executing command 'start':
Traceback (most recent call last):
File
"/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py",
line 95, in execute
method(env)
File
"/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/HDFS/package/scripts/namenode.py",
line 38, in start
namenode(action="start")
File
"/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/HDFS/package/scripts/hdfs_namenode.py",
line 45, in namenode
create_log_dir=True
File
"/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/HDFS/package/scripts/utils.py",
line 63, in service
not_if=service_is_up
File "/usr/lib/python2.6/site-packages/resource_management/core/base.py",
line 148, in __init__
self.env.run()
File
"/usr/lib/python2.6/site-packages/resource_management/core/environment.py",
line 149, in run
self.run_action(resource, action)
File
"/usr/lib/python2.6/site-packages/resource_management/core/environment.py",
line 115, in run_action
provider_action()
File
"/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py",
line 239, in action_run
raise ex
Fail: Execution of 'ulimit -c unlimited; export
HADOOP_LIBEXEC_DIR=/usr/lib/hadoop/libexec &&
/usr/lib/hadoop/sbin/hadoop-daemon.sh --config /etc/hadoop/conf start namenode'
returned 1. starting namenode, logging to
/var/log/hadoop/hdfs/hadoop-hdfs-namenode-vmhost4-vm0.frem.wandisco.com.out
--
This message was sent by Atlassian JIRA
(v6.2#6252)