anubhav singh created AMBARI-5566:
-------------------------------------

             Summary: Namenode fails to startup
                 Key: AMBARI-5566
                 URL: https://issues.apache.org/jira/browse/AMBARI-5566
             Project: Ambari
          Issue Type: Bug
          Components: Ambari-SCOM
    Affects Versions: 1.5.0
            Reporter: anubhav singh


Ambari server fails to startup even after successful installation.
2014-04-24 12:53:32,513 - Skipping Execute['sh /tmp/checkForFormat.sh hdfs 
/etc/hadoop/conf /var/run/hadoop/hdfs/namenode/formatted/ 
/data/hadoop/hdfs/namenode'] due to not_if
2014-04-24 12:53:32,513 - Execute['mkdir -p 
/var/run/hadoop/hdfs/namenode/formatted/'] {}
2014-04-24 12:53:32,535 - File['/etc/hadoop/conf/dfs.exclude'] {'owner': 
'hdfs', 'content': Template('exclude_hosts_list.j2'), 'group': 'hadoop'}
2014-04-24 12:53:32,539 - Directory['/var/run/hadoop/hdfs'] {'owner': 'hdfs', 
'recursive': True}
2014-04-24 12:53:32,539 - Directory['/var/log/hadoop/hdfs'] {'owner': 'hdfs', 
'recursive': True}
2014-04-24 12:53:32,540 - File['/var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid'] 
{'action': ['delete'], 'not_if': 'ls 
/var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps `cat 
/var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1', 
'ignore_failures': True}
2014-04-24 12:53:32,567 - Deleting 
File['/var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid']
2014-04-24 12:53:32,567 - Execute['ulimit -c unlimited;  export 
HADOOP_LIBEXEC_DIR=/usr/lib/hadoop/libexec && 
/usr/lib/hadoop/sbin/hadoop-daemon.sh --config /etc/hadoop/conf start 
namenode'] {'not_if': 'ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid 
>/dev/null 2>&1 && ps `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` 
>/dev/null 2>&1', 'user': 'hdfs'}
2014-04-24 12:53:36,689 - Error while executing command 'start':
Traceback (most recent call last):
  File 
"/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py",
 line 95, in execute
    method(env)
  File 
"/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/HDFS/package/scripts/namenode.py",
 line 38, in start
    namenode(action="start")
  File 
"/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/HDFS/package/scripts/hdfs_namenode.py",
 line 45, in namenode
    create_log_dir=True
  File 
"/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/HDFS/package/scripts/utils.py",
 line 63, in service
    not_if=service_is_up
  File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", 
line 148, in __init__
    self.env.run()
  File 
"/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
line 149, in run
    self.run_action(resource, action)
  File 
"/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
line 115, in run_action
    provider_action()
  File 
"/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py",
 line 239, in action_run
    raise ex
Fail: Execution of 'ulimit -c unlimited;  export 
HADOOP_LIBEXEC_DIR=/usr/lib/hadoop/libexec && 
/usr/lib/hadoop/sbin/hadoop-daemon.sh --config /etc/hadoop/conf start namenode' 
returned 1. starting namenode, logging to 
/var/log/hadoop/hdfs/hadoop-hdfs-namenode-vmhost4-vm0.frem.wandisco.com.out



--
This message was sent by Atlassian JIRA
(v6.2#6252)

Reply via email to