[ 
https://issues.apache.org/jira/browse/AMBARI-14596?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alejandro Fernandez updated AMBARI-14596:
-----------------------------------------
    Description: 
Cluster installation failed on Accumulo Client because it was one of the first 
tasks scheduled and HDFS Client had not been installed yet, which installs the 
hadoop rpm and creates the /etc/hadoop/conf folder.

If a host does not contain /etc/hadoop/conf, then we should not attempt to 
write config files to it during the after-install hooks. Once a component is 
installed that does contain the hadoop rpm, then it will be responsible for 
writing out the configs to it.

Ambari 2.2.1.0-71
HDP 2.4.0.0-47

{code}Traceback (most recent call last):
  File 
"/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/hook.py",
 line 38, in <module>
    AfterInstallHook().execute()
  File 
"/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py",
 line 219, in execute
    method(env)
  File 
"/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/hook.py",
 line 33, in hook
    setup_config()
  File 
"/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/shared_initialization.py",
 line 55, in setup_config
    only_if=format("ls {hadoop_conf_dir}"))
  File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", 
line 154, in __init__
    self.env.run()
  File 
"/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
line 158, in run
    self.run_action(resource, action)
  File 
"/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
line 121, in run_action
    provider_action()
  File 
"/usr/lib/python2.6/site-packages/resource_management/libraries/providers/xml_config.py",
 line 67, in action_create
    encoding = self.resource.encoding
  File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", 
line 154, in __init__
    self.env.run()
  File 
"/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
line 158, in run
    self.run_action(resource, action)
  File 
"/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
line 121, in run_action
    provider_action()
  File 
"/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py",
 line 87, in action_create
    raise Fail("Applying %s failed, parent directory %s doesn't exist" % 
(self.resource, dirname))
resource_management.core.exceptions.Fail: Applying 
File['/usr/hdp/current/hadoop-client/conf/core-site.xml'] failed, parent 
directory /usr/hdp/current/hadoop-client/conf doesn't exist{code}

  was:
Cluster installation failed on Accumulo Client because it was one of the first 
tasks scheduled and HDFS Client had not been installed yet, which installs the 
hadoop rpm and creates the /etc/hadoop/conf folder.

If a host does not contain /etc/hadoop/conf, then we should not attempt to 
write config files to it during the after-install hooks.

Ambari 2.2.1.0-71
HDP 2.4.0.0-47

{code}Traceback (most recent call last):
  File 
"/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/hook.py",
 line 38, in <module>
    AfterInstallHook().execute()
  File 
"/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py",
 line 219, in execute
    method(env)
  File 
"/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/hook.py",
 line 33, in hook
    setup_config()
  File 
"/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/shared_initialization.py",
 line 55, in setup_config
    only_if=format("ls {hadoop_conf_dir}"))
  File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", 
line 154, in __init__
    self.env.run()
  File 
"/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
line 158, in run
    self.run_action(resource, action)
  File 
"/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
line 121, in run_action
    provider_action()
  File 
"/usr/lib/python2.6/site-packages/resource_management/libraries/providers/xml_config.py",
 line 67, in action_create
    encoding = self.resource.encoding
  File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", 
line 154, in __init__
    self.env.run()
  File 
"/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
line 158, in run
    self.run_action(resource, action)
  File 
"/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
line 121, in run_action
    provider_action()
  File 
"/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py",
 line 87, in action_create
    raise Fail("Applying %s failed, parent directory %s doesn't exist" % 
(self.resource, dirname))
resource_management.core.exceptions.Fail: Applying 
File['/usr/hdp/current/hadoop-client/conf/core-site.xml'] failed, parent 
directory /usr/hdp/current/hadoop-client/conf doesn't exist{code}


> Install cluster failed on Accumulo as tried to write config when hadoop conf 
> dir is missing
> -------------------------------------------------------------------------------------------
>
>                 Key: AMBARI-14596
>                 URL: https://issues.apache.org/jira/browse/AMBARI-14596
>             Project: Ambari
>          Issue Type: Bug
>          Components: ambari-server
>    Affects Versions: 2.2.0
>            Reporter: Alejandro Fernandez
>            Assignee: Alejandro Fernandez
>             Fix For: 2.2.1
>
>         Attachments: AMBARI-14596.trunk.patch
>
>
> Cluster installation failed on Accumulo Client because it was one of the 
> first tasks scheduled and HDFS Client had not been installed yet, which 
> installs the hadoop rpm and creates the /etc/hadoop/conf folder.
> If a host does not contain /etc/hadoop/conf, then we should not attempt to 
> write config files to it during the after-install hooks. Once a component is 
> installed that does contain the hadoop rpm, then it will be responsible for 
> writing out the configs to it.
> Ambari 2.2.1.0-71
> HDP 2.4.0.0-47
> {code}Traceback (most recent call last):
>   File 
> "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/hook.py",
>  line 38, in <module>
>     AfterInstallHook().execute()
>   File 
> "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py",
>  line 219, in execute
>     method(env)
>   File 
> "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/hook.py",
>  line 33, in hook
>     setup_config()
>   File 
> "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/shared_initialization.py",
>  line 55, in setup_config
>     only_if=format("ls {hadoop_conf_dir}"))
>   File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", 
> line 154, in __init__
>     self.env.run()
>   File 
> "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
> line 158, in run
>     self.run_action(resource, action)
>   File 
> "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
> line 121, in run_action
>     provider_action()
>   File 
> "/usr/lib/python2.6/site-packages/resource_management/libraries/providers/xml_config.py",
>  line 67, in action_create
>     encoding = self.resource.encoding
>   File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", 
> line 154, in __init__
>     self.env.run()
>   File 
> "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
> line 158, in run
>     self.run_action(resource, action)
>   File 
> "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
> line 121, in run_action
>     provider_action()
>   File 
> "/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py",
>  line 87, in action_create
>     raise Fail("Applying %s failed, parent directory %s doesn't exist" % 
> (self.resource, dirname))
> resource_management.core.exceptions.Fail: Applying 
> File['/usr/hdp/current/hadoop-client/conf/core-site.xml'] failed, parent 
> directory /usr/hdp/current/hadoop-client/conf doesn't exist{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to