Repository: ambari Updated Branches: refs/heads/trunk d19f0bb76 -> b84cbbe7e
AMBARI-22083 - Wrong Hadoop Home Directory Is Being Picked Up on MAINT/PATCH Upgraded Clusters (part2) (jonathanhurley) Project: http://git-wip-us.apache.org/repos/asf/ambari/repo Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/b84cbbe7 Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/b84cbbe7 Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/b84cbbe7 Branch: refs/heads/trunk Commit: b84cbbe7e9a3afffab07da534541b4eb6076d50c Parents: d19f0bb Author: Jonathan Hurley <jhur...@hortonworks.com> Authored: Fri Sep 29 16:01:56 2017 -0400 Committer: Jonathan Hurley <jhur...@hortonworks.com> Committed: Fri Sep 29 16:01:56 2017 -0400 ---------------------------------------------------------------------- .../stacks/HDP/3.0/hooks/after-INSTALL/scripts/params.py | 2 +- .../stacks/HDP/3.0/hooks/before-ANY/scripts/params.py | 6 +++--- .../stacks/HDP/3.0/hooks/before-START/scripts/params.py | 2 +- 3 files changed, 5 insertions(+), 5 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/ambari/blob/b84cbbe7/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/after-INSTALL/scripts/params.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/after-INSTALL/scripts/params.py b/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/after-INSTALL/scripts/params.py index ddc6100..34dfe70 100644 --- a/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/after-INSTALL/scripts/params.py +++ b/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/after-INSTALL/scripts/params.py @@ -101,7 +101,7 @@ namenode_host = default("/clusterHostInfo/namenode_host", []) has_namenode = not len(namenode_host) == 0 if has_namenode or dfs_type == 'HCFS': - hadoop_conf_dir = conf_select.get_hadoop_conf_dir(force_latest_on_upgrade=True) + hadoop_conf_dir = conf_select.get_hadoop_conf_dir() link_configs_lock_file = get_config_lock_file() stack_select_lock_file = os.path.join(tmp_dir, "stack_select_lock_file") http://git-wip-us.apache.org/repos/asf/ambari/blob/b84cbbe7/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/before-ANY/scripts/params.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/before-ANY/scripts/params.py b/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/before-ANY/scripts/params.py index b111c02..cee0519 100644 --- a/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/before-ANY/scripts/params.py +++ b/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/before-ANY/scripts/params.py @@ -107,8 +107,8 @@ mapreduce_libs_path = format("{stack_root}/current/hadoop-mapreduce-client/*") # which would cause a lot of problems when writing out hadoop-env.sh; instead # force the use of "current" in the hook hdfs_user_nofile_limit = default("/configurations/hadoop-env/hdfs_user_nofile_limit", "128000") -hadoop_home = stack_select.get_hadoop_dir("home", force_latest_on_upgrade=True) -hadoop_libexec_dir = stack_select.get_hadoop_dir("libexec", force_latest_on_upgrade=True) +hadoop_home = stack_select.get_hadoop_dir("home") +hadoop_libexec_dir = stack_select.get_hadoop_dir("libexec") hadoop_conf_empty_dir = None hadoop_secure_dn_user = hdfs_user @@ -210,7 +210,7 @@ if dfs_ha_namenode_ids: if has_namenode or dfs_type == 'HCFS': - hadoop_conf_dir = conf_select.get_hadoop_conf_dir(force_latest_on_upgrade=True) + hadoop_conf_dir = conf_select.get_hadoop_conf_dir() hadoop_conf_secure_dir = os.path.join(hadoop_conf_dir, "secure") hbase_tmp_dir = "/tmp/hbase-hbase" http://git-wip-us.apache.org/repos/asf/ambari/blob/b84cbbe7/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/before-START/scripts/params.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/before-START/scripts/params.py b/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/before-START/scripts/params.py index d70030d..8555fea 100644 --- a/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/before-START/scripts/params.py +++ b/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/before-START/scripts/params.py @@ -176,7 +176,7 @@ if has_zk_host: if has_namenode or dfs_type == 'HCFS': hadoop_tmp_dir = format("/tmp/hadoop-{hdfs_user}") - hadoop_conf_dir = conf_select.get_hadoop_conf_dir(force_latest_on_upgrade=True) + hadoop_conf_dir = conf_select.get_hadoop_conf_dir() task_log4j_properties_location = os.path.join(hadoop_conf_dir, "task-log4j.properties") hadoop_pid_dir_prefix = config['configurations']['hadoop-env']['hadoop_pid_dir_prefix']