Repository: ambari
Updated Branches:
  refs/heads/trunk d19f0bb76 -> b84cbbe7e


AMBARI-22083 - Wrong Hadoop Home Directory Is Being Picked Up on MAINT/PATCH 
Upgraded Clusters (part2) (jonathanhurley)


Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/b84cbbe7
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/b84cbbe7
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/b84cbbe7

Branch: refs/heads/trunk
Commit: b84cbbe7e9a3afffab07da534541b4eb6076d50c
Parents: d19f0bb
Author: Jonathan Hurley <jhur...@hortonworks.com>
Authored: Fri Sep 29 16:01:56 2017 -0400
Committer: Jonathan Hurley <jhur...@hortonworks.com>
Committed: Fri Sep 29 16:01:56 2017 -0400

----------------------------------------------------------------------
 .../stacks/HDP/3.0/hooks/after-INSTALL/scripts/params.py       | 2 +-
 .../stacks/HDP/3.0/hooks/before-ANY/scripts/params.py          | 6 +++---
 .../stacks/HDP/3.0/hooks/before-START/scripts/params.py        | 2 +-
 3 files changed, 5 insertions(+), 5 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/ambari/blob/b84cbbe7/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/after-INSTALL/scripts/params.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/after-INSTALL/scripts/params.py
 
b/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/after-INSTALL/scripts/params.py
index ddc6100..34dfe70 100644
--- 
a/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/after-INSTALL/scripts/params.py
+++ 
b/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/after-INSTALL/scripts/params.py
@@ -101,7 +101,7 @@ namenode_host = default("/clusterHostInfo/namenode_host", 
[])
 has_namenode = not len(namenode_host) == 0
 
 if has_namenode or dfs_type == 'HCFS':
-  hadoop_conf_dir = 
conf_select.get_hadoop_conf_dir(force_latest_on_upgrade=True)
+  hadoop_conf_dir = conf_select.get_hadoop_conf_dir()
 
 link_configs_lock_file = get_config_lock_file()
 stack_select_lock_file = os.path.join(tmp_dir, "stack_select_lock_file")

http://git-wip-us.apache.org/repos/asf/ambari/blob/b84cbbe7/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/before-ANY/scripts/params.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/before-ANY/scripts/params.py
 
b/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/before-ANY/scripts/params.py
index b111c02..cee0519 100644
--- 
a/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/before-ANY/scripts/params.py
+++ 
b/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/before-ANY/scripts/params.py
@@ -107,8 +107,8 @@ mapreduce_libs_path = 
format("{stack_root}/current/hadoop-mapreduce-client/*")
 # which would cause a lot of problems when writing out hadoop-env.sh; instead
 # force the use of "current" in the hook
 hdfs_user_nofile_limit = 
default("/configurations/hadoop-env/hdfs_user_nofile_limit", "128000")
-hadoop_home = stack_select.get_hadoop_dir("home", force_latest_on_upgrade=True)
-hadoop_libexec_dir = stack_select.get_hadoop_dir("libexec", 
force_latest_on_upgrade=True)
+hadoop_home = stack_select.get_hadoop_dir("home")
+hadoop_libexec_dir = stack_select.get_hadoop_dir("libexec")
 
 hadoop_conf_empty_dir = None
 hadoop_secure_dn_user = hdfs_user
@@ -210,7 +210,7 @@ if dfs_ha_namenode_ids:
 
 
 if has_namenode or dfs_type == 'HCFS':
-    hadoop_conf_dir = 
conf_select.get_hadoop_conf_dir(force_latest_on_upgrade=True)
+    hadoop_conf_dir = conf_select.get_hadoop_conf_dir()
     hadoop_conf_secure_dir = os.path.join(hadoop_conf_dir, "secure")
 
 hbase_tmp_dir = "/tmp/hbase-hbase"

http://git-wip-us.apache.org/repos/asf/ambari/blob/b84cbbe7/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/before-START/scripts/params.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/before-START/scripts/params.py
 
b/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/before-START/scripts/params.py
index d70030d..8555fea 100644
--- 
a/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/before-START/scripts/params.py
+++ 
b/ambari-server/src/main/resources/stacks/HDP/3.0/hooks/before-START/scripts/params.py
@@ -176,7 +176,7 @@ if has_zk_host:
 
 if has_namenode or dfs_type == 'HCFS':
   hadoop_tmp_dir = format("/tmp/hadoop-{hdfs_user}")
-  hadoop_conf_dir = 
conf_select.get_hadoop_conf_dir(force_latest_on_upgrade=True)
+  hadoop_conf_dir = conf_select.get_hadoop_conf_dir()
   task_log4j_properties_location = os.path.join(hadoop_conf_dir, 
"task-log4j.properties")
 
 hadoop_pid_dir_prefix = 
config['configurations']['hadoop-env']['hadoop_pid_dir_prefix']

Reply via email to