AMBARI-21580 - Replace Hard Coded stack-select Structures (jonathanhurley)

Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/56d2ade2
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/56d2ade2
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/56d2ade2

Branch: refs/heads/branch-2.6
Commit: 56d2ade2b78adf170a92a99dbbd8d5d1a074389a
Parents: 863343a
Author: Jonathan Hurley <jhur...@hortonworks.com>
Authored: Wed Jul 26 15:01:33 2017 -0400
Committer: Jonathan Hurley <jhur...@hortonworks.com>
Committed: Thu Jul 27 15:29:07 2017 -0400

----------------------------------------------------------------------
 .../libraries/functions/stack_select.py         | 171 +++-
 .../libraries/script/script.py                  |  23 +-
 .../BlueprintConfigurationProcessor.java        |   5 +-
 .../ambari/server/state/ConfigHelper.java       |   1 +
 .../server/upgrade/FinalUpgradeCatalog.java     |   4 +-
 .../package/scripts/accumulo_client.py          |   5 +-
 .../package/scripts/accumulo_script.py          |  34 +-
 .../0.1.0.2.3/package/scripts/atlas_client.py   |   5 +-
 .../package/scripts/metadata_server.py          |   6 +-
 .../DRUID/0.9.2/package/scripts/druid_node.py   |   6 +-
 .../DRUID/0.9.2/package/scripts/superset.py     |   5 +-
 .../0.5.0.2.1/package/scripts/falcon_client.py  |   5 +-
 .../0.5.0.2.1/package/scripts/falcon_server.py  |   5 +-
 .../1.4.0.2.0/package/scripts/flume_handler.py  |   5 +-
 .../0.96.0.2.0/package/scripts/hbase_client.py  |   7 +-
 .../0.96.0.2.0/package/scripts/hbase_master.py  |   5 +-
 .../package/scripts/hbase_regionserver.py       |   5 +-
 .../package/scripts/phoenix_queryserver.py      |   6 +-
 .../HBASE/0.96.0.2.0/package/scripts/upgrade.py |   4 +-
 .../HDFS/2.1.0.2.0/package/scripts/datanode.py  |   7 +-
 .../2.1.0.2.0/package/scripts/hdfs_client.py    |   5 +-
 .../2.1.0.2.0/package/scripts/journalnode.py    |   5 +-
 .../HDFS/2.1.0.2.0/package/scripts/namenode.py  |   7 +-
 .../2.1.0.2.0/package/scripts/nfsgateway.py     |   6 +-
 .../HDFS/2.1.0.2.0/package/scripts/snamenode.py |   5 +-
 .../2.1.0.2.0/package/scripts/zkfc_slave.py     |   9 +-
 .../0.12.0.2.0/package/scripts/hcat_client.py   |   8 +-
 .../0.12.0.2.0/package/scripts/hive_client.py   |   5 +-
 .../package/scripts/hive_metastore.py           |   6 +-
 .../0.12.0.2.0/package/scripts/hive_server.py   |   5 +-
 .../package/scripts/hive_server_interactive.py  |   6 +-
 .../package/scripts/webhcat_server.py           |   5 +-
 .../KAFKA/0.8.1/package/scripts/kafka_broker.py |   5 +-
 .../0.5.0.2.2/package/scripts/knox_gateway.py   |   5 +-
 .../1.0.0.2.3/package/scripts/mahout_client.py  |   7 +-
 .../4.0.0.2.0/package/scripts/oozie_client.py   |   5 +-
 .../4.0.0.2.0/package/scripts/oozie_server.py   |   8 +-
 .../0.12.0.2.0/package/scripts/pig_client.py    |   5 +-
 .../R4ML/0.8.0/package/scripts/r4ml_client.py   |   5 +-
 .../0.4.0/package/scripts/ranger_admin.py       |   5 +-
 .../0.4.0/package/scripts/ranger_tagsync.py     |   7 +-
 .../0.4.0/package/scripts/ranger_usersync.py    |   3 -
 .../RANGER/0.4.0/package/scripts/upgrade.py     |   2 +-
 .../0.5.0.2.3/package/scripts/kms_server.py     |   3 -
 .../0.5.0.2.3/package/scripts/upgrade.py        |   2 +-
 .../0.60.0.2.2/package/scripts/slider_client.py |   6 +-
 .../1.2.1/package/scripts/job_history_server.py |   5 +-
 .../SPARK/1.2.1/package/scripts/livy_server.py  |   5 +-
 .../SPARK/1.2.1/package/scripts/spark_client.py |   5 +-
 .../package/scripts/spark_thrift_server.py      |   5 +-
 .../2.0.0/package/scripts/job_history_server.py |   5 +-
 .../2.0.0/package/scripts/livy2_server.py       |   5 +-
 .../2.0.0/package/scripts/spark_client.py       |   5 +-
 .../package/scripts/spark_thrift_server.py      |   5 +-
 .../1.4.4.2.0/package/scripts/service_check.py  |   3 -
 .../1.4.4.2.0/package/scripts/sqoop_client.py   |   5 +-
 .../STORM/0.9.1/package/scripts/drpc_server.py  |   5 +-
 .../STORM/0.9.1/package/scripts/nimbus.py       |   7 +-
 .../STORM/0.9.1/package/scripts/nimbus_prod.py  |   6 +-
 .../STORM/0.9.1/package/scripts/pacemaker.py    |   5 +-
 .../STORM/0.9.1/package/scripts/rest_api.py     |   3 -
 .../STORM/0.9.1/package/scripts/supervisor.py   |   6 +-
 .../0.9.1/package/scripts/supervisor_prod.py    |   6 +-
 .../STORM/0.9.1/package/scripts/ui_server.py    |   5 +-
 .../0.10.0/package/scripts/systemml_client.py   |   5 +-
 .../TEZ/0.4.0.2.1/package/scripts/tez_client.py |   5 +-
 .../TITAN/1.0.0/package/scripts/titan_client.py |   5 +-
 .../TITAN/1.0.0/package/scripts/titan_server.py |   5 +-
 .../scripts/application_timeline_server.py      |   5 +-
 .../2.1.0.2.0/package/scripts/historyserver.py  |   5 +-
 .../package/scripts/mapreduce2_client.py        |   5 +-
 .../2.1.0.2.0/package/scripts/nodemanager.py    |   5 +-
 .../package/scripts/resourcemanager.py          |   5 +-
 .../2.1.0.2.0/package/scripts/yarn_client.py    |   5 +-
 .../0.6.0.2.5/package/scripts/master.py         |   5 +-
 .../3.4.5/package/scripts/zookeeper_client.py   |   4 +-
 .../3.4.5/package/scripts/zookeeper_server.py   |   5 +-
 .../4.0/properties/stack_select_packages.json   | 775 +++++++++++++++
 .../FLUME/package/scripts/flume_handler.py      |   5 +-
 .../HBASE/package/scripts/hbase_client.py       |   6 +-
 .../HBASE/package/scripts/hbase_master.py       |   6 +-
 .../HBASE/package/scripts/hbase_regionserver.py |   5 +-
 .../package/scripts/hbase_restgatewayserver.py  |   5 +-
 .../services/HBASE/package/scripts/upgrade.py   |   4 +-
 .../services/HDFS/package/scripts/datanode.py   |   5 +-
 .../HDFS/package/scripts/hdfs_client.py         |   5 +-
 .../HDFS/package/scripts/journalnode.py         |   5 +-
 .../services/HDFS/package/scripts/namenode.py   |   5 +-
 .../services/HDFS/package/scripts/nfsgateway.py |   5 +-
 .../services/HDFS/package/scripts/snamenode.py  |   5 +-
 .../HIVE/package/scripts/hcat_client.py         |   4 +-
 .../HIVE/package/scripts/hive_client.py         |   7 +-
 .../HIVE/package/scripts/hive_metastore.py      |   5 +-
 .../HIVE/package/scripts/hive_server.py         |   5 +-
 .../HIVE/package/scripts/webhcat_server.py      |   5 +-
 .../KAFKA/package/scripts/kafka_broker.py       |   5 +-
 .../services/KAFKA/package/scripts/upgrade.py   |   4 +-
 .../KNOX/package/scripts/knox_gateway.py        |   6 +-
 .../OOZIE/package/scripts/oozie_client.py       |   5 +-
 .../OOZIE/package/scripts/oozie_server.py       |   8 +-
 .../services/PIG/package/scripts/pig_client.py  |   5 +-
 .../SLIDER/package/scripts/slider_client.py     |   6 +-
 .../SOLR/package/scripts/solr_server.py         |   2 +-
 .../SPARK/package/scripts/job_history_server.py |   5 +-
 .../SPARK/package/scripts/spark_client.py       |   5 +-
 .../package/scripts/spark_thrift_server.py      |   6 +-
 .../SQOOP/package/scripts/service_check.py      |   3 -
 .../SQOOP/package/scripts/sqoop_client.py       |   5 +-
 .../scripts/application_timeline_server.py      |   5 +-
 .../YARN/package/scripts/historyserver.py       |   5 +-
 .../YARN/package/scripts/mapreduce2_client.py   |   5 +-
 .../YARN/package/scripts/nodemanager.py         |   5 +-
 .../YARN/package/scripts/resourcemanager.py     |   5 +-
 .../YARN/package/scripts/yarn_client.py         |   5 +-
 .../ZOOKEEPER/package/scripts/zookeeper.py      |   2 +-
 .../package/scripts/zookeeper_client.py         |   5 +-
 .../package/scripts/zookeeper_server.py         |   5 +-
 .../scripts/shared_initialization.py            |  19 +-
 .../HBASE/package/scripts/hbase_client.py       |   8 +-
 .../HBASE/package/scripts/hbase_master.py       |   5 +-
 .../HBASE/package/scripts/hbase_regionserver.py |   5 +-
 .../package/scripts/hbase_restgatewayserver.py  |   5 +-
 .../package/scripts/phoenix_queryserver.py      |   6 +-
 .../services/HBASE/package/scripts/upgrade.py   |   4 +-
 .../FLUME/package/scripts/flume_handler.py      |   5 +-
 .../HBASE/package/scripts/hbase_client.py       |   7 +-
 .../HBASE/package/scripts/hbase_master.py       |   5 +-
 .../HBASE/package/scripts/hbase_regionserver.py |   5 +-
 .../package/scripts/hbase_restgatewayserver.py  |   5 +-
 .../package/scripts/phoenix_queryserver.py      |   7 +-
 .../services/HBASE/package/scripts/upgrade.py   |   4 +-
 .../services/HDFS/package/scripts/datanode.py   |   7 +-
 .../HDFS/package/scripts/hdfs_client.py         |   5 +-
 .../HDFS/package/scripts/journalnode.py         |   5 +-
 .../services/HDFS/package/scripts/namenode.py   |   7 +-
 .../services/HDFS/package/scripts/nfsgateway.py |   5 +-
 .../services/HDFS/package/scripts/snamenode.py  |   5 +-
 .../HIVE/package/scripts/hcat_client.py         |   4 +-
 .../HIVE/package/scripts/hive_client.py         |   7 +-
 .../HIVE/package/scripts/hive_metastore.py      |   5 +-
 .../HIVE/package/scripts/hive_server.py         |   5 +-
 .../HIVE/package/scripts/webhcat_server.py      |   5 +-
 .../KAFKA/package/scripts/kafka_broker.py       |   5 +-
 .../KNOX/package/scripts/knox_gateway.py        |   7 +-
 .../OOZIE/package/scripts/oozie_client.py       |   5 +-
 .../OOZIE/package/scripts/oozie_server.py       |  10 +-
 .../services/PIG/package/scripts/pig_client.py  |   5 +-
 .../RANGER/package/scripts/ranger_admin.py      |   5 +-
 .../RANGER/package/scripts/ranger_usersync.py   |   4 -
 .../services/RANGER/package/scripts/upgrade.py  |   2 +-
 .../RANGER_KMS/package/scripts/kms_server.py    |   3 -
 .../RANGER_KMS/package/scripts/upgrade.py       |   2 +-
 .../SLIDER/package/scripts/slider_client.py     |   6 +-
 .../SOLR/package/scripts/solr_server.py         |   2 +-
 .../SPARK/package/scripts/job_history_server.py |   5 +-
 .../SPARK/package/scripts/spark_client.py       |   5 +-
 .../package/scripts/spark_thrift_server.py      |   6 +-
 .../SQOOP/package/scripts/service_check.py      |   3 -
 .../SQOOP/package/scripts/sqoop_client.py       |   5 +-
 .../SYSTEMML/package/scripts/systemml_client.py |   5 +-
 .../TITAN/package/scripts/titan_client.py       |   5 +-
 .../scripts/application_timeline_server.py      |   5 +-
 .../YARN/package/scripts/historyserver.py       |   5 +-
 .../YARN/package/scripts/mapreduce2_client.py   |   5 +-
 .../YARN/package/scripts/nodemanager.py         |   5 +-
 .../YARN/package/scripts/resourcemanager.py     |   5 +-
 .../YARN/package/scripts/yarn_client.py         |   5 +-
 .../ZOOKEEPER/package/scripts/zookeeper.py      |   2 +-
 .../package/scripts/zookeeper_client.py         |   5 +-
 .../package/scripts/zookeeper_server.py         |   5 +-
 .../HDP/2.0.6/configuration/cluster-env.xml     |  15 +
 .../2.0.6/hooks/after-INSTALL/scripts/hook.py   |   2 +-
 .../2.0.6/hooks/after-INSTALL/scripts/params.py |   3 -
 .../scripts/shared_initialization.py            |  39 +-
 .../2.0.6/properties/stack_select_packages.json | 952 +++++++++++++++++++
 .../python/stacks/2.0.6/FLUME/test_flume.py     |   5 +-
 .../stacks/2.0.6/HBASE/test_hbase_client.py     |   5 +-
 .../stacks/2.0.6/HBASE/test_hbase_master.py     |   4 +
 .../2.0.6/HBASE/test_hbase_regionserver.py      |   4 +
 .../2.0.6/HBASE/test_phoenix_queryserver.py     |   3 +
 .../python/stacks/2.0.6/HDFS/test_datanode.py   |   4 +
 .../stacks/2.0.6/HDFS/test_hdfs_client.py       |   5 +
 .../stacks/2.0.6/HDFS/test_journalnode.py       |   4 +
 .../python/stacks/2.0.6/HDFS/test_namenode.py   |   4 +
 .../python/stacks/2.0.6/HDFS/test_nfsgateway.py |   3 +
 .../stacks/2.0.6/HIVE/test_hcat_client.py       |   3 +
 .../stacks/2.0.6/HIVE/test_hive_client.py       |   6 +
 .../stacks/2.0.6/HIVE/test_webhcat_server.py    |   4 +
 .../stacks/2.0.6/OOZIE/test_oozie_client.py     |   4 +
 .../stacks/2.0.6/OOZIE/test_oozie_server.py     |  11 +
 .../python/stacks/2.0.6/PIG/test_pig_client.py  |   4 +
 .../python/stacks/2.0.6/SQOOP/test_sqoop.py     |   3 +
 .../stacks/2.0.6/YARN/test_historyserver.py     |   5 +-
 .../stacks/2.0.6/YARN/test_mapreduce2_client.py |   5 +
 .../stacks/2.0.6/YARN/test_nodemanager.py       |   6 +
 .../stacks/2.0.6/YARN/test_resourcemanager.py   |   3 +
 .../stacks/2.0.6/YARN/test_yarn_client.py       |   4 +
 .../2.0.6/ZOOKEEPER/test_zookeeper_client.py    |   4 +
 .../2.0.6/ZOOKEEPER/test_zookeeper_server.py    |   5 +
 .../hooks/after-INSTALL/test_after_install.py   |  43 +-
 .../stacks/2.1/FALCON/test_falcon_client.py     |   4 +
 .../stacks/2.1/FALCON/test_falcon_server.py     |   5 +
 .../stacks/2.1/HIVE/test_hive_metastore.py      |  26 +-
 .../stacks/2.1/STORM/test_storm_drpc_server.py  |   4 +
 .../stacks/2.1/STORM/test_storm_nimbus.py       |   3 +
 .../stacks/2.1/STORM/test_storm_nimbus_prod.py  |   3 +
 .../stacks/2.1/STORM/test_storm_supervisor.py   |   4 +
 .../2.1/STORM/test_storm_supervisor_prod.py     |   4 +
 .../stacks/2.1/STORM/test_storm_ui_server.py    |   5 +-
 .../python/stacks/2.1/TEZ/test_tez_client.py    |   4 +
 .../stacks/2.1/YARN/test_apptimelineserver.py   |   3 +
 .../stacks/2.2/ACCUMULO/test_accumulo_client.py |   4 +
 .../stacks/2.2/KAFKA/test_kafka_broker.py       |   4 +
 .../stacks/2.2/SLIDER/test_slider_client.py     |   6 +
 .../stacks/2.2/SPARK/test_job_history_server.py |   3 +
 .../stacks/2.2/SPARK/test_spark_client.py       |   4 +
 .../stacks/2.3/MAHOUT/test_mahout_client.py     |   4 +
 .../2.3/SPARK/test_spark_thrift_server.py       |   3 +
 .../src/test/python/stacks/utils/RMFTestCase.py |  12 +
 219 files changed, 2320 insertions(+), 807 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-common/src/main/python/resource_management/libraries/functions/stack_select.py
----------------------------------------------------------------------
diff --git 
a/ambari-common/src/main/python/resource_management/libraries/functions/stack_select.py
 
b/ambari-common/src/main/python/resource_management/libraries/functions/stack_select.py
index 265e7df..9e28907 100644
--- 
a/ambari-common/src/main/python/resource_management/libraries/functions/stack_select.py
+++ 
b/ambari-common/src/main/python/resource_management/libraries/functions/stack_select.py
@@ -22,6 +22,7 @@ limitations under the License.
 import os
 import sys
 import re
+import ambari_simplejson as json
 
 # Local Imports
 from resource_management.core.logger import Logger
@@ -40,50 +41,6 @@ from resource_management.libraries.functions import 
StackFeature
 
 STACK_SELECT_PREFIX = 'ambari-python-wrap'
 
-# a mapping of Ambari server role to <stack-selector-tool> component name for 
all
-# non-clients
-SERVER_ROLE_DIRECTORY_MAP = {
-  'ACCUMULO_MASTER' : 'accumulo-master',
-  'ACCUMULO_MONITOR' : 'accumulo-monitor',
-  'ACCUMULO_GC' : 'accumulo-gc',
-  'ACCUMULO_TRACER' : 'accumulo-tracer',
-  'ACCUMULO_TSERVER' : 'accumulo-tablet',
-  'ATLAS_SERVER' : 'atlas-server',
-  'FLUME_HANDLER' : 'flume-server',
-  'FALCON_SERVER' : 'falcon-server',
-  'NAMENODE' : 'hadoop-hdfs-namenode',
-  'DATANODE' : 'hadoop-hdfs-datanode',
-  'SECONDARY_NAMENODE' : 'hadoop-hdfs-secondarynamenode',
-  'NFS_GATEWAY' : 'hadoop-hdfs-nfs3',
-  'JOURNALNODE' : 'hadoop-hdfs-journalnode',
-  'HBASE_MASTER' : 'hbase-master',
-  'HBASE_REGIONSERVER' : 'hbase-regionserver',
-  'HIVE_METASTORE' : 'hive-metastore',
-  'HIVE_SERVER' : 'hive-server2',
-  'HIVE_SERVER_INTERACTIVE' : 'hive-server2-hive2',
-  'WEBHCAT_SERVER' : 'hive-webhcat',
-  'KAFKA_BROKER' : 'kafka-broker',
-  'KNOX_GATEWAY' : 'knox-server',
-  'OOZIE_SERVER' : 'oozie-server',
-  'RANGER_ADMIN' : 'ranger-admin',
-  'RANGER_USERSYNC' : 'ranger-usersync',
-  'RANGER_TAGSYNC' : 'ranger-tagsync',
-  'RANGER_KMS' : 'ranger-kms',
-  'SPARK_JOBHISTORYSERVER' : 'spark-historyserver',
-  'SPARK_THRIFTSERVER' : 'spark-thriftserver',
-  'NIMBUS' : 'storm-nimbus',
-  'SUPERVISOR' : 'storm-supervisor',
-  'HISTORYSERVER' : 'hadoop-mapreduce-historyserver',
-  'APP_TIMELINE_SERVER' : 'hadoop-yarn-timelineserver',
-  'NODEMANAGER' : 'hadoop-yarn-nodemanager',
-  'RESOURCEMANAGER' : 'hadoop-yarn-resourcemanager',
-  'ZOOKEEPER_SERVER' : 'zookeeper-server',
-
-  # ZKFC is tied to NN since it doesn't have its own componnet in 
<stack-selector-tool> and there is
-  # a requirement that the ZKFC is installed on each NN
-  'ZKFC' : 'hadoop-hdfs-namenode'
-}
-
 # mapping of service check to <stack-selector-tool> component
 SERVICE_CHECK_DIRECTORY_MAP = {
   "HDFS_SERVICE_CHECK" : "hadoop-client",
@@ -113,6 +70,110 @@ HADOOP_DIR_DEFAULTS = {
   "lib": "/usr/lib/hadoop/lib"
 }
 
+PACKAGE_SCOPE_INSTALL = "INSTALL"
+PACKAGE_SCOPE_STANDARD = "STANDARD"
+PACKAGE_SCOPE_PATCH = "PATCH"
+PACKAGE_SCOPE_STACK_SELECT = "STACK-SELECT-PACKAGE"
+_PACKAGE_SCOPES = (PACKAGE_SCOPE_INSTALL, PACKAGE_SCOPE_STANDARD, 
PACKAGE_SCOPE_PATCH, PACKAGE_SCOPE_STACK_SELECT)
+
+
+def get_package_name(default_package = None):
+  """
+  Gets the stack-select package name for the service name and
+  component from the current command. Not all services/components are used 
with the
+  stack-select tools, so those will return no packages.
+
+  :return:  the stack-select package name for the command's component or None
+  """
+  config = Script.get_config()
+
+  if 'role' not in config or 'serviceName' not in config:
+    raise Fail("Both the role and the service name must be included in the 
command in order to determine which packages to use with the stack-select tool")
+
+  service_name = config['serviceName']
+  component_name = config['role']
+
+  # should return a single item
+  try:
+    package = get_packages(PACKAGE_SCOPE_STACK_SELECT, service_name, 
component_name)
+    if package is None:
+      package = default_package
+
+    return package
+  except:
+    if default_package is not None:
+      return default_package
+    else:
+      raise
+
+
+
+def get_packages(scope, service_name = None, component_name = None):
+  """
+  Gets the packages which should be used with the stack's stack-select tool 
for the
+  specified service/component. Not all services/components are used with the 
stack-select tools,
+  so those will return no packages.
+
+  :param scope: the scope of the command
+  :param service_name:  the service name, such as ZOOKEEPER
+  :param component_name: the component name, such as ZOOKEEPER_SERVER
+  :return:  the packages to use with stack-select or None
+  """
+  from resource_management.libraries.functions.default import default
+
+  import time
+
+  if scope not in _PACKAGE_SCOPES:
+    raise Fail("The specified scope of {0} is not valid".format(scope))
+
+  config = Script.get_config()
+
+  if service_name is None or component_name is None:
+    if 'role' not in config or 'serviceName' not in config:
+      raise Fail("Both the role and the service name must be included in the 
command in order to determine which packages to use with the stack-select tool")
+
+    service_name = config['serviceName']
+    component_name = config['role']
+
+
+  stack_name = default("/hostLevelParams/stack_name", None)
+  if stack_name is None:
+    raise Fail("The stack name is not present in the command. Packages for 
stack-select tool cannot be loaded.")
+
+  stack_select_packages_config = 
default("/configurations/cluster-env/stack_select_packages", None)
+  if stack_select_packages_config is None:
+    raise Fail("The stack packages are not defined on the command. Unable to 
load packages for the stack-select tool")
+
+  data = json.loads(stack_select_packages_config)
+
+  if stack_name not in data:
+    raise Fail(
+      "Cannot find stack-select packages for the {0} stack".format(stack_name))
+
+  stack_select_key = "stack-select"
+  data = data[stack_name]
+  if stack_select_key not in data:
+    raise Fail(
+      "There are no stack-select packages defined for this command for the {0} 
stack".format(stack_name))
+
+  # this should now be the dictionary of role name to package name
+  data = data[stack_select_key]
+  service_name = service_name.upper()
+  component_name = component_name.upper()
+
+  if service_name not in data:
+    Logger.info("Skipping stack-select on {0} because it does not exist in the 
stack-select package structure.".format(service_name))
+    return None
+
+  data = data[service_name]
+
+  if component_name not in data:
+    Logger.info("Skipping stack-select on {0} because it does not exist in the 
stack-select package structure.".format(component_name))
+    return None
+
+  return data[component_name][scope]
+
+
 def select_all(version_to_select):
   """
   Executes <stack-selector-tool> on every component for the specified version. 
If the value passed in is a
@@ -135,6 +196,20 @@ def select_all(version_to_select):
   Execute(command, only_if = only_if_command)
 
 
+def select_packages(version):
+  """
+  Uses the command's service and role to determine the stack-select packages 
which need to be invoked.
+  :param version: the version to select
+  :return: None
+  """
+  stack_select_packages = get_packages(PACKAGE_SCOPE_STANDARD)
+  if stack_select_packages is None:
+    return
+
+  for stack_select_package_name in stack_select_packages:
+    select(stack_select_package_name, version)
+
+
 def select(component, version):
   """
   Executes <stack-selector-tool> on the specific component and version. Some 
global
@@ -170,15 +245,15 @@ def get_role_component_current_stack_version():
   Gets the current HDP version of the component that this role command is for.
   :return:  the current HDP version of the specified component or None
   """
-  stack_select_component = None
   role = default("/role", "")
   role_command =  default("/roleCommand", "")
+
   stack_selector_name = 
stack_tools.get_stack_tool_name(stack_tools.STACK_SELECTOR_NAME)
 
-  if role in SERVER_ROLE_DIRECTORY_MAP:
-    stack_select_component = SERVER_ROLE_DIRECTORY_MAP[role]
-  elif role_command == "SERVICE_CHECK" and role in SERVICE_CHECK_DIRECTORY_MAP:
+  if role_command == "SERVICE_CHECK" and role in SERVICE_CHECK_DIRECTORY_MAP:
     stack_select_component = SERVICE_CHECK_DIRECTORY_MAP[role]
+  else:
+    stack_select_component = get_package_name()
 
   if stack_select_component is None:
     return None

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-common/src/main/python/resource_management/libraries/script/script.py
----------------------------------------------------------------------
diff --git 
a/ambari-common/src/main/python/resource_management/libraries/script/script.py 
b/ambari-common/src/main/python/resource_management/libraries/script/script.py
index cce3ac4..abbada4 100644
--- 
a/ambari-common/src/main/python/resource_management/libraries/script/script.py
+++ 
b/ambari-common/src/main/python/resource_management/libraries/script/script.py
@@ -182,12 +182,6 @@ class Script(object):
     except IOError, err:
       Script.structuredOut.update({"errMsg" : "Unable to write to " + 
self.stroutfile})
 
-  def get_component_name(self):
-    """
-    To be overridden by subclasses.
-     Returns a string with the component name used in selecting the version.
-    """
-    pass
 
   def get_config_dir_during_stack_upgrade(self, env, base_dir, 
conf_select_name):
     """
@@ -217,11 +211,13 @@ class Script(object):
     :param stack_name: One of HDP, HDPWIN, PHD, BIGTOP.
     :return: Append the version number to the structured out.
     """
+    from resource_management.libraries.functions import stack_select
+
     stack_name = Script.get_stack_name()
-    component_name = self.get_component_name()
+    stack_select_package_name = stack_select.get_package_name()
 
-    if component_name and stack_name:
-      component_version = get_component_version(stack_name, component_name)
+    if stack_select_package_name and stack_name:
+      component_version = get_component_version(stack_name, 
stack_select_package_name)
 
       if component_version:
         self.put_structured_out({"version": component_version})
@@ -426,11 +422,12 @@ class Script(object):
 
     :return: stack version including the build number. e.g.: 2.3.4.0-1234.
     """
+    from resource_management.libraries.functions import stack_select
+
     # preferred way is to get the actual selected version of current component
-    component_name = self.get_component_name()
-    if not Script.stack_version_from_distro_select and component_name:
-      from resource_management.libraries.functions import stack_select
-      Script.stack_version_from_distro_select = 
stack_select.get_stack_version_before_install(component_name)
+    stack_select_package_name = stack_select.get_package_name()
+    if not Script.stack_version_from_distro_select and 
stack_select_package_name:
+      Script.stack_version_from_distro_select = 
stack_select.get_stack_version_before_install(stack_select_package_name)
 
     # If <stack-selector-tool> has not yet been done (situations like first 
install),
     # we can use <stack-selector-tool> version itself.

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/java/org/apache/ambari/server/controller/internal/BlueprintConfigurationProcessor.java
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/java/org/apache/ambari/server/controller/internal/BlueprintConfigurationProcessor.java
 
b/ambari-server/src/main/java/org/apache/ambari/server/controller/internal/BlueprintConfigurationProcessor.java
index 50cea9e..ec8b38c 100644
--- 
a/ambari-server/src/main/java/org/apache/ambari/server/controller/internal/BlueprintConfigurationProcessor.java
+++ 
b/ambari-server/src/main/java/org/apache/ambari/server/controller/internal/BlueprintConfigurationProcessor.java
@@ -2953,14 +2953,15 @@ public class BlueprintConfigurationProcessor {
 
     Set<String> properties = 
Sets.newHashSet(ConfigHelper.CLUSTER_ENV_STACK_NAME_PROPERTY,
         ConfigHelper.CLUSTER_ENV_STACK_ROOT_PROPERTY, 
ConfigHelper.CLUSTER_ENV_STACK_TOOLS_PROPERTY,
-        ConfigHelper.CLUSTER_ENV_STACK_FEATURES_PROPERTY);
+        ConfigHelper.CLUSTER_ENV_STACK_FEATURES_PROPERTY,
+        ConfigHelper.CLUSTER_ENV_STACK_SELECT_PACKAGES_PROPERTY);
 
     try {
       Map<String, Map<String, String>> defaultStackProperties = 
configHelper.getDefaultStackProperties(stackId);
       Map<String,String> clusterEnvDefaultProperties = 
defaultStackProperties.get(CLUSTER_ENV_CONFIG_TYPE_NAME);
 
       for( String property : properties ){
-        if (defaultStackProperties.containsKey(property)) {
+        if (clusterEnvDefaultProperties.containsKey(property)) {
           configuration.setProperty(CLUSTER_ENV_CONFIG_TYPE_NAME, property,
               clusterEnvDefaultProperties.get(property));
 

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/java/org/apache/ambari/server/state/ConfigHelper.java
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/java/org/apache/ambari/server/state/ConfigHelper.java 
b/ambari-server/src/main/java/org/apache/ambari/server/state/ConfigHelper.java
index 974ad4f..8270370 100644
--- 
a/ambari-server/src/main/java/org/apache/ambari/server/state/ConfigHelper.java
+++ 
b/ambari-server/src/main/java/org/apache/ambari/server/state/ConfigHelper.java
@@ -92,6 +92,7 @@ public class ConfigHelper {
   public static final String CLUSTER_ENV_STACK_FEATURES_PROPERTY = 
"stack_features";
   public static final String CLUSTER_ENV_STACK_TOOLS_PROPERTY = "stack_tools";
   public static final String CLUSTER_ENV_STACK_ROOT_PROPERTY = "stack_root";
+  public static final String CLUSTER_ENV_STACK_SELECT_PACKAGES_PROPERTY = 
"stack_select_packages";
 
   public static final String HTTP_ONLY = "HTTP_ONLY";
   public static final String HTTPS_ONLY = "HTTPS_ONLY";

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/java/org/apache/ambari/server/upgrade/FinalUpgradeCatalog.java
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/java/org/apache/ambari/server/upgrade/FinalUpgradeCatalog.java
 
b/ambari-server/src/main/java/org/apache/ambari/server/upgrade/FinalUpgradeCatalog.java
index dad0ecf..55f9665 100644
--- 
a/ambari-server/src/main/java/org/apache/ambari/server/upgrade/FinalUpgradeCatalog.java
+++ 
b/ambari-server/src/main/java/org/apache/ambari/server/upgrade/FinalUpgradeCatalog.java
@@ -63,6 +63,7 @@ public class FinalUpgradeCatalog extends 
AbstractFinalUpgradeCatalog {
    * <ul>
    * <li>Adds/Updates {@link ConfigHelper#CLUSTER_ENV_STACK_FEATURES_PROPERTY} 
from stack</li>
    * <li>Adds/Updates {@link ConfigHelper#CLUSTER_ENV_STACK_TOOLS_PROPERTY} 
from stack</li>
+   * <li>Adds/Updates {@link 
ConfigHelper#CLUSTER_ENV_STACK_SELECT_PACKAGES_PROPERTY} from stack</li>
    * </ul>
    *
    * Note: Config properties stack_features and stack_tools should always be 
updated to latest values as defined
@@ -84,7 +85,8 @@ public class FinalUpgradeCatalog extends 
AbstractFinalUpgradeCatalog {
       List<PropertyInfo> properties = stackInfo.getProperties();
       for(PropertyInfo property : properties) {
         
if(property.getName().equals(ConfigHelper.CLUSTER_ENV_STACK_FEATURES_PROPERTY) 
||
-            
property.getName().equals(ConfigHelper.CLUSTER_ENV_STACK_TOOLS_PROPERTY)) {
+            
property.getName().equals(ConfigHelper.CLUSTER_ENV_STACK_TOOLS_PROPERTY) || 
+            
property.getName().equals(ConfigHelper.CLUSTER_ENV_STACK_SELECT_PACKAGES_PROPERTY))
 {
           propertyMap.put(property.getName(), property.getValue());
         }
       }

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/ACCUMULO/1.6.1.2.2.0/package/scripts/accumulo_client.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/ACCUMULO/1.6.1.2.2.0/package/scripts/accumulo_client.py
 
b/ambari-server/src/main/resources/common-services/ACCUMULO/1.6.1.2.2.0/package/scripts/accumulo_client.py
index 67ca525..856446c 100644
--- 
a/ambari-server/src/main/resources/common-services/ACCUMULO/1.6.1.2.2.0/package/scripts/accumulo_client.py
+++ 
b/ambari-server/src/main/resources/common-services/ACCUMULO/1.6.1.2.2.0/package/scripts/accumulo_client.py
@@ -30,9 +30,6 @@ from accumulo_configuration import setup_conf_dir
 
 
 class AccumuloClient(Script):
-  def get_component_name(self):
-    return "accumulo-client"
-
   def install(self, env):
     self.install_packages(env)
     self.configure(env)
@@ -60,7 +57,7 @@ class AccumuloClient(Script):
 
     Logger.info("Executing Accumulo Client Upgrade pre-restart")
     conf_select.select(params.stack_name, "accumulo", params.version)
-    stack_select.select("accumulo-client", params.version)
+    stack_select.select_packages(params.version)
 
 if __name__ == "__main__":
   AccumuloClient().execute()

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/ACCUMULO/1.6.1.2.2.0/package/scripts/accumulo_script.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/ACCUMULO/1.6.1.2.2.0/package/scripts/accumulo_script.py
 
b/ambari-server/src/main/resources/common-services/ACCUMULO/1.6.1.2.2.0/package/scripts/accumulo_script.py
index ebd418d..a0519b2 100644
--- 
a/ambari-server/src/main/resources/common-services/ACCUMULO/1.6.1.2.2.0/package/scripts/accumulo_script.py
+++ 
b/ambari-server/src/main/resources/common-services/ACCUMULO/1.6.1.2.2.0/package/scripts/accumulo_script.py
@@ -37,33 +37,10 @@ from accumulo_service import accumulo_service
 
 class AccumuloScript(Script):
 
-  # a mapping between the component named used by these scripts and the name
-  # which is used by <stack-selector-tool>
-  COMPONENT_TO_STACK_SELECT_MAPPING = {
-    "gc" : "accumulo-gc",
-    "master" : "accumulo-master",
-    "monitor" : "accumulo-monitor",
-    "tserver" : "accumulo-tablet",
-    "tracer" : "accumulo-tracer"
-  }
-
   def __init__(self, component):
     self.component = component
 
 
-  def get_component_name(self):
-    """
-    Gets the <stack-selector-tool> component name given the script component
-    :return:  the name of the component on the stack which is used by
-              <stack-selector-tool>
-    """
-    if self.component not in self.COMPONENT_TO_STACK_SELECT_MAPPING:
-      return None
-
-    stack_component = self.COMPONENT_TO_STACK_SELECT_MAPPING[self.component]
-    return stack_component
-
-
   def install(self, env):
     self.install_packages(env)
 
@@ -107,19 +84,12 @@ class AccumuloScript(Script):
     if not (params.stack_version_formatted and 
check_stack_feature(StackFeature.ROLLING_UPGRADE, 
params.stack_version_formatted)):
       return
 
-    if self.component not in self.COMPONENT_TO_STACK_SELECT_MAPPING:
-      Logger.info("Unable to execute an upgrade for unknown component 
{0}".format(self.component))
-      raise Fail("Unable to execute an upgrade for unknown component 
{0}".format(self.component))
-
-    stack_component = self.COMPONENT_TO_STACK_SELECT_MAPPING[self.component]
+    stack_component = stack_select.get_package_name()
 
     Logger.info("Executing Accumulo Upgrade pre-restart for 
{0}".format(stack_component))
     conf_select.select(params.stack_name, "accumulo", params.version)
-    stack_select.select(stack_component, params.version)
+    stack_select.select_packages(params.version)
 
-    # some accumulo components depend on the client, so update that too
-    stack_select.select("accumulo-client", params.version)
-      
   def get_log_folder(self):
     import params
     return params.log_dir

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/ATLAS/0.1.0.2.3/package/scripts/atlas_client.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/ATLAS/0.1.0.2.3/package/scripts/atlas_client.py
 
b/ambari-server/src/main/resources/common-services/ATLAS/0.1.0.2.3/package/scripts/atlas_client.py
index d01ff84..6e4a168 100644
--- 
a/ambari-server/src/main/resources/common-services/ATLAS/0.1.0.2.3/package/scripts/atlas_client.py
+++ 
b/ambari-server/src/main/resources/common-services/ATLAS/0.1.0.2.3/package/scripts/atlas_client.py
@@ -30,16 +30,13 @@ from metadata import metadata
 
 class AtlasClient(Script):
 
-  def get_component_name(self):
-    return "atlas-client"
-
   def pre_upgrade_restart(self, env, upgrade_type=None):
     import params
     env.set_params(params)
 
     if check_stack_feature(StackFeature.ATLAS_UPGRADE_SUPPORT, 
params.version_for_stack_feature_checks):
       conf_select.select(params.stack_name, "atlas", params.version)
-      stack_select.select("atlas-client", params.version)
+      stack_select.select_packages(params.version)
 
   def install(self, env):
     self.install_packages(env)

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/ATLAS/0.1.0.2.3/package/scripts/metadata_server.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/ATLAS/0.1.0.2.3/package/scripts/metadata_server.py
 
b/ambari-server/src/main/resources/common-services/ATLAS/0.1.0.2.3/package/scripts/metadata_server.py
index 38f9a41..ff7bf5f 100644
--- 
a/ambari-server/src/main/resources/common-services/ATLAS/0.1.0.2.3/package/scripts/metadata_server.py
+++ 
b/ambari-server/src/main/resources/common-services/ATLAS/0.1.0.2.3/package/scripts/metadata_server.py
@@ -38,10 +38,6 @@ from setup_ranger_atlas import setup_ranger_atlas
 from resource_management.core.resources.zkmigrator import ZkMigrator
 
 class MetadataServer(Script):
-
-  def get_component_name(self):
-    return "atlas-server"
-
   def install(self, env):
     import params
     env.set_params(params)
@@ -63,7 +59,7 @@ class MetadataServer(Script):
 
     if check_stack_feature(StackFeature.ATLAS_UPGRADE_SUPPORT, params.version):
       conf_select.select(params.stack_name, "atlas", params.version)
-      stack_select.select("atlas-server", params.version)
+      stack_select.select_packages(params.version)
 
   def start(self, env, upgrade_type=None):
     import params

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/DRUID/0.9.2/package/scripts/druid_node.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/DRUID/0.9.2/package/scripts/druid_node.py
 
b/ambari-server/src/main/resources/common-services/DRUID/0.9.2/package/scripts/druid_node.py
index 7c6bf39..20623f7 100644
--- 
a/ambari-server/src/main/resources/common-services/DRUID/0.9.2/package/scripts/druid_node.py
+++ 
b/ambari-server/src/main/resources/common-services/DRUID/0.9.2/package/scripts/druid_node.py
@@ -35,10 +35,6 @@ class DruidBase(Script):
   def __init__(self, nodeType=None):
     self.nodeType = nodeType
 
-  def get_component_name(self):
-    node_type_lower = self.nodeType.lower()
-    return format("druid-{node_type_lower}")
-
   def install(self, env):
     self.install_packages(env)
 
@@ -55,7 +51,7 @@ class DruidBase(Script):
     env.set_params(params)
 
     if params.stack_version and 
check_stack_feature(StackFeature.ROLLING_UPGRADE, params.stack_version):
-      stack_select.select(self.get_component_name(), params.stack_version)
+      stack_select.select_packages(params.stack_version)
     if params.stack_version and 
check_stack_feature(StackFeature.CONFIG_VERSIONING, params.stack_version):
       conf_select.select(params.stack_name, "druid", params.stack_version)
 

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/DRUID/0.9.2/package/scripts/superset.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/DRUID/0.9.2/package/scripts/superset.py
 
b/ambari-server/src/main/resources/common-services/DRUID/0.9.2/package/scripts/superset.py
index b837b24..36dab51 100644
--- 
a/ambari-server/src/main/resources/common-services/DRUID/0.9.2/package/scripts/superset.py
+++ 
b/ambari-server/src/main/resources/common-services/DRUID/0.9.2/package/scripts/superset.py
@@ -36,9 +36,6 @@ from resource_management.libraries.resources.properties_file 
import PropertiesFi
 
 class Superset(Script):
 
-  def get_component_name(self):
-    return format("druid-superset")
-
   def install(self, env):
     self.install_packages(env)
 
@@ -98,7 +95,7 @@ class Superset(Script):
     env.set_params(params)
 
     if params.stack_version and 
check_stack_feature(StackFeature.ROLLING_UPGRADE, params.stack_version):
-      stack_select.select(self.get_component_name(), params.stack_version)
+      stack_select.select_packages(params.version)
     if params.stack_version and 
check_stack_feature(StackFeature.CONFIG_VERSIONING, params.stack_version):
       conf_select.select(params.stack_name, "superset", params.stack_version)
 

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/falcon_client.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/falcon_client.py
 
b/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/falcon_client.py
index 365f661..85d80ba 100644
--- 
a/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/falcon_client.py
+++ 
b/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/falcon_client.py
@@ -38,9 +38,6 @@ class FalconClient(Script):
 
 @OsFamilyImpl(os_family=OsFamilyImpl.DEFAULT)
 class FalconClientLinux(FalconClient):
-  def get_component_name(self):
-    return "falcon-client"
-
   def install(self, env):
     self.install_packages(env)
     self.configure(env)
@@ -57,7 +54,7 @@ class FalconClientLinux(FalconClient):
 
     Logger.info("Executing Falcon Client Stack Upgrade pre-restart")
     conf_select.select(params.stack_name, "falcon", params.version)
-    stack_select.select("falcon-client", params.version)
+    stack_select.select_packages(params.version)
 
 @OsFamilyImpl(os_family=OSConst.WINSRV_FAMILY)
 class FalconClientWindows(FalconClient):

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/falcon_server.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/falcon_server.py
 
b/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/falcon_server.py
index 5b2db44..c7ba556 100644
--- 
a/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/falcon_server.py
+++ 
b/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/falcon_server.py
@@ -60,9 +60,6 @@ class FalconServer(Script):
 
 @OsFamilyImpl(os_family=OsFamilyImpl.DEFAULT)
 class FalconServerLinux(FalconServer):
-  def get_component_name(self):
-    return "falcon-server"
-
   def install(self, env):
     import params
     self.install_packages(env)
@@ -85,7 +82,7 @@ class FalconServerLinux(FalconServer):
 
     Logger.info("Executing Falcon Server Stack Upgrade pre-restart")
     conf_select.select(params.stack_name, "falcon", params.version)
-    stack_select.select("falcon-server", params.version)
+    stack_select.select_packages(params.version)
 
     falcon_server_upgrade.pre_start_restore()
 

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/flume_handler.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/flume_handler.py
 
b/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/flume_handler.py
index 107ce6d..f57fae7 100644
--- 
a/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/flume_handler.py
+++ 
b/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/flume_handler.py
@@ -40,9 +40,6 @@ class FlumeHandler(Script):
 
 @OsFamilyImpl(os_family=OsFamilyImpl.DEFAULT)
 class FlumeHandlerLinux(FlumeHandler):
-  def get_component_name(self):
-    return "flume-server"
-
   def install(self, env):
     import params
     self.install_packages(env)
@@ -90,7 +87,7 @@ class FlumeHandlerLinux(FlumeHandler):
 
     Logger.info("Executing Flume Stack Upgrade pre-restart")
     conf_select.select(params.stack_name, "flume", params.version)
-    stack_select.select("flume-server", params.version)
+    stack_select.select_packages(params.version)
 
   def get_log_folder(self):
     import params

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/hbase_client.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/hbase_client.py
 
b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/hbase_client.py
index f18a96a..4820d24 100644
--- 
a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/hbase_client.py
+++ 
b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/hbase_client.py
@@ -52,20 +52,16 @@ class HbaseClientWindows(HbaseClient):
 
 @OsFamilyImpl(os_family=OsFamilyImpl.DEFAULT)
 class HbaseClientDefault(HbaseClient):
-  def get_component_name(self):
-    return "hbase-client"
-
   def pre_upgrade_restart(self, env, upgrade_type=None):
     import params
     env.set_params(params)
 
     if params.version and check_stack_feature(StackFeature.ROLLING_UPGRADE, 
params.version): 
       conf_select.select(params.stack_name, "hbase", params.version)
-      stack_select.select("hbase-client", params.version)
 
       # phoenix may not always be deployed
       try:
-        stack_select.select("phoenix-client", params.version)
+        stack_select.select_packages(params.version)
       except Exception as e:
         print "Ignoring error due to missing phoenix-client"
         print str(e)
@@ -75,7 +71,6 @@ class HbaseClientDefault(HbaseClient):
       # of the final "CLIENTS" group and we need to ensure that hadoop-client
       # is also set
       conf_select.select(params.stack_name, "hadoop", params.version)
-      stack_select.select("hadoop-client", params.version)
 
 
 if __name__ == "__main__":

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/hbase_master.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/hbase_master.py
 
b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/hbase_master.py
index 8151572..5932c65 100644
--- 
a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/hbase_master.py
+++ 
b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/hbase_master.py
@@ -76,13 +76,10 @@ class HbaseMasterWindows(HbaseMaster):
 
 @OsFamilyImpl(os_family=OsFamilyImpl.DEFAULT)
 class HbaseMasterDefault(HbaseMaster):
-  def get_component_name(self):
-    return "hbase-master"
-
   def pre_upgrade_restart(self, env, upgrade_type=None):
     import params
     env.set_params(params)
-    upgrade.prestart(env, "hbase-master")
+    upgrade.prestart(env)
 
   def start(self, env, upgrade_type=None):
     import params

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/hbase_regionserver.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/hbase_regionserver.py
 
b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/hbase_regionserver.py
index 9194991..10dde75 100644
--- 
a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/hbase_regionserver.py
+++ 
b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/hbase_regionserver.py
@@ -70,13 +70,10 @@ class HbaseRegionServerWindows(HbaseRegionServer):
 
 @OsFamilyImpl(os_family=OsFamilyImpl.DEFAULT)
 class HbaseRegionServerDefault(HbaseRegionServer):
-  def get_component_name(self):
-    return "hbase-regionserver"
-
   def pre_upgrade_restart(self, env, upgrade_type=None):
     import params
     env.set_params(params)
-    upgrade.prestart(env, "hbase-regionserver")
+    upgrade.prestart(env)
 
   def post_upgrade_restart(self, env, upgrade_type=None):
     import params

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/phoenix_queryserver.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/phoenix_queryserver.py
 
b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/phoenix_queryserver.py
index b1bdb78..f2b44bd 100644
--- 
a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/phoenix_queryserver.py
+++ 
b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/phoenix_queryserver.py
@@ -34,10 +34,6 @@ class PhoenixQueryServer(Script):
     self.install_packages(env)
 
 
-  def get_component_name(self):
-    return "phoenix-server"
-
-
   def configure(self, env):
     import params
     env.set_params(params)
@@ -64,7 +60,7 @@ class PhoenixQueryServer(Script):
     if params.stack_version_formatted and 
check_stack_feature(StackFeature.PHOENIX, params.stack_version_formatted):     
       # phoenix uses hbase configs
       conf_select.select(params.stack_name, "hbase", params.version)
-      stack_select.select("phoenix-server", params.version)
+      stack_select.select_packages(params.version)
 
 
   def status(self, env):

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/upgrade.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/upgrade.py
 
b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/upgrade.py
index 43c7ff4..8ea3616 100644
--- 
a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/upgrade.py
+++ 
b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/upgrade.py
@@ -33,12 +33,12 @@ from resource_management.libraries.functions.format import 
format
 from resource_management.libraries.functions import check_process_status
 
 
-def prestart(env, stack_component):
+def prestart(env):
   import params
 
   if params.version and check_stack_feature(StackFeature.ROLLING_UPGRADE, 
params.version):
     conf_select.select(params.stack_name, "hbase", params.version)
-    stack_select.select(stack_component, params.version)
+    stack_select.select_packages(params.version)
 
 def post_regionserver(env):
   import params

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/datanode.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/datanode.py
 
b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/datanode.py
index c7b813f..ba926cb 100644
--- 
a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/datanode.py
+++ 
b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/datanode.py
@@ -38,14 +38,11 @@ from utils import get_dfsadmin_base_command
 
 class DataNode(Script):
 
-  def get_component_name(self):
-    return "hadoop-hdfs-datanode"
-
   def get_hdfs_binary(self):
     """
     Get the name or path to the hdfs binary depending on the component name.
     """
-    component_name = self.get_component_name()
+    component_name = stack_select.get_package_name()
     return get_hdfs_binary(component_name)
 
 
@@ -133,7 +130,7 @@ class DataNodeDefault(DataNode):
     env.set_params(params)
     if params.version and check_stack_feature(StackFeature.ROLLING_UPGRADE, 
params.version):
       conf_select.select(params.stack_name, "hadoop", params.version)
-      stack_select.select("hadoop-hdfs-datanode", params.version)
+      stack_select.select_packages(params.version)
 
   def post_upgrade_restart(self, env, upgrade_type=None):
     Logger.info("Executing DataNode Stack Upgrade post-restart")

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_client.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_client.py
 
b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_client.py
index 87a6f52..f320eee 100644
--- 
a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_client.py
+++ 
b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_client.py
@@ -57,15 +57,12 @@ class HdfsClient(Script):
 @OsFamilyImpl(os_family=OsFamilyImpl.DEFAULT)
 class HdfsClientDefault(HdfsClient):
 
-  def get_component_name(self):
-    return "hadoop-client"
-
   def pre_upgrade_restart(self, env, upgrade_type=None):
     import params
     env.set_params(params)
     if params.version and check_stack_feature(StackFeature.ROLLING_UPGRADE, 
params.version):
       conf_select.select(params.stack_name, "hadoop", params.version)
-      stack_select.select("hadoop-client", params.version)
+      stack_select.select_packages(params.version)
 
 @OsFamilyImpl(os_family=OSConst.WINSRV_FAMILY)
 class HdfsClientWindows(HdfsClient):

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/journalnode.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/journalnode.py
 
b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/journalnode.py
index 0805ff4..402e7ad 100644
--- 
a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/journalnode.py
+++ 
b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/journalnode.py
@@ -42,9 +42,6 @@ class JournalNode(Script):
 @OsFamilyImpl(os_family=OsFamilyImpl.DEFAULT)
 class JournalNodeDefault(JournalNode):
 
-  def get_component_name(self):
-    return "hadoop-hdfs-journalnode"
-
   def pre_upgrade_restart(self, env, upgrade_type=None):
     Logger.info("Executing Stack Upgrade pre-restart")
     import params
@@ -52,7 +49,7 @@ class JournalNodeDefault(JournalNode):
 
     if params.version and check_stack_feature(StackFeature.ROLLING_UPGRADE, 
params.version):
       conf_select.select(params.stack_name, "hadoop", params.version)
-      stack_select.select("hadoop-hdfs-journalnode", params.version)
+      stack_select.select_packages(params.version)
 
   def start(self, env, upgrade_type=None):
     import params

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/namenode.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/namenode.py
 
b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/namenode.py
index 7f7e30c..aac7977 100644
--- 
a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/namenode.py
+++ 
b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/namenode.py
@@ -67,14 +67,11 @@ except ImportError:
 
 class NameNode(Script):
 
-  def get_component_name(self):
-    return "hadoop-hdfs-namenode"
-
   def get_hdfs_binary(self):
     """
     Get the name or path to the hdfs binary depending on the component name.
     """
-    component_name = self.get_component_name()
+    component_name = stack_select.get_package_name()
     return get_hdfs_binary(component_name)
 
   def install(self, env):
@@ -204,7 +201,7 @@ class NameNodeDefault(NameNode):
     if upgrade_type != constants.UPGRADE_TYPE_NON_ROLLING or 
params.upgrade_direction != Direction.DOWNGRADE:
       conf_select.select(params.stack_name, "hadoop", params.version)
 
-    stack_select.select("hadoop-hdfs-namenode", params.version)
+    stack_select.select_packages(params.version)
 
   def post_upgrade_restart(self, env, upgrade_type=None):
     Logger.info("Executing Stack Upgrade post-restart")

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/nfsgateway.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/nfsgateway.py
 
b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/nfsgateway.py
index 03a497a..dea9d43 100644
--- 
a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/nfsgateway.py
+++ 
b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/nfsgateway.py
@@ -31,10 +31,6 @@ from resource_management.libraries.functions.stack_features 
import check_stack_f
 
 
 class NFSGateway(Script):
-
-  def get_component_name(self):
-    return "hadoop-hdfs-nfs3"
-
   def install(self, env):
     import params
 
@@ -48,7 +44,7 @@ class NFSGateway(Script):
 
     if params.stack_version_formatted and 
check_stack_feature(StackFeature.NFS, params.stack_version_formatted):
       conf_select.select(params.stack_name, "hadoop", params.version)
-      stack_select.select("hadoop-hdfs-nfs3", params.version)
+      stack_select.select_packages(params.version)
 
   def start(self, env, upgrade_type=None):
     import params

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/snamenode.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/snamenode.py
 
b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/snamenode.py
index ac45ffd..115a830 100644
--- 
a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/snamenode.py
+++ 
b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/snamenode.py
@@ -64,9 +64,6 @@ class SNameNode(Script):
 @OsFamilyImpl(os_family=OsFamilyImpl.DEFAULT)
 class SNameNodeDefault(SNameNode):
 
-  def get_component_name(self):
-    return "hadoop-hdfs-secondarynamenode"
-
   def pre_upgrade_restart(self, env, upgrade_type=None):
     Logger.info("Executing Stack Upgrade pre-restart")
     import params
@@ -74,7 +71,7 @@ class SNameNodeDefault(SNameNode):
 
     if params.version and check_stack_feature(StackFeature.ROLLING_UPGRADE, 
params.version):
       conf_select.select(params.stack_name, "hadoop", params.version)
-      stack_select.select("hadoop-hdfs-secondarynamenode", params.version)
+      stack_select.select_packages(params.version)
       
   def get_log_folder(self):
     import params

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/zkfc_slave.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/zkfc_slave.py
 
b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/zkfc_slave.py
index db68544..fabbabb 100644
--- 
a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/zkfc_slave.py
+++ 
b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/zkfc_slave.py
@@ -35,13 +35,6 @@ from resource_management.libraries.script import Script
 from resource_management.core.resources.zkmigrator import ZkMigrator
 
 class ZkfcSlave(Script):
-  def get_component_name(self):
-    import params
-    if params.version_for_stack_feature_checks and 
check_stack_feature(StackFeature.ZKFC_VERSION_ADVERTISED, 
params.version_for_stack_feature_checks):
-      # params.version is not defined when installing cluster from blueprint
-      return "hadoop-hdfs-zkfc"
-    pass
-
   def install(self, env):
     import params
     env.set_params(params)
@@ -137,7 +130,7 @@ class ZkfcSlaveDefault(ZkfcSlave):
     env.set_params(params)
     if check_stack_feature(StackFeature.ZKFC_VERSION_ADVERTISED, 
params.version_for_stack_feature_checks):
       conf_select.select(params.stack_name, "hadoop", params.version)
-      stack_select.select("hadoop-hdfs-zkfc", params.version)
+      stack_select.select_packages(params.version)
 
 def initialize_ha_zookeeper(params):
   try:

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hcat_client.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hcat_client.py
 
b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hcat_client.py
index b37698e..677479f 100644
--- 
a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hcat_client.py
+++ 
b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hcat_client.py
@@ -51,12 +51,6 @@ class HCatClientWindows(HCatClient):
 
 @OsFamilyImpl(os_family=OsFamilyImpl.DEFAULT)
 class HCatClientDefault(HCatClient):
-  def get_component_name(self):
-    # HCat client doesn't have a first-class entry in <stack-selector-tool>. 
Since clients always
-    # update after daemons, this ensures that the hcat directories are correct 
on hosts
-    # which do not include the WebHCat daemon
-    return "hive-webhcat"
-
 
   def pre_upgrade_restart(self, env, upgrade_type=None):
     """
@@ -78,7 +72,7 @@ class HCatClientDefault(HCatClient):
     # HCat client doesn't have a first-class entry in <stack-selector-tool>. 
Since clients always
     # update after daemons, this ensures that the hcat directories are correct 
on hosts
     # which do not include the WebHCat daemon
-    stack_select.select("hive-webhcat", params.version)
+    stack_select.select_packages(params.version)
 
 
 if __name__ == "__main__":

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_client.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_client.py
 
b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_client.py
index 3d9bfd7..1694816 100644
--- 
a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_client.py
+++ 
b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_client.py
@@ -50,9 +50,6 @@ class HiveClientWindows(HiveClient):
 
 @OsFamilyImpl(os_family=OsFamilyImpl.DEFAULT)
 class HiveClientDefault(HiveClient):
-  def get_component_name(self):
-    return "hadoop-client"
-
   def pre_upgrade_restart(self, env, upgrade_type=None):
     Logger.info("Executing Hive client Stack Upgrade pre-restart")
 
@@ -61,7 +58,7 @@ class HiveClientDefault(HiveClient):
     if params.version and check_stack_feature(StackFeature.ROLLING_UPGRADE, 
params.version):
       conf_select.select(params.stack_name, "hive", params.version)
       conf_select.select(params.stack_name, "hadoop", params.version)
-      stack_select.select("hadoop-client", params.version)
+      stack_select.select_packages(params.version)
 
 
 if __name__ == "__main__":

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_metastore.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_metastore.py
 
b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_metastore.py
index 633e82e..c84f47e 100644
--- 
a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_metastore.py
+++ 
b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_metastore.py
@@ -87,10 +87,6 @@ class HiveMetastoreWindows(HiveMetastore):
 
 @OsFamilyImpl(os_family=OsFamilyImpl.DEFAULT)
 class HiveMetastoreDefault(HiveMetastore):
-  def get_component_name(self):
-    return "hive-metastore"
-
-
   def status(self, env):
     import status_params
     from resource_management.libraries.functions import check_process_status
@@ -111,7 +107,7 @@ class HiveMetastoreDefault(HiveMetastore):
 
     if params.version and check_stack_feature(StackFeature.ROLLING_UPGRADE, 
params.version):
       conf_select.select(params.stack_name, "hive", params.version)
-      stack_select.select("hive-metastore", params.version)
+      stack_select.select_packages(params.version)
 
     if is_upgrade and params.stack_version_formatted_major and \
             check_stack_feature(StackFeature.HIVE_METASTORE_UPGRADE_SCHEMA, 
params.stack_version_formatted_major):

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_server.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_server.py
 
b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_server.py
index 8f7d068..a1212e0 100644
--- 
a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_server.py
+++ 
b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_server.py
@@ -72,9 +72,6 @@ class HiveServerWindows(HiveServer):
 
 @OsFamilyImpl(os_family=OsFamilyImpl.DEFAULT)
 class HiveServerDefault(HiveServer):
-  def get_component_name(self):
-    return "hive-server2"
-
   def start(self, env, upgrade_type=None):
     import params
     env.set_params(params)
@@ -118,7 +115,7 @@ class HiveServerDefault(HiveServer):
 
     if params.version and check_stack_feature(StackFeature.ROLLING_UPGRADE, 
params.version):
       conf_select.select(params.stack_name, "hive", params.version)
-      stack_select.select("hive-server2", params.version)
+      stack_select.select_packages(params.version)
 
       # Copy mapreduce.tar.gz and tez.tar.gz to HDFS
       resource_created = copy_to_hdfs(

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_server_interactive.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_server_interactive.py
 
b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_server_interactive.py
index c9582a3..16d7907 100644
--- 
a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_server_interactive.py
+++ 
b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_server_interactive.py
@@ -67,10 +67,6 @@ class HiveServerInteractive(Script):
 
 @OsFamilyImpl(os_family=OsFamilyImpl.DEFAULT)
 class HiveServerInteractiveDefault(HiveServerInteractive):
-
-    def get_component_name(self):
-      return "hive-server2-hive2"
-
     def install(self, env):
       import params
       self.install_packages(env)
@@ -86,7 +82,7 @@ class HiveServerInteractiveDefault(HiveServerInteractive):
       env.set_params(params)
 
       if params.version and check_stack_feature(StackFeature.ROLLING_UPGRADE, 
params.version):
-        stack_select.select("hive-server2-hive2", params.version)
+        stack_select.select_packages(params.version)
         conf_select.select(params.stack_name, "hive2", params.version)
 
         # Copy hive.tar.gz and tez.tar.gz used by Hive Interactive to HDFS

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/webhcat_server.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/webhcat_server.py
 
b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/webhcat_server.py
index da5e82b..db86d9e 100644
--- 
a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/webhcat_server.py
+++ 
b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/webhcat_server.py
@@ -64,9 +64,6 @@ class WebHCatServerWindows(WebHCatServer):
 
 @OsFamilyImpl(os_family=OsFamilyImpl.DEFAULT)
 class WebHCatServerDefault(WebHCatServer):
-  def get_component_name(self):
-    return "hive-webhcat"
-
   def status(self, env):
     import status_params
     env.set_params(status_params)
@@ -81,7 +78,7 @@ class WebHCatServerDefault(WebHCatServer):
       # webhcat has no conf, but uses hadoop home, so verify that regular 
hadoop conf is set
       conf_select.select(params.stack_name, "hive-hcatalog", params.version)
       conf_select.select(params.stack_name, "hadoop", params.version)
-      stack_select.select("hive-webhcat", params.version)
+      stack_select.select_packages(params.version)
 
   def get_log_folder(self):
     import params

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/KAFKA/0.8.1/package/scripts/kafka_broker.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/KAFKA/0.8.1/package/scripts/kafka_broker.py
 
b/ambari-server/src/main/resources/common-services/KAFKA/0.8.1/package/scripts/kafka_broker.py
index 4512038..08c3dd1 100644
--- 
a/ambari-server/src/main/resources/common-services/KAFKA/0.8.1/package/scripts/kafka_broker.py
+++ 
b/ambari-server/src/main/resources/common-services/KAFKA/0.8.1/package/scripts/kafka_broker.py
@@ -36,9 +36,6 @@ from setup_ranger_kafka import setup_ranger_kafka
 
 class KafkaBroker(Script):
 
-  def get_component_name(self):
-    return "kafka-broker"
-
   def install(self, env):
     self.install_packages(env)
 
@@ -52,7 +49,7 @@ class KafkaBroker(Script):
     env.set_params(params)
 
     if params.version and check_stack_feature(StackFeature.ROLLING_UPGRADE, 
params.version):
-      stack_select.select("kafka-broker", params.version)
+      stack_select.select_packages(params.version)
 
     if params.version and check_stack_feature(StackFeature.CONFIG_VERSIONING, 
params.version):
       conf_select.select(params.stack_name, "kafka", params.version)

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/knox_gateway.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/knox_gateway.py
 
b/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/knox_gateway.py
index c91b394..6df68cc 100644
--- 
a/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/knox_gateway.py
+++ 
b/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/knox_gateway.py
@@ -49,9 +49,6 @@ from setup_ranger_knox import setup_ranger_knox
 
 
 class KnoxGateway(Script):
-  def get_component_name(self):
-    return "knox-server"
-
   def install(self, env):
     import params
     env.set_params(params)
@@ -122,7 +119,7 @@ class KnoxGatewayDefault(KnoxGateway):
 
     # <conf-selector-tool> will change the symlink to the conf folder.
     conf_select.select(params.stack_name, "knox", params.version)
-    stack_select.select("knox-server", params.version)
+    stack_select.select_packages(params.version)
 
     # seed the new Knox data directory with the keystores of yesteryear
     if params.upgrade_direction == Direction.UPGRADE:

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/MAHOUT/1.0.0.2.3/package/scripts/mahout_client.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/MAHOUT/1.0.0.2.3/package/scripts/mahout_client.py
 
b/ambari-server/src/main/resources/common-services/MAHOUT/1.0.0.2.3/package/scripts/mahout_client.py
index 3486add..b598d17 100644
--- 
a/ambari-server/src/main/resources/common-services/MAHOUT/1.0.0.2.3/package/scripts/mahout_client.py
+++ 
b/ambari-server/src/main/resources/common-services/MAHOUT/1.0.0.2.3/package/scripts/mahout_client.py
@@ -29,18 +29,13 @@ from resource_management.libraries.functions.default import 
default
 
 class MahoutClient(Script):
 
-  def get_component_name(self):
-    return "mahout-client"
-
-
   def pre_upgrade_restart(self, env, upgrade_type=None):
     Logger.info("Executing Stack Upgrade pre-restart")
     import params
     env.set_params(params)
 
     conf_select.select(params.stack_name, "mahout", params.version)
-    stack_select.select("mahout-client", params.version )
-
+    stack_select.select_packages(params.version)
 
   def install(self, env):
     self.install_packages(env)

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/oozie_client.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/oozie_client.py
 
b/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/oozie_client.py
index 9bbca9f..ac8b078 100644
--- 
a/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/oozie_client.py
+++ 
b/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/oozie_client.py
@@ -30,9 +30,6 @@ from resource_management.core.exceptions import 
ClientComponentHasNoStatus
 
 class OozieClient(Script):
 
-  def get_component_name(self):
-    return "oozie-client"
-
   def install(self, env):
     self.install_packages(env)
     self.configure(env)
@@ -59,7 +56,7 @@ class OozieClient(Script):
 
     Logger.info("Executing Oozie Client Stack Upgrade pre-restart")
     conf_select.select(params.stack_name, "oozie", params.version)
-    stack_select.select("oozie-client", params.version)
+    stack_select.select_packages(params.version)
 
   # We substitute some configs (oozie.authentication.kerberos.principal) 
before generation (see oozie.py and params.py).
   # This function returns changed configs (it's used for config generation 
before config download)

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/oozie_server.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/oozie_server.py
 
b/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/oozie_server.py
index b3a8643..82a764d 100644
--- 
a/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/oozie_server.py
+++ 
b/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/oozie_server.py
@@ -46,9 +46,6 @@ from resource_management.core.resources.zkmigrator import 
ZkMigrator
 
 class OozieServer(Script):
 
-  def get_component_name(self):
-    return "oozie-server"
-
   def install(self, env):
     self.install_packages(env)
 
@@ -67,9 +64,8 @@ class OozieServer(Script):
         # This is required as both need to be pointing to new installed oozie 
version.
 
         # Sets the symlink : eg: <stack-root>/current/oozie-client -> 
<stack-root>/a.b.c.d-<version>/oozie
-        stack_select.select("oozie-client", params.version)
         # Sets the symlink : eg: <stack-root>/current/oozie-server -> 
<stack-root>/a.b.c.d-<version>/oozie
-        stack_select.select("oozie-server", params.version)
+        stack_select.select_packages(params.version)
 
       if params.version and 
check_stack_feature(StackFeature.CONFIG_VERSIONING, params.version):
         conf_select.select(params.stack_name, "oozie", params.version)
@@ -127,7 +123,7 @@ class OozieServerDefault(OozieServer):
 
     if params.version and check_stack_feature(StackFeature.ROLLING_UPGRADE, 
params.version):
       conf_select.select(params.stack_name, "oozie", params.version)
-      stack_select.select("oozie-server", params.version)
+      stack_select.select_packages(params.version)
 
     OozieUpgrade.prepare_libext_directory(upgrade_type=upgrade_type)
 

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/pig_client.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/pig_client.py
 
b/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/pig_client.py
index 3233381..693af5c 100644
--- 
a/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/pig_client.py
+++ 
b/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/pig_client.py
@@ -43,9 +43,6 @@ class PigClient(Script):
 
 @OsFamilyImpl(os_family=OsFamilyImpl.DEFAULT)
 class PigClientLinux(PigClient):
-  def get_component_name(self):
-    return "hadoop-client"
-
   def pre_upgrade_restart(self, env, upgrade_type=None):
     import params
     env.set_params(params)
@@ -53,7 +50,7 @@ class PigClientLinux(PigClient):
     if params.version and check_stack_feature(StackFeature.ROLLING_UPGRADE, 
params.version): 
       conf_select.select(params.stack_name, "pig", params.version)
       conf_select.select(params.stack_name, "hadoop", params.version)
-      stack_select.select("hadoop-client", params.version) # includes 
pig-client
+      stack_select.select_packages(params.version)
 
   def install(self, env):
     self.install_packages(env)

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/R4ML/0.8.0/package/scripts/r4ml_client.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/R4ML/0.8.0/package/scripts/r4ml_client.py
 
b/ambari-server/src/main/resources/common-services/R4ML/0.8.0/package/scripts/r4ml_client.py
index 3dbce5c..f973e69 100755
--- 
a/ambari-server/src/main/resources/common-services/R4ML/0.8.0/package/scripts/r4ml_client.py
+++ 
b/ambari-server/src/main/resources/common-services/R4ML/0.8.0/package/scripts/r4ml_client.py
@@ -34,9 +34,6 @@ class R4MLClient(Script):
     import params
     env.set_params(params)
 
-  def get_component_name(self):
-    return "r4ml-client"
-
   def pre_upgrade_restart(self, env, upgrade_type=None):
     import params
 
@@ -44,7 +41,7 @@ class R4MLClient(Script):
     if params.version and check_stack_feature(StackFeature.ROLLING_UPGRADE, 
params.version):
       Logger.info("Executing R4ML Client Stack Upgrade pre-restart")
       conf_select.select(params.stack_name, "r4ml", params.version)
-      stack_select.select("r4ml-client", params.version)
+      stack_select.select_packages(params.version)
 
   def stack_upgrade_save_new_config(self, env):
     import params

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/ranger_admin.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/ranger_admin.py
 
b/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/ranger_admin.py
index ecbacbf..e27a03e 100644
--- 
a/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/ranger_admin.py
+++ 
b/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/ranger_admin.py
@@ -38,9 +38,6 @@ import os, errno
 
 class RangerAdmin(Script):
 
-  def get_component_name(self):
-    return "ranger-admin"
-
   def install(self, env):
     self.install_packages(env)
     import params
@@ -210,7 +207,7 @@ class RangerAdmin(Script):
     stack_name = upgrade_stack[0]
     stack_version = upgrade_stack[1]
 
-    stack_select.select("ranger-admin", stack_version)
+    stack_select.select_packages(params.version)
     conf_select.select(stack_name, "ranger-admin", stack_version)
 
   def get_log_folder(self):

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/ranger_tagsync.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/ranger_tagsync.py
 
b/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/ranger_tagsync.py
index a474e76..5ff498f 100644
--- 
a/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/ranger_tagsync.py
+++ 
b/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/ranger_tagsync.py
@@ -86,10 +86,7 @@ class RangerTagsync(Script):
     if params.stack_supports_ranger_tagsync:
       Logger.info("Executing Ranger Tagsync Stack Upgrade pre-restart")
       conf_select.select(params.stack_name, "ranger-tagsync", params.version)
-      stack_select.select("ranger-tagsync", params.version)
-
-  def get_component_name(self):
-    return "ranger-tagsync"
+      stack_select.select_packages(params.version)
 
   def get_log_folder(self):
     import params
@@ -111,7 +108,7 @@ class RangerTagsync(Script):
     stack_name = upgrade_stack[0]
     stack_version = upgrade_stack[1]
 
-    stack_select.select("ranger-tagsync", stack_version)
+    stack_select.select_packages(params.version)
     conf_select.select(stack_name, "ranger-tagsync", stack_version)
     if params.stack_supports_ranger_tagsync_ssl_xml_support:
       Logger.info("Upgrading Tagsync, stack support Atlas user for Tagsync, 
creating keystore for same.")

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/ranger_usersync.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/ranger_usersync.py
 
b/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/ranger_usersync.py
index b9366f6..8654bc2 100644
--- 
a/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/ranger_usersync.py
+++ 
b/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/ranger_usersync.py
@@ -109,9 +109,6 @@ class RangerUsersync(Script):
     env.set_params(params)
     upgrade.prestart(env, "ranger-usersync")
 
-  def get_component_name(self):
-    return "ranger-usersync"
-
   def get_log_folder(self):
     import params
     return params.usersync_log_dir

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/upgrade.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/upgrade.py
 
b/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/upgrade.py
index a07a1fd..ca1b2bf 100644
--- 
a/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/upgrade.py
+++ 
b/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/upgrade.py
@@ -28,4 +28,4 @@ def prestart(env, stack_component):
 
   if params.version and params.stack_supports_rolling_upgrade:
     conf_select.select(params.stack_name, stack_component, params.version)
-    stack_select.select(stack_component, params.version)
+    stack_select.select_packages(params.version)

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/kms_server.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/kms_server.py
 
b/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/kms_server.py
index 829a998..0bd11f3 100755
--- 
a/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/kms_server.py
+++ 
b/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/kms_server.py
@@ -33,9 +33,6 @@ import upgrade
 
 class KmsServer(Script):
 
-  def get_component_name(self):
-    return "ranger-kms"
-
   def install(self, env):
     self.install_packages(env)
     import params

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/upgrade.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/upgrade.py
 
b/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/upgrade.py
index 8478bb8..73c32ff 100644
--- 
a/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/upgrade.py
+++ 
b/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/upgrade.py
@@ -27,4 +27,4 @@ def prestart(env, stack_component):
 
   if params.version and params.stack_supports_config_versioning:
     conf_select.select(params.stack_name, stack_component, params.version)
-    stack_select.select(stack_component, params.version)
+    stack_select.select_packages(params.version)

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/slider_client.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/slider_client.py
 
b/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/slider_client.py
index 08c8569..adc7026 100644
--- 
a/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/slider_client.py
+++ 
b/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/slider_client.py
@@ -35,22 +35,18 @@ class SliderClient(Script):
 
 @OsFamilyImpl(os_family=OsFamilyImpl.DEFAULT)
 class SliderClientLinux(SliderClient):
-  def get_component_name(self):
-    return "slider-client"
-
   def pre_upgrade_restart(self, env,  upgrade_type=None):
     import params
     env.set_params(params)
 
     if params.version and check_stack_feature(StackFeature.ROLLING_UPGRADE, 
params.version):
       conf_select.select(params.stack_name, "slider", params.version)
-      stack_select.select("slider-client", params.version)
+      stack_select.select_packages(params.version)
 
       # also set all of the hadoop clients since slider client is upgraded as
       # part of the final "CLIENTS" group and we need to ensure that
       # hadoop-client is also set
       conf_select.select(params.stack_name, "hadoop", params.version)
-      stack_select.select("hadoop-client", params.version)
 
   def install(self, env):
     self.install_packages(env)

http://git-wip-us.apache.org/repos/asf/ambari/blob/56d2ade2/ambari-server/src/main/resources/common-services/SPARK/1.2.1/package/scripts/job_history_server.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/SPARK/1.2.1/package/scripts/job_history_server.py
 
b/ambari-server/src/main/resources/common-services/SPARK/1.2.1/package/scripts/job_history_server.py
index 4fe2f54..c36def4 100644
--- 
a/ambari-server/src/main/resources/common-services/SPARK/1.2.1/package/scripts/job_history_server.py
+++ 
b/ambari-server/src/main/resources/common-services/SPARK/1.2.1/package/scripts/job_history_server.py
@@ -68,9 +68,6 @@ class JobHistoryServer(Script):
     check_process_status(status_params.spark_history_server_pid_file)
     
 
-  def get_component_name(self):
-    return "spark-historyserver"
-
   def pre_upgrade_restart(self, env, upgrade_type=None):
     import params
 
@@ -78,7 +75,7 @@ class JobHistoryServer(Script):
     if params.version and check_stack_feature(StackFeature.ROLLING_UPGRADE, 
params.version):
       Logger.info("Executing Spark Job History Server Stack Upgrade 
pre-restart")
       conf_select.select(params.stack_name, "spark", params.version)
-      stack_select.select("spark-historyserver", params.version)
+      stack_select.select_packages(params.version)
 
       # Spark 1.3.1.2.3, and higher, which was included in HDP 2.3, does not 
have a dependency on Tez, so it does not
       # need to copy the tarball, otherwise, copy it.

Reply via email to