AMBARI-22513 Make yumrpm.py functions to use global defined commands (dgrinenko)
Project: http://git-wip-us.apache.org/repos/asf/ambari/repo Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/02490734 Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/02490734 Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/02490734 Branch: refs/heads/trunk Commit: 0249073469de5cd11f032705d609f8cb19ff28c0 Parents: 99b19e5 Author: Dmytro Grinenko <hapyles...@apache.org> Authored: Fri Nov 24 16:41:12 2017 +0200 Committer: Dmytro Grinenko <hapyles...@apache.org> Committed: Fri Nov 24 16:41:12 2017 +0200 ---------------------------------------------------------------------- .../core/providers/package/yumrpm.py | 4 ++-- .../HIVE/0.12.0.2.0/package/scripts/hive.py | 12 ++++++------ 2 files changed, 8 insertions(+), 8 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/ambari/blob/02490734/ambari-common/src/main/python/resource_management/core/providers/package/yumrpm.py ---------------------------------------------------------------------- diff --git a/ambari-common/src/main/python/resource_management/core/providers/package/yumrpm.py b/ambari-common/src/main/python/resource_management/core/providers/package/yumrpm.py index 8426479..367e2af 100644 --- a/ambari-common/src/main/python/resource_management/core/providers/package/yumrpm.py +++ b/ambari-common/src/main/python/resource_management/core/providers/package/yumrpm.py @@ -138,7 +138,7 @@ class YumProvider(RPMBasedPackageProvider): :rtype list[list,] """ - cmd = [AMBARI_SUDO_BINARY, "yum", "list", "available"] + cmd = list(ALL_AVAILABLE_PACKAGES_CMD) if repo_filter: cmd.extend(["--disablerepo=*", "--enablerepo=" + repo_filter]) @@ -154,7 +154,7 @@ class YumProvider(RPMBasedPackageProvider): :rtype list[list,] """ - packages = self._lookup_packages([AMBARI_SUDO_BINARY, "yum", "list", "installed"], "Installed Packages") + packages = self._lookup_packages(list(ALL_INSTALLED_PACKAGES_CMD), "Installed Packages") if repo_filter: packages = [item for item in packages if item[2].lower() == repo_filter.lower()] http://git-wip-us.apache.org/repos/asf/ambari/blob/02490734/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive.py b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive.py index 716a37c..d9cc55f 100644 --- a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive.py +++ b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive.py @@ -24,7 +24,7 @@ from urlparse import urlparse from resource_management.libraries.script.script import Script from resource_management.libraries.resources.hdfs_resource import HdfsResource -from resource_management.libraries.functions.copy_tarball import copy_to_hdfs +from resource_management.libraries.functions import copy_tarball from resource_management.libraries.functions.get_config import get_config from resource_management.libraries.functions import StackFeature from resource_management.libraries.functions.stack_features import check_stack_feature @@ -177,19 +177,19 @@ def setup_hiveserver2(): # ********************************* # if copy tarball to HDFS feature supported copy mapreduce.tar.gz and tez.tar.gz to HDFS if params.stack_version_formatted_major and check_stack_feature(StackFeature.COPY_TARBALL_TO_HDFS, params.stack_version_formatted_major): - copy_to_hdfs("mapreduce", params.user_group, params.hdfs_user, skip=params.sysprep_skip_copy_tarballs_hdfs) - copy_to_hdfs("tez", params.user_group, params.hdfs_user, skip=params.sysprep_skip_copy_tarballs_hdfs) + copy_tarball.copy_to_hdfs("mapreduce", params.user_group, params.hdfs_user, skip=params.sysprep_skip_copy_tarballs_hdfs) + copy_tarball.copy_to_hdfs("tez", params.user_group, params.hdfs_user, skip=params.sysprep_skip_copy_tarballs_hdfs) # Always copy pig.tar.gz and hive.tar.gz using the appropriate mode. # This can use a different source and dest location to account - copy_to_hdfs("pig", + copy_tarball.copy_to_hdfs("pig", params.user_group, params.hdfs_user, file_mode=params.tarballs_mode, custom_source_file=params.pig_tar_source, custom_dest_file=params.pig_tar_dest_file, skip=params.sysprep_skip_copy_tarballs_hdfs) - copy_to_hdfs("hive", + copy_tarball.copy_to_hdfs("hive", params.user_group, params.hdfs_user, file_mode=params.tarballs_mode, @@ -210,7 +210,7 @@ def setup_hiveserver2(): src_filename = os.path.basename(source_file) dest_file = os.path.join(dest_dir, src_filename) - copy_to_hdfs(tarball_name, + copy_tarball.copy_to_hdfs(tarball_name, params.user_group, params.hdfs_user, file_mode=params.tarballs_mode,