http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/setup_ranger_hive.py ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/setup_ranger_hive.py b/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/setup_ranger_hive.py deleted file mode 100755 index 81a4e3e..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/setup_ranger_hive.py +++ /dev/null @@ -1,98 +0,0 @@ -#!/usr/bin/env python -""" -Licensed to the Apache Software Foundation (ASF) under one -or more contributor license agreements. See the NOTICE file -distributed with this work for additional information -regarding copyright ownership. The ASF licenses this file -to you under the Apache License, Version 2.0 (the -"License"); you may not use this file except in compliance -with the License. You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - -Unless required by applicable law or agreed to in writing, software -distributed under the License is distributed on an "AS IS" BASIS, -WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -See the License for the specific language governing permissions and -limitations under the License. - -""" -from resource_management.core.logger import Logger - -def setup_ranger_hive(upgrade_type = None): - import params - - if params.has_ranger_admin: - - stack_version = None - - if upgrade_type is not None: - stack_version = params.version - - if params.retryAble: - Logger.info("Hive: Setup ranger: command retry enables thus retrying if ranger admin is down !") - else: - Logger.info("Hive: Setup ranger: command retry not enabled thus skipping if ranger admin is down !") - - if params.xml_configurations_supported and params.enable_ranger_hive and params.xa_audit_hdfs_is_enabled: - params.HdfsResource("/ranger/audit", - type="directory", - action="create_on_execute", - owner=params.hdfs_user, - group=params.hdfs_user, - mode=0755, - recursive_chmod=True - ) - params.HdfsResource("/ranger/audit/hiveServer2", - type="directory", - action="create_on_execute", - owner=params.hive_user, - group=params.hive_user, - mode=0700, - recursive_chmod=True - ) - params.HdfsResource(None, action="execute") - - if params.xml_configurations_supported: - api_version=None - if params.stack_supports_ranger_kerberos: - api_version='v2' - from resource_management.libraries.functions.setup_ranger_plugin_xml import setup_ranger_plugin - setup_ranger_plugin('hive-server2', 'hive', params.ranger_previous_jdbc_jar, - params.ranger_downloaded_custom_connector, params.ranger_driver_curl_source, - params.ranger_driver_curl_target, params.java64_home, - params.repo_name, params.hive_ranger_plugin_repo, - params.ranger_env, params.ranger_plugin_properties, - params.policy_user, params.policymgr_mgr_url, - params.enable_ranger_hive, conf_dict=params.hive_server_conf_dir, - component_user=params.hive_user, component_group=params.user_group, cache_service_list=['hiveServer2'], - plugin_audit_properties=params.config['configurations']['ranger-hive-audit'], plugin_audit_attributes=params.config['configuration_attributes']['ranger-hive-audit'], - plugin_security_properties=params.config['configurations']['ranger-hive-security'], plugin_security_attributes=params.config['configuration_attributes']['ranger-hive-security'], - plugin_policymgr_ssl_properties=params.config['configurations']['ranger-hive-policymgr-ssl'], plugin_policymgr_ssl_attributes=params.config['configuration_attributes']['ranger-hive-policymgr-ssl'], - component_list=['hive-client', 'hive-metastore', 'hive-server2'], audit_db_is_enabled=params.xa_audit_db_is_enabled, - credential_file=params.credential_file, xa_audit_db_password=params.xa_audit_db_password, - ssl_truststore_password=params.ssl_truststore_password, ssl_keystore_password=params.ssl_keystore_password, - stack_version_override = stack_version, skip_if_rangeradmin_down= not params.retryAble, api_version=api_version, - is_security_enabled = params.security_enabled, - is_stack_supports_ranger_kerberos = params.stack_supports_ranger_kerberos, - component_user_principal=params.hive_principal if params.security_enabled else None, - component_user_keytab=params.hive_server2_keytab if params.security_enabled else None) - else: - from resource_management.libraries.functions.setup_ranger_plugin import setup_ranger_plugin - setup_ranger_plugin('hive-server2', 'hive', params.ranger_previous_jdbc_jar, - params.ranger_downloaded_custom_connector, params.ranger_driver_curl_source, - params.ranger_driver_curl_target, params.java64_home, - params.repo_name, params.hive_ranger_plugin_repo, - params.ranger_env, params.ranger_plugin_properties, - params.policy_user, params.policymgr_mgr_url, - params.enable_ranger_hive, conf_dict=params.hive_server_conf_dir, - component_user=params.hive_user, component_group=params.user_group, cache_service_list=['hiveServer2'], - plugin_audit_properties=params.config['configurations']['ranger-hive-audit'], plugin_audit_attributes=params.config['configuration_attributes']['ranger-hive-audit'], - plugin_security_properties=params.config['configurations']['ranger-hive-security'], plugin_security_attributes=params.config['configuration_attributes']['ranger-hive-security'], - plugin_policymgr_ssl_properties=params.config['configurations']['ranger-hive-policymgr-ssl'], plugin_policymgr_ssl_attributes=params.config['configuration_attributes']['ranger-hive-policymgr-ssl'], - component_list=['hive-client', 'hive-metastore', 'hive-server2'], audit_db_is_enabled=params.xa_audit_db_is_enabled, - credential_file=params.credential_file, xa_audit_db_password=params.xa_audit_db_password, - ssl_truststore_password=params.ssl_truststore_password, ssl_keystore_password=params.ssl_keystore_password, - stack_version_override = stack_version, skip_if_rangeradmin_down= not params.retryAble) - else: - Logger.info('Ranger admin not installed')
http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/setup_ranger_hive_interactive.py ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/setup_ranger_hive_interactive.py b/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/setup_ranger_hive_interactive.py deleted file mode 100755 index 0b5d5db..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/setup_ranger_hive_interactive.py +++ /dev/null @@ -1,78 +0,0 @@ -#!/usr/bin/env python -""" -Licensed to the Apache Software Foundation (ASF) under one -or more contributor license agreements. See the NOTICE file -distributed with this work for additional information -regarding copyright ownership. The ASF licenses this file -to you under the Apache License, Version 2.0 (the -"License"); you may not use this file except in compliance -with the License. You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - -Unless required by applicable law or agreed to in writing, software -distributed under the License is distributed on an "AS IS" BASIS, -WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -See the License for the specific language governing permissions and -limitations under the License. - -""" -from resource_management.core.logger import Logger - -def setup_ranger_hive_interactive(upgrade_type = None): - import params - - if params.has_ranger_admin: - - stack_version = None - - if upgrade_type is not None: - stack_version = params.version - - if params.retryAble: - Logger.info("Hive2: Setup ranger: command retry enabled thus retrying if ranger admin is down !") - else: - Logger.info("Hive2: Setup ranger: command retry not enabled thus skipping if ranger admin is down !") - - if params.xml_configurations_supported and params.enable_ranger_hive and params.xa_audit_hdfs_is_enabled: - params.HdfsResource("/ranger/audit", - type="directory", - action="create_on_execute", - owner=params.hdfs_user, - group=params.hdfs_user, - mode=0755, - recursive_chmod=True - ) - params.HdfsResource("/ranger/audit/hive2", - type="directory", - action="create_on_execute", - owner=params.hive_user, - group=params.hive_user, - mode=0700, - recursive_chmod=True - ) - params.HdfsResource(None, action="execute") - - from resource_management.libraries.functions.setup_ranger_plugin_xml import setup_ranger_plugin - setup_ranger_plugin('hive-server2-hive2', 'hive', params.ranger_previous_jdbc_jar, - params.ranger_downloaded_custom_connector, params.ranger_driver_curl_source, - params.ranger_driver_curl_target, params.java64_home, - params.repo_name, params.hive_ranger_plugin_repo, - params.ranger_env, params.ranger_plugin_properties, - params.policy_user, params.policymgr_mgr_url, - params.enable_ranger_hive, conf_dict=params.hive_server_interactive_conf_dir, - component_user=params.hive_user, component_group=params.user_group, cache_service_list=['hive-server2-hive2'], - plugin_audit_properties=params.config['configurations']['ranger-hive-audit'], plugin_audit_attributes=params.config['configuration_attributes']['ranger-hive-audit'], - plugin_security_properties=params.config['configurations']['ranger-hive-security'], plugin_security_attributes=params.config['configuration_attributes']['ranger-hive-security'], - plugin_policymgr_ssl_properties=params.config['configurations']['ranger-hive-policymgr-ssl'], plugin_policymgr_ssl_attributes=params.config['configuration_attributes']['ranger-hive-policymgr-ssl'], - component_list=['hive-client', 'hive-metastore', 'hive-server2','hive-server2-hive2'], audit_db_is_enabled=False, - credential_file=params.credential_file, xa_audit_db_password=None, - ssl_truststore_password=params.ssl_truststore_password, ssl_keystore_password=params.ssl_keystore_password, - stack_version_override = stack_version, skip_if_rangeradmin_down= not params.retryAble, api_version='v2', - is_security_enabled = params.security_enabled, - is_stack_supports_ranger_kerberos = params.stack_supports_ranger_kerberos, - component_user_principal=params.hive_principal if params.security_enabled else None, - component_user_keytab=params.hive_server2_keytab if params.security_enabled else None) - - else: - Logger.info('Ranger admin not installed') http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/status_params.py ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/status_params.py b/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/status_params.py deleted file mode 100755 index a7b2e3f..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/status_params.py +++ /dev/null @@ -1,123 +0,0 @@ -#!/usr/bin/env python -""" -Licensed to the Apache Software Foundation (ASF) under one -or more contributor license agreements. See the NOTICE file -distributed with this work for additional information -regarding copyright ownership. The ASF licenses this file -to you under the Apache License, Version 2.0 (the -"License"); you may not use this file except in compliance -with the License. You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - -Unless required by applicable law or agreed to in writing, software -distributed under the License is distributed on an "AS IS" BASIS, -WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -See the License for the specific language governing permissions and -limitations under the License. - -""" - -from ambari_commons import OSCheck - -from resource_management.libraries.functions import conf_select -from resource_management.libraries.functions import stack_select -from resource_management.libraries.functions import format -from resource_management.libraries.functions import StackFeature -from resource_management.libraries.functions.stack_features import check_stack_feature -from resource_management.libraries.functions.version import format_stack_version -from resource_management.libraries.functions.default import default -from resource_management.libraries.functions import get_kinit_path -from resource_management.libraries.script.script import Script - - -# a map of the Ambari role to the component name -# for use with <stack-root>/current/<component> -SERVER_ROLE_DIRECTORY_MAP = { - 'HIVE_METASTORE' : 'hive-metastore', - 'HIVE_SERVER' : 'hive-server2', - 'WEBHCAT_SERVER' : 'hive-webhcat', - 'HIVE_CLIENT' : 'hive-client', - 'HCAT' : 'hive-client', - 'HIVE_SERVER_INTERACTIVE' : 'hive-server2-hive2' -} - - -# Either HIVE_METASTORE, HIVE_SERVER, WEBHCAT_SERVER, HIVE_CLIENT, HCAT, HIVE_SERVER_INTERACTIVE -role = default("/role", None) -component_directory = Script.get_component_from_role(SERVER_ROLE_DIRECTORY_MAP, "HIVE_CLIENT") -component_directory_interactive = Script.get_component_from_role(SERVER_ROLE_DIRECTORY_MAP, "HIVE_SERVER_INTERACTIVE") - -config = Script.get_config() - -stack_root = Script.get_stack_root() -stack_version_unformatted = config['hostLevelParams']['stack_version'] -stack_version_formatted_major = format_stack_version(stack_version_unformatted) - -if OSCheck.is_windows_family(): - hive_metastore_win_service_name = "metastore" - hive_client_win_service_name = "hwi" - hive_server_win_service_name = "hiveserver2" - webhcat_server_win_service_name = "templeton" -else: - hive_pid_dir = config['configurations']['hive-env']['hive_pid_dir'] - hive_pid = 'hive-server.pid' - hive_interactive_pid = 'hive-interactive.pid' - hive_metastore_pid = 'hive.pid' - - hcat_pid_dir = config['configurations']['hive-env']['hcat_pid_dir'] #hcat_pid_dir - webhcat_pid_file = format('{hcat_pid_dir}/webhcat.pid') - - process_name = 'mysqld' - if OSCheck.is_suse_family() or OSCheck.is_ubuntu_family(): - daemon_name = 'mysql' - else: - daemon_name = 'mysqld' - - # Security related/required params - hostname = config['hostname'] - security_enabled = config['configurations']['cluster-env']['security_enabled'] - kinit_path_local = get_kinit_path(default('/configurations/kerberos-env/executable_search_paths', None)) - tmp_dir = Script.get_tmp_dir() - hdfs_user = config['configurations']['hadoop-env']['hdfs_user'] - hive_user = config['configurations']['hive-env']['hive_user'] - webhcat_user = config['configurations']['hive-env']['webhcat_user'] - - # default configuration directories - hadoop_conf_dir = conf_select.get_hadoop_conf_dir() - hadoop_bin_dir = stack_select.get_hadoop_dir("bin") - hive_etc_dir_prefix = "/etc/hive" - hive_interactive_etc_dir_prefix = "/etc/hive2" - - hive_server_conf_dir = "/etc/hive/conf.server" - hive_server_interactive_conf_dir = "/etc/hive2/conf.server" - -# webhcat_conf_dir = format("{stack_root}/current/hive-webhcat/conf") -# hive_home_dir = format("{stack_root}/current/{component_directory}") -# hive_conf_dir = format("{stack_root}/current/{component_directory}/conf") -# hive_client_conf_dir = format("{stack_root}/current/{component_directory}/conf") - webhcat_conf_dir = '/etc/hive/conf' - hive_home_dir = '/usr/lib/hive' - hive_conf_dir = '/usr/lib/hive/conf' - hive_client_conf_dir = '/etc/hive/conf' - - if check_stack_feature(StackFeature.CONFIG_VERSIONING, stack_version_formatted_major): - hive_server_conf_dir = format("{stack_root}/current/{component_directory}/conf/conf.server") - hive_conf_dir = hive_server_conf_dir - - if check_stack_feature(StackFeature.HIVE_WEBHCAT_SPECIFIC_CONFIGS, stack_version_formatted_major): - # this is NOT a typo. Configs for hcatalog/webhcat point to a - # specific directory which is NOT called 'conf' - # FIXME ODPi: webhcat_conf_dir = format("{stack_root}/current/hive-webhcat/etc/webhcat") - webhcat_conf_dir = format("/etc/hive-webhcat/conf") - - # if stack version supports hive serve interactive - if check_stack_feature(StackFeature.HIVE_SERVER_INTERACTIVE, stack_version_formatted_major): - hive_server_interactive_conf_dir = format("{stack_root}/current/{component_directory_interactive}/conf/conf.server") - - hive_config_dir = hive_client_conf_dir - - if 'role' in config and config['role'] in ["HIVE_SERVER", "HIVE_METASTORE", "HIVE_SERVER_INTERACTIVE"]: - hive_config_dir = hive_server_conf_dir - -stack_name = default("/hostLevelParams/stack_name", None) http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/webhcat.py ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/webhcat.py b/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/webhcat.py deleted file mode 100755 index fe3f34a..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/webhcat.py +++ /dev/null @@ -1,145 +0,0 @@ -""" -Licensed to the Apache Software Foundation (ASF) under one -or more contributor license agreements. See the NOTICE file -distributed with this work for additional information -regarding copyright ownership. The ASF licenses this file -to you under the Apache License, Version 2.0 (the -"License"); you may not use this file except in compliance -with the License. You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - -Unless required by applicable law or agreed to in writing, software -distributed under the License is distributed on an "AS IS" BASIS, -WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -See the License for the specific language governing permissions and -limitations under the License. - -Ambari Agent - -""" -import sys -import os.path -from resource_management import * -from resource_management.core.resources.system import Execute -from resource_management.libraries.functions import StackFeature -from resource_management.libraries.functions.stack_features import check_stack_feature -from ambari_commons.os_family_impl import OsFamilyFuncImpl, OsFamilyImpl -from resource_management.libraries.functions.setup_atlas_hook import has_atlas_in_cluster, setup_atlas_hook -from ambari_commons import OSConst -from ambari_commons.constants import SERVICE - - -@OsFamilyFuncImpl(os_family=OSConst.WINSRV_FAMILY) -def webhcat(): - import params - XmlConfig("webhcat-site.xml", - conf_dir=params.hcat_config_dir, - configurations=params.config['configurations']['webhcat-site'] - ) - # Manually overriding service logon user & password set by the installation package - ServiceConfig(params.webhcat_server_win_service_name, - action="change_user", - username = params.hcat_user, - password = Script.get_password(params.hcat_user)) - - -@OsFamilyFuncImpl(os_family=OsFamilyImpl.DEFAULT) -def webhcat(): - import params - - Directory(params.templeton_pid_dir, - owner=params.webhcat_user, - mode=0755, - group=params.user_group, - create_parents = True) - - Directory(params.templeton_log_dir, - owner=params.webhcat_user, - mode=0755, - group=params.user_group, - create_parents = True) - - Directory(params.config_dir, - create_parents = True, - owner=params.webhcat_user, - group=params.user_group, - cd_access="a") - - if params.security_enabled: - kinit_if_needed = format("{kinit_path_local} -kt {hdfs_user_keytab} {hdfs_principal_name};") - else: - kinit_if_needed = "" - - if kinit_if_needed: - Execute(kinit_if_needed, - user=params.webhcat_user, - path='/bin' - ) - - # Replace _HOST with hostname in relevant principal-related properties - webhcat_site = params.config['configurations']['webhcat-site'].copy() - for prop_name in ['templeton.hive.properties', 'templeton.kerberos.principal']: - if prop_name in webhcat_site: - webhcat_site[prop_name] = webhcat_site[prop_name].replace("_HOST", params.hostname) - - XmlConfig("webhcat-site.xml", - conf_dir=params.config_dir, - configurations=webhcat_site, - configuration_attributes=params.config['configuration_attributes']['webhcat-site'], - owner=params.webhcat_user, - group=params.user_group, - ) - - # if we're in an upgrade of a secure cluster, make sure hive-site and yarn-site are created - if params.stack_version_formatted_major and check_stack_feature(StackFeature.CONFIG_VERSIONING, params.stack_version_formatted_major) and \ - params.version and params.stack_root: - XmlConfig("hive-site.xml", - conf_dir = format("{stack_root}/{version}/hive/conf"), - configurations = params.config['configurations']['hive-site'], - configuration_attributes = params.config['configuration_attributes']['hive-site'], - owner = params.hive_user, - group = params.user_group, - ) - - XmlConfig("yarn-site.xml", - conf_dir = format("{stack_root}/{version}/hadoop/conf"), - configurations = params.config['configurations']['yarn-site'], - configuration_attributes = params.config['configuration_attributes']['yarn-site'], - owner = params.yarn_user, - group = params.user_group, - ) - - - File(format("{config_dir}/webhcat-env.sh"), - owner=params.webhcat_user, - group=params.user_group, - content=InlineTemplate(params.webhcat_env_sh_template) - ) - - Directory(params.webhcat_conf_dir, - cd_access='a', - create_parents = True - ) - - log4j_webhcat_filename = 'webhcat-log4j.properties' - if (params.log4j_webhcat_props != None): - File(format("{config_dir}/{log4j_webhcat_filename}"), - mode=0644, - group=params.user_group, - owner=params.webhcat_user, - content=params.log4j_webhcat_props - ) - elif (os.path.exists("{config_dir}/{log4j_webhcat_filename}.template")): - File(format("{config_dir}/{log4j_webhcat_filename}"), - mode=0644, - group=params.user_group, - owner=params.webhcat_user, - content=StaticFile(format("{config_dir}/{log4j_webhcat_filename}.template")) - ) - - # Generate atlas-application.properties.xml file - if has_atlas_in_cluster(): - # WebHCat uses a different config dir than the rest of the daemons in Hive. - atlas_hook_filepath = os.path.join(params.config_dir, params.atlas_hook_filename) - setup_atlas_hook(SERVICE.HIVE, params.hive_atlas_application_properties, atlas_hook_filepath, params.hive_user, params.user_group) http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/webhcat_server.py ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/webhcat_server.py b/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/webhcat_server.py deleted file mode 100755 index 34687c4..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/webhcat_server.py +++ /dev/null @@ -1,164 +0,0 @@ -""" -Licensed to the Apache Software Foundation (ASF) under one -or more contributor license agreements. See the NOTICE file -distributed with this work for additional information -regarding copyright ownership. The ASF licenses this file -to you under the Apache License, Version 2.0 (the -"License"); you may not use this file except in compliance -with the License. You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - -Unless required by applicable law or agreed to in writing, software -distributed under the License is distributed on an "AS IS" BASIS, -WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -See the License for the specific language governing permissions and -limitations under the License. - -Ambari Agent - -""" -from resource_management import * -from resource_management.libraries.functions import conf_select -from resource_management.libraries.functions import stack_select -from resource_management.libraries.functions import StackFeature -from resource_management.libraries.functions.stack_features import check_stack_feature -from resource_management.libraries.functions.security_commons import build_expectations, \ - cached_kinit_executor, get_params_from_filesystem, validate_security_config_properties, \ - FILE_TYPE_XML -from webhcat import webhcat -from webhcat_service import webhcat_service -from ambari_commons import OSConst -from ambari_commons.os_family_impl import OsFamilyImpl - - -class WebHCatServer(Script): - def install(self, env): - import params - self.install_packages(env) - - def start(self, env, upgrade_type=None): - import params - env.set_params(params) - self.configure(env) # FOR SECURITY - webhcat_service(action='start', upgrade_type=upgrade_type) - - def stop(self, env, upgrade_type=None): - import params - env.set_params(params) - webhcat_service(action='stop') - - def configure(self, env): - import params - env.set_params(params) - webhcat() - - -@OsFamilyImpl(os_family=OSConst.WINSRV_FAMILY) -class WebHCatServerWindows(WebHCatServer): - def status(self, env): - import status_params - env.set_params(status_params) - check_windows_service_status(status_params.webhcat_server_win_service_name) - - -@OsFamilyImpl(os_family=OsFamilyImpl.DEFAULT) -class WebHCatServerDefault(WebHCatServer): - def get_component_name(self): - return "hive-webhcat" - - def status(self, env): - import status_params - env.set_params(status_params) - check_process_status(status_params.webhcat_pid_file) - - def pre_upgrade_restart(self, env, upgrade_type=None): - Logger.info("Executing WebHCat Stack Upgrade pre-restart") - import params - env.set_params(params) - - if params.version and check_stack_feature(StackFeature.ROLLING_UPGRADE, params.version): - # webhcat has no conf, but uses hadoop home, so verify that regular hadoop conf is set - conf_select.select(params.stack_name, "hive-hcatalog", params.version) - conf_select.select(params.stack_name, "hadoop", params.version) - stack_select.select("hive-webhcat", params.version) - - def security_status(self, env): - import status_params - env.set_params(status_params) - - if status_params.security_enabled: - expectations ={} - expectations.update( - build_expectations( - 'webhcat-site', - { - "templeton.kerberos.secret": "secret" - }, - [ - "templeton.kerberos.keytab", - "templeton.kerberos.principal" - ], - [ - "templeton.kerberos.keytab" - ] - ) - ) - expectations.update( - build_expectations( - 'hive-site', - { - "hive.server2.authentication": "KERBEROS", - "hive.metastore.sasl.enabled": "true", - "hive.security.authorization.enabled": "true" - }, - None, - None - ) - ) - - security_params = {} - security_params.update(get_params_from_filesystem(status_params.hive_conf_dir, - {'hive-site.xml': FILE_TYPE_XML})) - security_params.update(get_params_from_filesystem(status_params.webhcat_conf_dir, - {'webhcat-site.xml': FILE_TYPE_XML})) - result_issues = validate_security_config_properties(security_params, expectations) - if not result_issues: # If all validations passed successfully - try: - # Double check the dict before calling execute - if 'webhcat-site' not in security_params \ - or 'templeton.kerberos.keytab' not in security_params['webhcat-site'] \ - or 'templeton.kerberos.principal' not in security_params['webhcat-site']: - self.put_structured_out({"securityState": "UNSECURED"}) - self.put_structured_out({"securityIssuesFound": "Keytab file or principal are not set property."}) - return - - cached_kinit_executor(status_params.kinit_path_local, - status_params.webhcat_user, - security_params['webhcat-site']['templeton.kerberos.keytab'], - security_params['webhcat-site']['templeton.kerberos.principal'], - status_params.hostname, - status_params.tmp_dir) - self.put_structured_out({"securityState": "SECURED_KERBEROS"}) - except Exception as e: - self.put_structured_out({"securityState": "ERROR"}) - self.put_structured_out({"securityStateErrorInfo": str(e)}) - else: - issues = [] - for cf in result_issues: - issues.append("Configuration file %s did not pass the validation. Reason: %s" % (cf, result_issues[cf])) - self.put_structured_out({"securityIssuesFound": ". ".join(issues)}) - self.put_structured_out({"securityState": "UNSECURED"}) - else: - self.put_structured_out({"securityState": "UNSECURED"}) - - def get_log_folder(self): - import params - return params.hcat_log_dir - - def get_user(self): - import params - return params.webhcat_user - -if __name__ == "__main__": - WebHCatServer().execute() http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/webhcat_service.py ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/webhcat_service.py b/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/webhcat_service.py deleted file mode 100755 index c24db4c..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/webhcat_service.py +++ /dev/null @@ -1,96 +0,0 @@ -""" -Licensed to the Apache Software Foundation (ASF) under one -or more contributor license agreements. See the NOTICE file -distributed with this work for additional information -regarding copyright ownership. The ASF licenses this file -to you under the Apache License, Version 2.0 (the -"License"); you may not use this file except in compliance -with the License. You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - -Unless required by applicable law or agreed to in writing, software -distributed under the License is distributed on an "AS IS" BASIS, -WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -See the License for the specific language governing permissions and -limitations under the License. - -Ambari Agent - -""" -from resource_management import * -from ambari_commons.os_family_impl import OsFamilyFuncImpl, OsFamilyImpl -from ambari_commons import OSConst -from resource_management.core.shell import as_user -from resource_management.core.logger import Logger -import traceback - - -@OsFamilyFuncImpl(os_family=OSConst.WINSRV_FAMILY) -def webhcat_service(action='start', rolling_restart=False): - import params - if action == 'start' or action == 'stop': - Service(params.webhcat_server_win_service_name, action=action) - - -@OsFamilyFuncImpl(os_family=OsFamilyImpl.DEFAULT) -def webhcat_service(action='start', upgrade_type=None): - import params - - environ = { - 'HADOOP_HOME': params.hadoop_home - } - - cmd = format('{webhcat_bin_dir}/webhcat_server.sh') - - if action == 'start': - if upgrade_type is not None and params.version and params.stack_root: - environ['HADOOP_HOME'] = format("{stack_root}/{version}/hadoop") - - daemon_cmd = format('cd {hcat_pid_dir} ; {cmd} start') - no_op_test = as_user(format('ls {webhcat_pid_file} >/dev/null 2>&1 && ps -p `cat {webhcat_pid_file}` >/dev/null 2>&1'), user=params.webhcat_user) - try: - Execute(daemon_cmd, - user=params.webhcat_user, - not_if=no_op_test, - environment = environ) - except: - show_logs(params.hcat_log_dir, params.webhcat_user) - raise - elif action == 'stop': - try: - graceful_stop(cmd, environ) - except Fail: - show_logs(params.hcat_log_dir, params.webhcat_user) - Logger.info(traceback.format_exc()) - - pid_expression = "`" + as_user(format("cat {webhcat_pid_file}"), user=params.webhcat_user) + "`" - process_id_exists_command = format("ls {webhcat_pid_file} >/dev/null 2>&1 && ps -p {pid_expression} >/dev/null 2>&1") - daemon_hard_kill_cmd = format("{sudo} kill -9 {pid_expression}") - wait_time = 10 - Execute(daemon_hard_kill_cmd, - not_if = format("! ({process_id_exists_command}) || ( sleep {wait_time} && ! ({process_id_exists_command}) )"), - ignore_failures = True - ) - - try: - # check if stopped the process, else fail the task - Execute(format("! ({process_id_exists_command})"), - tries=20, - try_sleep=3, - ) - except: - show_logs(params.hcat_log_dir, params.webhcat_user) - raise - - File(params.webhcat_pid_file, - action="delete", - ) - -def graceful_stop(cmd, environ): - import params - daemon_cmd = format('{cmd} stop') - - Execute(daemon_cmd, - user = params.webhcat_user, - environment = environ) http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/webhcat_service_check.py ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/webhcat_service_check.py b/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/webhcat_service_check.py deleted file mode 100755 index 8e80d48..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/scripts/webhcat_service_check.py +++ /dev/null @@ -1,128 +0,0 @@ -#!/usr/bin/env python -""" -Licensed to the Apache Software Foundation (ASF) under one -or more contributor license agreements. See the NOTICE file -distributed with this work for additional information -regarding copyright ownership. The ASF licenses this file -to you under the Apache License, Version 2.0 (the -"License"); you may not use this file except in compliance -with the License. You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - -Unless required by applicable law or agreed to in writing, software -distributed under the License is distributed on an "AS IS" BASIS, -WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -See the License for the specific language governing permissions and -limitations under the License. - -""" -import urllib2 - -from resource_management import * -from ambari_commons.os_family_impl import OsFamilyFuncImpl, OsFamilyImpl -from ambari_commons import OSConst -import time - -@OsFamilyFuncImpl(os_family=OSConst.WINSRV_FAMILY) -def webhcat_service_check(): - Logger.info("Webhcat smoke test - service status") - - import params - # AMBARI-11633 [WinTP2] Webhcat service check fails - # Hive doesn't pass the environment variables correctly to child processes, which fails the smoke test. - # Reducing the amount of URLs checked to the minimum required. - #smoke_cmd = os.path.join(params.stack_root,"Run-SmokeTests.cmd") - #service = "WEBHCAT" - #Execute(format("cmd /C {smoke_cmd} {service}"), user=params.hcat_user, logoutput=True) - - url_tests = [ - "status", - #These are the failing ones: - #"ddl/database?user.name=hadoop", - #"ddl/database/default/table?user.name=hadoop" - ] - - - import socket - - url_host = socket.getfqdn() - url_port = params.config["configurations"]["webhcat-site"]["templeton.port"] - - for url_test in url_tests: - url_request = "http://{0}:{1}/templeton/v1/{2}".format(url_host, url_port, url_test) - url_response = None - - try: - # execute the query for the JSON that includes WebHCat status - url_response = urllib2.urlopen(url_request, timeout=30) - - status = url_response.getcode() - response = url_response.read() - - if status != 200: - Logger.warning("Webhcat service check status: {0}".format(status)) - Logger.info("Webhcat service check response: {0}".format(response)) - except urllib2.HTTPError as he: - raise Fail("Webhcat check {0} failed: {1}".format(url_request, he.msg)) - finally: - if url_response is not None: - try: - url_response.close() - except: - pass - - -@OsFamilyFuncImpl(os_family=OsFamilyImpl.DEFAULT) -def webhcat_service_check(): - import params - File(format("{tmp_dir}/templetonSmoke.sh"), - content= StaticFile('templetonSmoke.sh'), - mode=0755 - ) - - if params.security_enabled: - smokeuser_keytab=params.smoke_user_keytab - smoke_user_principal=params.smokeuser_principal - else: - smokeuser_keytab= "no_keytab" - smoke_user_principal="no_principal" - - unique_name = format("{smokeuser}.{timestamp}", timestamp = time.time()) - templeton_test_script = format("idtest.{unique_name}.pig") - templeton_test_input = format("/tmp/idtest.{unique_name}.in") - templeton_test_output = format("/tmp/idtest.{unique_name}.out") - - File(format("{tmp_dir}/{templeton_test_script}"), - content = Template("templeton_smoke.pig.j2", templeton_test_input=templeton_test_input, templeton_test_output=templeton_test_output), - owner=params.hdfs_user - ) - - params.HdfsResource(format("/tmp/{templeton_test_script}"), - action = "create_on_execute", - type = "file", - source = format("{tmp_dir}/{templeton_test_script}"), - owner = params.smokeuser - ) - - params.HdfsResource(templeton_test_input, - action = "create_on_execute", - type = "file", - source = "/etc/passwd", - owner = params.smokeuser - ) - - params.HdfsResource(None, action = "execute") - - cmd = format("{tmp_dir}/templetonSmoke.sh {webhcat_server_host[0]} {smokeuser} {templeton_port} {templeton_test_script} {smokeuser_keytab}" - " {security_param} {kinit_path_local} {smoke_user_principal}" - " {tmp_dir}") - - Execute(cmd, - tries=3, - try_sleep=5, - path='/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin', - logoutput=True) - - - http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/hadoop-metrics2-hivemetastore.properties.j2 ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/hadoop-metrics2-hivemetastore.properties.j2 b/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/hadoop-metrics2-hivemetastore.properties.j2 deleted file mode 100755 index e4d88bc..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/hadoop-metrics2-hivemetastore.properties.j2 +++ /dev/null @@ -1,54 +0,0 @@ -{# -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -#} - -# Licensed to the Apache Software Foundation (ASF) under one or more -# contributor license agreements. See the NOTICE file distributed with -# this work for additional information regarding copyright ownership. -# The ASF licenses this file to You under the Apache License, Version 2.0 -# (the "License"); you may not use this file except in compliance with -# the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -# syntax: [prefix].[source|sink|jmx].[instance].[options] -# See package.html for org.apache.hadoop.metrics2 for details - -{% if has_metric_collector %} - - *.period={{metrics_collection_period}} - *.sink.timeline.plugin.urls=file:///usr/lib/ambari-metrics-hadoop-sink/ambari-metrics-hadoop-sink.jar - *.sink.timeline.class=org.apache.hadoop.metrics2.sink.timeline.HadoopTimelineMetricsSink - *.sink.timeline.period={{metrics_collection_period}} - *.sink.timeline.sendInterval={{metrics_report_interval}}000 - *.sink.timeline.slave.host.name = {{hostname}} - - # HTTPS properties - *.sink.timeline.truststore.path = {{metric_truststore_path}} - *.sink.timeline.truststore.type = {{metric_truststore_type}} - *.sink.timeline.truststore.password = {{metric_truststore_password}} - - hivemetastore.sink.timeline.collector={{metric_collector_protocol}}://{{metric_collector_host}}:{{metric_collector_port}} - - -{% endif %} http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/hadoop-metrics2-hiveserver2.properties.j2 ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/hadoop-metrics2-hiveserver2.properties.j2 b/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/hadoop-metrics2-hiveserver2.properties.j2 deleted file mode 100755 index b5c4891..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/hadoop-metrics2-hiveserver2.properties.j2 +++ /dev/null @@ -1,54 +0,0 @@ -{# -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -#} - -# Licensed to the Apache Software Foundation (ASF) under one or more -# contributor license agreements. See the NOTICE file distributed with -# this work for additional information regarding copyright ownership. -# The ASF licenses this file to You under the Apache License, Version 2.0 -# (the "License"); you may not use this file except in compliance with -# the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -# syntax: [prefix].[source|sink|jmx].[instance].[options] -# See package.html for org.apache.hadoop.metrics2 for details - -{% if has_metric_collector %} - - *.period={{metrics_collection_period}} - *.sink.timeline.plugin.urls=file:///usr/lib/ambari-metrics-hadoop-sink/ambari-metrics-hadoop-sink.jar - *.sink.timeline.class=org.apache.hadoop.metrics2.sink.timeline.HadoopTimelineMetricsSink - *.sink.timeline.period={{metrics_collection_period}} - *.sink.timeline.sendInterval={{metrics_report_interval}}000 - *.sink.timeline.slave.host.name = {{hostname}} - - # HTTPS properties - *.sink.timeline.truststore.path = {{metric_truststore_path}} - *.sink.timeline.truststore.type = {{metric_truststore_type}} - *.sink.timeline.truststore.password = {{metric_truststore_password}} - - hiveserver2.sink.timeline.collector={{metric_collector_protocol}}://{{metric_collector_host}}:{{metric_collector_port}} - - -{% endif %} http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/hadoop-metrics2-llapdaemon.j2 ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/hadoop-metrics2-llapdaemon.j2 b/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/hadoop-metrics2-llapdaemon.j2 deleted file mode 100755 index 1d75ccf..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/hadoop-metrics2-llapdaemon.j2 +++ /dev/null @@ -1,52 +0,0 @@ -{# -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -#} - -# Licensed to the Apache Software Foundation (ASF) under one or more -# contributor license agreements. See the NOTICE file distributed with -# this work for additional information regarding copyright ownership. -# The ASF licenses this file to You under the Apache License, Version 2.0 -# (the "License"); you may not use this file except in compliance with -# the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -# syntax: [prefix].[source|sink|jmx].[instance].[options] -# See package.html for org.apache.hadoop.metrics2 for details - -{% if has_metric_collector %} - - *.period={{metrics_collection_period}} - *.sink.timeline.plugin.urls=file:///usr/lib/ambari-metrics-hadoop-sink/ambari-metrics-hadoop-sink.jar - *.sink.timeline.class=org.apache.hadoop.metrics2.sink.timeline.HadoopTimelineMetricsSink - *.sink.timeline.period={{metrics_collection_period}} - *.sink.timeline.sendInterval={{metrics_report_interval}}000 - - # HTTPS properties - *.sink.timeline.truststore.path = {{metric_truststore_path}} - *.sink.timeline.truststore.type = {{metric_truststore_type}} - *.sink.timeline.truststore.password = {{metric_truststore_password}} - - llapdaemon.sink.timeline.collector={{metric_collector_protocol}}://{{metric_collector_host}}:{{metric_collector_port}} - -{% endif %} \ No newline at end of file http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/hadoop-metrics2-llaptaskscheduler.j2 ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/hadoop-metrics2-llaptaskscheduler.j2 b/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/hadoop-metrics2-llaptaskscheduler.j2 deleted file mode 100755 index 5ab787c..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/hadoop-metrics2-llaptaskscheduler.j2 +++ /dev/null @@ -1,52 +0,0 @@ -{# -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -#} - -# Licensed to the Apache Software Foundation (ASF) under one or more -# contributor license agreements. See the NOTICE file distributed with -# this work for additional information regarding copyright ownership. -# The ASF licenses this file to You under the Apache License, Version 2.0 -# (the "License"); you may not use this file except in compliance with -# the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -# syntax: [prefix].[source|sink|jmx].[instance].[options] -# See package.html for org.apache.hadoop.metrics2 for details - -{% if has_metric_collector %} - - *.period={{metrics_collection_period}} - *.sink.timeline.plugin.urls=file:///usr/lib/ambari-metrics-hadoop-sink/ambari-metrics-hadoop-sink.jar - *.sink.timeline.class=org.apache.hadoop.metrics2.sink.timeline.HadoopTimelineMetricsSink - *.sink.timeline.period={{metrics_collection_period}} - *.sink.timeline.sendInterval={{metrics_report_interval}}000 - - # HTTPS properties - *.sink.timeline.truststore.path = {{metric_truststore_path}} - *.sink.timeline.truststore.type = {{metric_truststore_type}} - *.sink.timeline.truststore.password = {{metric_truststore_password}} - - llaptaskscheduler.sink.timeline.collector={{metric_collector_protocol}}://{{metric_collector_host}}:{{metric_collector_port}} - -{% endif %} \ No newline at end of file http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/hive.conf.j2 ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/hive.conf.j2 b/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/hive.conf.j2 deleted file mode 100755 index 5af53d0..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/hive.conf.j2 +++ /dev/null @@ -1,35 +0,0 @@ -{# -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -#} - -# Licensed to the Apache Software Foundation (ASF) under one or more -# contributor license agreements. See the NOTICE file distributed with -# this work for additional information regarding copyright ownership. -# The ASF licenses this file to You under the Apache License, Version 2.0 -# (the "License"); you may not use this file except in compliance with -# the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -{{hive_user}} - nofile {{hive_user_nofile_limit}} -{{hive_user}} - nproc {{hive_user_nproc_limit}} http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/startHiveserver2.sh.j2 ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/startHiveserver2.sh.j2 b/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/startHiveserver2.sh.j2 deleted file mode 100755 index 70b418c..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/startHiveserver2.sh.j2 +++ /dev/null @@ -1,24 +0,0 @@ -# -# -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, -# software distributed under the License is distributed on an -# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY -# KIND, either express or implied. See the License for the -# specific language governing permissions and limitations -# under the License. -# -# - -HIVE_SERVER2_OPTS=" -hiveconf hive.log.file=hiveserver2.log -hiveconf hive.log.dir=$5" -HIVE_CONF_DIR=$4 {{hive_bin}}/hiveserver2 -hiveconf hive.metastore.uris=" " ${HIVE_SERVER2_OPTS} > $1 2> $2 & -echo $!|cat>$3 http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/startHiveserver2Interactive.sh.j2 ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/startHiveserver2Interactive.sh.j2 b/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/startHiveserver2Interactive.sh.j2 deleted file mode 100755 index 6062a7e..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/startHiveserver2Interactive.sh.j2 +++ /dev/null @@ -1,24 +0,0 @@ -# -# -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, -# software distributed under the License is distributed on an -# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY -# KIND, either express or implied. See the License for the -# specific language governing permissions and limitations -# under the License. -# -# - -HIVE_SERVER2_INTERACTIVE_OPTS=" -hiveconf hive.log.file=hiveserver2Interactive.log -hiveconf hive.log.dir=$5" -HIVE_INTERACTIVE_CONF_DIR=$4 {{hive_interactive_bin}}/hiveserver2 -hiveconf hive.metastore.uris=" " ${HIVE_SERVER2_INTERACTIVE_OPTS} > $1 2> $2 & -echo $!|cat>$3 \ No newline at end of file http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/templeton_smoke.pig.j2 ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/templeton_smoke.pig.j2 b/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/templeton_smoke.pig.j2 deleted file mode 100755 index 3153e81..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/package/templates/templeton_smoke.pig.j2 +++ /dev/null @@ -1,24 +0,0 @@ -# -# -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, -# software distributed under the License is distributed on an -# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY -# KIND, either express or implied. See the License for the -# specific language governing permissions and limitations -# under the License. -# -# - -A = load '{{templeton_test_input}}' using PigStorage(':'); -B = foreach A generate \$0 as id; -store B into '{{templeton_test_output}}'; \ No newline at end of file http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/services/KERBEROS/metainfo.xml ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/services/KERBEROS/metainfo.xml b/bigtop-packages/src/common/ambari/ODPi/1.0/services/KERBEROS/metainfo.xml deleted file mode 100755 index 25cfcc6..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/services/KERBEROS/metainfo.xml +++ /dev/null @@ -1,26 +0,0 @@ -<?xml version="1.0"?> -<!-- - Licensed to the Apache Software Foundation (ASF) under one or more - contributor license agreements. See the NOTICE file distributed with - this work for additional information regarding copyright ownership. - The ASF licenses this file to You under the Apache License, Version 2.0 - (the "License"); you may not use this file except in compliance with - the License. You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - - Unless required by applicable law or agreed to in writing, software - distributed under the License is distributed on an "AS IS" BASIS, - WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - See the License for the specific language governing permissions and - limitations under the License. ---> -<metainfo> - <schemaVersion>2.0</schemaVersion> - <services> - <service> - <name>KERBEROS</name> - <extends>common-services/KERBEROS/1.10.3-10</extends> - </service> - </services> -</metainfo>
