http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/scripts/params.py ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/scripts/params.py b/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/scripts/params.py deleted file mode 100755 index fc2c61f..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/scripts/params.py +++ /dev/null @@ -1,318 +0,0 @@ -""" -Licensed to the Apache Software Foundation (ASF) under one -or more contributor license agreements. See the NOTICE file -distributed with this work for additional information -regarding copyright ownership. The ASF licenses this file -to you under the Apache License, Version 2.0 (the -"License"); you may not use this file except in compliance -with the License. You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - -Unless required by applicable law or agreed to in writing, software -distributed under the License is distributed on an "AS IS" BASIS, -WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -See the License for the specific language governing permissions and -limitations under the License. - -""" - -import os - -from resource_management.libraries.functions import conf_select -from resource_management.libraries.functions import stack_select -from resource_management.libraries.functions import default -from resource_management.libraries.functions import format_jvm_option -from resource_management.libraries.functions import format -from resource_management.libraries.functions.version import format_stack_version, compare_versions -from ambari_commons.os_check import OSCheck -from resource_management.libraries.script.script import Script -from resource_management.libraries.functions import get_kinit_path -from resource_management.libraries.functions.get_not_managed_resources import get_not_managed_resources -from resource_management.libraries.resources.hdfs_resource import HdfsResource - -config = Script.get_config() - -host_sys_prepped = default("/hostLevelParams/host_sys_prepped", False) - -stack_version_unformatted = config['hostLevelParams']['stack_version'] -stack_version_formatted = format_stack_version(stack_version_unformatted) - -dfs_type = default("/commandParams/dfs_type", "") -hadoop_conf_dir = "/etc/hadoop/conf" - -component_list = default("/localComponents", []) - -hdfs_tmp_dir = config['configurations']['hadoop-env']['hdfs_tmp_dir'] - -# hadoop default params -mapreduce_libs_path = "/usr/lib/hadoop-mapreduce/*" - -hadoop_libexec_dir = stack_select.get_hadoop_dir("libexec") -hadoop_lib_home = stack_select.get_hadoop_dir("lib") -hadoop_bin = stack_select.get_hadoop_dir("sbin") -hadoop_home = '/usr' -create_lib_snappy_symlinks = True - -# HDP 2.2+ params -if Script.is_stack_greater_or_equal("2.2"): - mapreduce_libs_path = "/usr/hdp/current/hadoop-mapreduce-client/*" - hadoop_home = stack_select.get_hadoop_dir("home") - create_lib_snappy_symlinks = False - -current_service = config['serviceName'] - -#security params -security_enabled = config['configurations']['cluster-env']['security_enabled'] - -#users and groups -has_hadoop_env = 'hadoop-env' in config['configurations'] -mapred_user = config['configurations']['mapred-env']['mapred_user'] -hdfs_user = config['configurations']['hadoop-env']['hdfs_user'] -yarn_user = config['configurations']['yarn-env']['yarn_user'] - -user_group = config['configurations']['cluster-env']['user_group'] - -#hosts -hostname = config["hostname"] -ambari_server_hostname = config['clusterHostInfo']['ambari_server_host'][0] -rm_host = default("/clusterHostInfo/rm_host", []) -slave_hosts = default("/clusterHostInfo/slave_hosts", []) -oozie_servers = default("/clusterHostInfo/oozie_server", []) -hcat_server_hosts = default("/clusterHostInfo/webhcat_server_host", []) -hive_server_host = default("/clusterHostInfo/hive_server_host", []) -hbase_master_hosts = default("/clusterHostInfo/hbase_master_hosts", []) -hs_host = default("/clusterHostInfo/hs_host", []) -jtnode_host = default("/clusterHostInfo/jtnode_host", []) -namenode_host = default("/clusterHostInfo/namenode_host", []) -zk_hosts = default("/clusterHostInfo/zookeeper_hosts", []) -ganglia_server_hosts = default("/clusterHostInfo/ganglia_server_host", []) -ams_collector_hosts = default("/clusterHostInfo/metrics_collector_hosts", []) - -has_namenode = not len(namenode_host) == 0 -has_resourcemanager = not len(rm_host) == 0 -has_slaves = not len(slave_hosts) == 0 -has_oozie_server = not len(oozie_servers) == 0 -has_hcat_server_host = not len(hcat_server_hosts) == 0 -has_hive_server_host = not len(hive_server_host) == 0 -has_hbase_masters = not len(hbase_master_hosts) == 0 -has_zk_host = not len(zk_hosts) == 0 -has_ganglia_server = not len(ganglia_server_hosts) == 0 -has_metric_collector = not len(ams_collector_hosts) == 0 - -is_namenode_master = hostname in namenode_host -is_jtnode_master = hostname in jtnode_host -is_rmnode_master = hostname in rm_host -is_hsnode_master = hostname in hs_host -is_hbase_master = hostname in hbase_master_hosts -is_slave = hostname in slave_hosts -if has_ganglia_server: - ganglia_server_host = ganglia_server_hosts[0] -if has_metric_collector: - if 'cluster-env' in config['configurations'] and \ - 'metrics_collector_vip_host' in config['configurations']['cluster-env']: - metric_collector_host = config['configurations']['cluster-env']['metrics_collector_vip_host'] - else: - metric_collector_host = ams_collector_hosts[0] - if 'cluster-env' in config['configurations'] and \ - 'metrics_collector_vip_port' in config['configurations']['cluster-env']: - metric_collector_port = config['configurations']['cluster-env']['metrics_collector_vip_port'] - else: - metric_collector_web_address = default("/configurations/ams-site/timeline.metrics.service.webapp.address", "localhost:6188") - if metric_collector_web_address.find(':') != -1: - metric_collector_port = metric_collector_web_address.split(':')[1] - else: - metric_collector_port = '6188' - if default("/configurations/ams-site/timeline.metrics.service.http.policy", "HTTP_ONLY") == "HTTPS_ONLY": - metric_collector_protocol = 'https' - else: - metric_collector_protocol = 'http' - metric_truststore_path= default("/configurations/ams-ssl-client/ssl.client.truststore.location", "") - metric_truststore_type= default("/configurations/ams-ssl-client/ssl.client.truststore.type", "") - metric_truststore_password= default("/configurations/ams-ssl-client/ssl.client.truststore.password", "") - - pass -metrics_report_interval = default("/configurations/ams-site/timeline.metrics.sink.report.interval", 60) -metrics_collection_period = default("/configurations/ams-site/timeline.metrics.sink.collection.period", 10) - -#hadoop params - -if has_namenode or dfs_type == 'HCFS': - hadoop_tmp_dir = format("/tmp/hadoop-{hdfs_user}") - hadoop_conf_dir = conf_select.get_hadoop_conf_dir(force_latest_on_upgrade=True) - task_log4j_properties_location = os.path.join(hadoop_conf_dir, "task-log4j.properties") - -hadoop_pid_dir_prefix = config['configurations']['hadoop-env']['hadoop_pid_dir_prefix'] -hdfs_log_dir_prefix = config['configurations']['hadoop-env']['hdfs_log_dir_prefix'] -hbase_tmp_dir = "/tmp/hbase-hbase" -#db params -server_db_name = config['hostLevelParams']['db_name'] -db_driver_filename = config['hostLevelParams']['db_driver_filename'] -oracle_driver_url = config['hostLevelParams']['oracle_jdbc_url'] -mysql_driver_url = config['hostLevelParams']['mysql_jdbc_url'] -ambari_server_resources = config['hostLevelParams']['jdk_location'] -oracle_driver_symlink_url = format("{ambari_server_resources}oracle-jdbc-driver.jar") -mysql_driver_symlink_url = format("{ambari_server_resources}mysql-jdbc-driver.jar") - -ambari_db_rca_url = config['hostLevelParams']['ambari_db_rca_url'][0] -ambari_db_rca_driver = config['hostLevelParams']['ambari_db_rca_driver'][0] -ambari_db_rca_username = config['hostLevelParams']['ambari_db_rca_username'][0] -ambari_db_rca_password = config['hostLevelParams']['ambari_db_rca_password'][0] - -if has_namenode and 'rca_enabled' in config['configurations']['hadoop-env']: - rca_enabled = config['configurations']['hadoop-env']['rca_enabled'] -else: - rca_enabled = False -rca_disabled_prefix = "###" -if rca_enabled == True: - rca_prefix = "" -else: - rca_prefix = rca_disabled_prefix - -#hadoop-env.sh -java_home = config['hostLevelParams']['java_home'] - -jsvc_path = "/usr/lib/bigtop-utils" - -hadoop_heapsize = config['configurations']['hadoop-env']['hadoop_heapsize'] -namenode_heapsize = config['configurations']['hadoop-env']['namenode_heapsize'] -namenode_opt_newsize = config['configurations']['hadoop-env']['namenode_opt_newsize'] -namenode_opt_maxnewsize = config['configurations']['hadoop-env']['namenode_opt_maxnewsize'] -namenode_opt_permsize = format_jvm_option("/configurations/hadoop-env/namenode_opt_permsize","128m") -namenode_opt_maxpermsize = format_jvm_option("/configurations/hadoop-env/namenode_opt_maxpermsize","256m") - -jtnode_opt_newsize = "200m" -jtnode_opt_maxnewsize = "200m" -jtnode_heapsize = "1024m" -ttnode_heapsize = "1024m" - -dtnode_heapsize = config['configurations']['hadoop-env']['dtnode_heapsize'] -mapred_pid_dir_prefix = default("/configurations/mapred-env/mapred_pid_dir_prefix","/var/run/hadoop-mapreduce") -mapred_log_dir_prefix = default("/configurations/mapred-env/mapred_log_dir_prefix","/var/log/hadoop-mapreduce") - -#log4j.properties - -yarn_log_dir_prefix = default("/configurations/yarn-env/yarn_log_dir_prefix","/var/log/hadoop-yarn") - -dfs_hosts = default('/configurations/hdfs-site/dfs.hosts', None) - -#log4j.properties -if (('hdfs-log4j' in config['configurations']) and ('content' in config['configurations']['hdfs-log4j'])): - log4j_props = config['configurations']['hdfs-log4j']['content'] - if (('yarn-log4j' in config['configurations']) and ('content' in config['configurations']['yarn-log4j'])): - log4j_props += config['configurations']['yarn-log4j']['content'] -else: - log4j_props = None - -refresh_topology = False -command_params = config["commandParams"] if "commandParams" in config else None -if command_params is not None: - refresh_topology = bool(command_params["refresh_topology"]) if "refresh_topology" in command_params else False - -ambari_libs_dir = "/var/lib/ambari-agent/lib" -is_webhdfs_enabled = config['configurations']['hdfs-site']['dfs.webhdfs.enabled'] -default_fs = config['configurations']['core-site']['fs.defaultFS'] - -#host info -all_hosts = default("/clusterHostInfo/all_hosts", []) -all_racks = default("/clusterHostInfo/all_racks", []) -all_ipv4_ips = default("/clusterHostInfo/all_ipv4_ips", []) -slave_hosts = default("/clusterHostInfo/slave_hosts", []) - -#topology files -net_topology_script_file_path = "/etc/hadoop/conf/topology_script.py" -net_topology_script_dir = os.path.dirname(net_topology_script_file_path) -net_topology_mapping_data_file_name = 'topology_mappings.data' -net_topology_mapping_data_file_path = os.path.join(net_topology_script_dir, net_topology_mapping_data_file_name) - -#Added logic to create /tmp and /user directory for HCFS stack. -has_core_site = 'core-site' in config['configurations'] -hdfs_user_keytab = config['configurations']['hadoop-env']['hdfs_user_keytab'] -kinit_path_local = get_kinit_path() -stack_version_unformatted = config['hostLevelParams']['stack_version'] -stack_version_formatted = format_stack_version(stack_version_unformatted) -hadoop_bin_dir = stack_select.get_hadoop_dir("bin") -hdfs_principal_name = default('/configurations/hadoop-env/hdfs_principal_name', None) -hdfs_site = config['configurations']['hdfs-site'] -default_fs = config['configurations']['core-site']['fs.defaultFS'] -smoke_user = config['configurations']['cluster-env']['smokeuser'] -smoke_hdfs_user_dir = format("/user/{smoke_user}") -smoke_hdfs_user_mode = 0770 - - -##### Namenode RPC ports - metrics config section start ##### - -# Figure out the rpc ports for current namenode -nn_rpc_client_port = None -nn_rpc_dn_port = None -nn_rpc_healthcheck_port = None - -namenode_id = None -namenode_rpc = None - -dfs_ha_enabled = False -dfs_ha_nameservices = default('/configurations/hdfs-site/dfs.internal.nameservices', None) -if dfs_ha_nameservices is None: - dfs_ha_nameservices = default('/configurations/hdfs-site/dfs.nameservices', None) -dfs_ha_namenode_ids = default(format("/configurations/hdfs-site/dfs.ha.namenodes.{dfs_ha_nameservices}"), None) - -dfs_ha_namemodes_ids_list = [] -other_namenode_id = None - -if dfs_ha_namenode_ids: - dfs_ha_namemodes_ids_list = dfs_ha_namenode_ids.split(",") - dfs_ha_namenode_ids_array_len = len(dfs_ha_namemodes_ids_list) - if dfs_ha_namenode_ids_array_len > 1: - dfs_ha_enabled = True - -if dfs_ha_enabled: - for nn_id in dfs_ha_namemodes_ids_list: - nn_host = config['configurations']['hdfs-site'][format('dfs.namenode.rpc-address.{dfs_ha_nameservices}.{nn_id}')] - if hostname in nn_host: - namenode_id = nn_id - namenode_rpc = nn_host - pass - pass -else: - namenode_rpc = default('/configurations/hdfs-site/dfs.namenode.rpc-address', None) - -if namenode_rpc: - nn_rpc_client_port = namenode_rpc.split(':')[1].strip() - -if dfs_ha_enabled: - dfs_service_rpc_address = default(format('/configurations/hdfs-site/dfs.namenode.servicerpc-address.{dfs_ha_nameservices}.{namenode_id}'), None) - dfs_lifeline_rpc_address = default(format('/configurations/hdfs-site/dfs.namenode.lifeline.rpc-address.{dfs_ha_nameservices}.{namenode_id}'), None) -else: - dfs_service_rpc_address = default('/configurations/hdfs-site/dfs.namenode.servicerpc-address', None) - dfs_lifeline_rpc_address = default(format('/configurations/hdfs-site/dfs.namenode.lifeline.rpc-address'), None) - -if dfs_service_rpc_address: - nn_rpc_dn_port = dfs_service_rpc_address.split(':')[1].strip() - -if dfs_lifeline_rpc_address: - nn_rpc_healthcheck_port = dfs_lifeline_rpc_address.split(':')[1].strip() - -is_nn_client_port_configured = False if nn_rpc_client_port is None else True -is_nn_dn_port_configured = False if nn_rpc_dn_port is None else True -is_nn_healthcheck_port_configured = False if nn_rpc_healthcheck_port is None else True - -##### end ##### - -import functools -#create partial functions with common arguments for every HdfsResource call -#to create/delete/copyfromlocal hdfs directories/files we need to call params.HdfsResource in code -HdfsResource = functools.partial( - HdfsResource, - user=hdfs_user, - hdfs_resource_ignore_file = "/var/lib/ambari-agent/data/.hdfs_resource_ignore", - security_enabled = security_enabled, - keytab = hdfs_user_keytab, - kinit_path_local = kinit_path_local, - hadoop_bin_dir = hadoop_bin_dir, - hadoop_conf_dir = hadoop_conf_dir, - principal_name = hdfs_principal_name, - hdfs_site = hdfs_site, - default_fs = default_fs, - immutable_paths = get_not_managed_resources(), - dfs_type = dfs_type -)
http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/scripts/rack_awareness.py ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/scripts/rack_awareness.py b/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/scripts/rack_awareness.py deleted file mode 100755 index 548f051..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/scripts/rack_awareness.py +++ /dev/null @@ -1,47 +0,0 @@ -#!/usr/bin/env python - -""" -Licensed to the Apache Software Foundation (ASF) under one -or more contributor license agreements. See the NOTICE file -distributed with this work for additional information -regarding copyright ownership. The ASF licenses this file -to you under the Apache License, Version 2.0 (the -"License"); you may not use this file except in compliance -with the License. You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - -Unless required by applicable law or agreed to in writing, software -distributed under the License is distributed on an "AS IS" BASIS, -WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -See the License for the specific language governing permissions and -limitations under the License. -""" - -from resource_management.core.resources import File -from resource_management.core.source import StaticFile, Template -from resource_management.libraries.functions import format - - -def create_topology_mapping(): - import params - - File(params.net_topology_mapping_data_file_path, - content=Template("topology_mappings.data.j2"), - owner=params.hdfs_user, - group=params.user_group, - only_if=format("test -d {net_topology_script_dir}")) - -def create_topology_script(): - import params - - File(params.net_topology_script_file_path, - content=StaticFile('topology_script.py'), - mode=0755, - only_if=format("test -d {net_topology_script_dir}")) - -def create_topology_script_and_mapping(): - import params - if params.has_hadoop_env: - create_topology_mapping() - create_topology_script() http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/scripts/shared_initialization.py ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/scripts/shared_initialization.py b/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/scripts/shared_initialization.py deleted file mode 100755 index ba9c8fb..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/scripts/shared_initialization.py +++ /dev/null @@ -1,175 +0,0 @@ -""" -Licensed to the Apache Software Foundation (ASF) under one -or more contributor license agreements. See the NOTICE file -distributed with this work for additional information -regarding copyright ownership. The ASF licenses this file -to you under the Apache License, Version 2.0 (the -"License"); you may not use this file except in compliance -with the License. You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - -Unless required by applicable law or agreed to in writing, software -distributed under the License is distributed on an "AS IS" BASIS, -WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -See the License for the specific language governing permissions and -limitations under the License. - -""" - -import os -from resource_management.libraries.providers.hdfs_resource import WebHDFSUtil - -from resource_management import * - -def setup_hadoop(): - """ - Setup hadoop files and directories - """ - import params - - Execute(("setenforce","0"), - only_if="test -f /selinux/enforce", - not_if="(! which getenforce ) || (which getenforce && getenforce | grep -q Disabled)", - sudo=True, - ) - - #directories - if params.has_namenode or params.dfs_type == 'HCFS': - Directory(params.hdfs_log_dir_prefix, - create_parents = True, - owner='root', - group=params.user_group, - mode=0775, - cd_access='a', - ) - if params.has_namenode: - Directory(params.hadoop_pid_dir_prefix, - create_parents = True, - owner='root', - group='root', - cd_access='a', - ) - Directory(params.hadoop_tmp_dir, - create_parents = True, - owner=params.hdfs_user, - cd_access='a', - ) - #files - if params.security_enabled: - tc_owner = "root" - else: - tc_owner = params.hdfs_user - - # if WebHDFS is not enabled we need this jar to create hadoop folders. - if params.host_sys_prepped: - print "Skipping copying of fast-hdfs-resource.jar as host is sys prepped" - elif params.dfs_type == 'HCFS' or not WebHDFSUtil.is_webhdfs_available(params.is_webhdfs_enabled, params.default_fs): - # for source-code of jar goto contrib/fast-hdfs-resource - File(format("{ambari_libs_dir}/fast-hdfs-resource.jar"), - mode=0644, - content=StaticFile("fast-hdfs-resource.jar") - ) - - if os.path.exists(params.hadoop_conf_dir): - File(os.path.join(params.hadoop_conf_dir, 'commons-logging.properties'), - owner=tc_owner, - content=Template('commons-logging.properties.j2') - ) - - health_check_template_name = "health_check" - File(os.path.join(params.hadoop_conf_dir, health_check_template_name), - owner=tc_owner, - content=Template(health_check_template_name + ".j2") - ) - - log4j_filename = os.path.join(params.hadoop_conf_dir, "log4j.properties") - if (params.log4j_props != None): - File(log4j_filename, - mode=0644, - group=params.user_group, - owner=params.hdfs_user, - content=params.log4j_props - ) - elif (os.path.exists(format("{params.hadoop_conf_dir}/log4j.properties"))): - File(log4j_filename, - mode=0644, - group=params.user_group, - owner=params.hdfs_user, - ) - - File(os.path.join(params.hadoop_conf_dir, "hadoop-metrics2.properties"), - owner=params.hdfs_user, - group=params.user_group, - content=Template("hadoop-metrics2.properties.j2") - ) - - if params.dfs_type == 'HCFS' and params.has_core_site and 'ECS_CLIENT' in params.component_list: - create_dirs() - - -def setup_configs(): - """ - Creates configs for services HDFS mapred - """ - import params - - if params.has_namenode or params.dfs_type == 'HCFS': - if os.path.exists(params.hadoop_conf_dir): - File(params.task_log4j_properties_location, - content=StaticFile("task-log4j.properties"), - mode=0755 - ) - - if os.path.exists(os.path.join(params.hadoop_conf_dir, 'configuration.xsl')): - File(os.path.join(params.hadoop_conf_dir, 'configuration.xsl'), - owner=params.hdfs_user, - group=params.user_group - ) - if os.path.exists(os.path.join(params.hadoop_conf_dir, 'masters')): - File(os.path.join(params.hadoop_conf_dir, 'masters'), - owner=params.hdfs_user, - group=params.user_group - ) - - generate_include_file() - - -def generate_include_file(): - import params - - if params.has_namenode and params.dfs_hosts and params.has_slaves: - include_hosts_list = params.slave_hosts - File(params.dfs_hosts, - content=Template("include_hosts_list.j2"), - owner=params.hdfs_user, - group=params.user_group - ) - -def create_javahome_symlink(): - if os.path.exists("/usr/jdk/jdk1.6.0_31") and not os.path.exists("/usr/jdk64/jdk1.6.0_31"): - Directory("/usr/jdk64/", - create_parents = True, - ) - Link("/usr/jdk/jdk1.6.0_31", - to="/usr/jdk64/jdk1.6.0_31", - ) - -def create_dirs(): - import params - params.HdfsResource(params.hdfs_tmp_dir, - type="directory", - action="create_on_execute", - owner=params.hdfs_user, - mode=0777 - ) - params.HdfsResource(params.smoke_hdfs_user_dir, - type="directory", - action="create_on_execute", - owner=params.smoke_user, - mode=params.smoke_hdfs_user_mode - ) - params.HdfsResource(None, - action="execute" - ) - http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/commons-logging.properties.j2 ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/commons-logging.properties.j2 b/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/commons-logging.properties.j2 deleted file mode 100755 index 2197ba5..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/commons-logging.properties.j2 +++ /dev/null @@ -1,43 +0,0 @@ -{# -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -#} - -#/* -# * Licensed to the Apache Software Foundation (ASF) under one -# * or more contributor license agreements. See the NOTICE file -# * distributed with this work for additional information -# * regarding copyright ownership. The ASF licenses this file -# * to you under the Apache License, Version 2.0 (the -# * "License"); you may not use this file except in compliance -# * with the License. You may obtain a copy of the License at -# * -# * http://www.apache.org/licenses/LICENSE-2.0 -# * -# * Unless required by applicable law or agreed to in writing, software -# * distributed under the License is distributed on an "AS IS" BASIS, -# * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# * See the License for the specific language governing permissions and -# * limitations under the License. -# */ - -#Logging Implementation - -#Log4J -org.apache.commons.logging.Log=org.apache.commons.logging.impl.Log4JLogger - -#JDK Logger -#org.apache.commons.logging.Log=org.apache.commons.logging.impl.Jdk14Logger http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/exclude_hosts_list.j2 ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/exclude_hosts_list.j2 b/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/exclude_hosts_list.j2 deleted file mode 100755 index 1adba80..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/exclude_hosts_list.j2 +++ /dev/null @@ -1,21 +0,0 @@ -{# -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -#} - -{% for host in hdfs_exclude_file %} -{{host}} -{% endfor %} http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/hadoop-metrics2.properties.j2 ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/hadoop-metrics2.properties.j2 b/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/hadoop-metrics2.properties.j2 deleted file mode 100755 index fcd9b23..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/hadoop-metrics2.properties.j2 +++ /dev/null @@ -1,104 +0,0 @@ -{# -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -#} - -# Licensed to the Apache Software Foundation (ASF) under one or more -# contributor license agreements. See the NOTICE file distributed with -# this work for additional information regarding copyright ownership. -# The ASF licenses this file to You under the Apache License, Version 2.0 -# (the "License"); you may not use this file except in compliance with -# the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -# syntax: [prefix].[source|sink|jmx].[instance].[options] -# See package.html for org.apache.hadoop.metrics2 for details - -{% if has_ganglia_server %} -*.period=60 - -*.sink.ganglia.class=org.apache.hadoop.metrics2.sink.ganglia.GangliaSink31 -*.sink.ganglia.period=10 - -# default for supportsparse is false -*.sink.ganglia.supportsparse=true - -.sink.ganglia.slope=jvm.metrics.gcCount=zero,jvm.metrics.memHeapUsedM=both -.sink.ganglia.dmax=jvm.metrics.threadsBlocked=70,jvm.metrics.memHeapUsedM=40 - -# Hook up to the server -namenode.sink.ganglia.servers={{ganglia_server_host}}:8661 -datanode.sink.ganglia.servers={{ganglia_server_host}}:8659 -jobtracker.sink.ganglia.servers={{ganglia_server_host}}:8662 -tasktracker.sink.ganglia.servers={{ganglia_server_host}}:8658 -maptask.sink.ganglia.servers={{ganglia_server_host}}:8660 -reducetask.sink.ganglia.servers={{ganglia_server_host}}:8660 -resourcemanager.sink.ganglia.servers={{ganglia_server_host}}:8664 -nodemanager.sink.ganglia.servers={{ganglia_server_host}}:8657 -historyserver.sink.ganglia.servers={{ganglia_server_host}}:8666 -journalnode.sink.ganglia.servers={{ganglia_server_host}}:8654 -nimbus.sink.ganglia.servers={{ganglia_server_host}}:8649 -supervisor.sink.ganglia.servers={{ganglia_server_host}}:8650 - -resourcemanager.sink.ganglia.tagsForPrefix.yarn=Queue - -{% endif %} - -{% if has_metric_collector %} - -*.period={{metrics_collection_period}} -*.sink.timeline.plugin.urls=file:///usr/lib/ambari-metrics-hadoop-sink/ambari-metrics-hadoop-sink.jar -*.sink.timeline.class=org.apache.hadoop.metrics2.sink.timeline.HadoopTimelineMetricsSink -*.sink.timeline.period={{metrics_collection_period}} -*.sink.timeline.sendInterval={{metrics_report_interval}}000 -*.sink.timeline.slave.host.name = {{hostname}} - -# HTTPS properties -*.sink.timeline.truststore.path = {{metric_truststore_path}} -*.sink.timeline.truststore.type = {{metric_truststore_type}} -*.sink.timeline.truststore.password = {{metric_truststore_password}} - -datanode.sink.timeline.collector={{metric_collector_protocol}}://{{metric_collector_host}}:{{metric_collector_port}} -namenode.sink.timeline.collector={{metric_collector_protocol}}://{{metric_collector_host}}:{{metric_collector_port}} -resourcemanager.sink.timeline.collector={{metric_collector_protocol}}://{{metric_collector_host}}:{{metric_collector_port}} -nodemanager.sink.timeline.collector={{metric_collector_protocol}}://{{metric_collector_host}}:{{metric_collector_port}} -jobhistoryserver.sink.timeline.collector={{metric_collector_protocol}}://{{metric_collector_host}}:{{metric_collector_port}} -journalnode.sink.timeline.collector={{metric_collector_protocol}}://{{metric_collector_host}}:{{metric_collector_port}} -maptask.sink.timeline.collector={{metric_collector_protocol}}://{{metric_collector_host}}:{{metric_collector_port}} -reducetask.sink.timeline.collector={{metric_collector_protocol}}://{{metric_collector_host}}:{{metric_collector_port}} -applicationhistoryserver.sink.timeline.collector={{metric_collector_protocol}}://{{metric_collector_host}}:{{metric_collector_port}} - -resourcemanager.sink.timeline.tagsForPrefix.yarn=Queue - -{% if is_nn_client_port_configured %} -# Namenode rpc ports customization -namenode.sink.timeline.metric.rpc.client.port={{nn_rpc_client_port}} -{% endif %} -{% if is_nn_dn_port_configured %} -namenode.sink.timeline.metric.rpc.datanode.port={{nn_rpc_dn_port}} -{% endif %} -{% if is_nn_healthcheck_port_configured %} -namenode.sink.timeline.metric.rpc.healthcheck.port={{nn_rpc_healthcheck_port}} -{% endif %} - -{% endif %} http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/health_check.j2 ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/health_check.j2 b/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/health_check.j2 deleted file mode 100755 index 0a03d17..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/health_check.j2 +++ /dev/null @@ -1,81 +0,0 @@ -{# -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -#} - -#!/bin/bash -# -#/* -# * Licensed to the Apache Software Foundation (ASF) under one -# * or more contributor license agreements. See the NOTICE file -# * distributed with this work for additional information -# * regarding copyright ownership. The ASF licenses this file -# * to you under the Apache License, Version 2.0 (the -# * "License"); you may not use this file except in compliance -# * with the License. You may obtain a copy of the License at -# * -# * http://www.apache.org/licenses/LICENSE-2.0 -# * -# * Unless required by applicable law or agreed to in writing, software -# * distributed under the License is distributed on an "AS IS" BASIS, -# * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# * See the License for the specific language governing permissions and -# * limitations under the License. -# */ - -err=0; - -function check_disks { - - for m in `awk '$3~/ext3/ {printf" %s ",$2}' /etc/fstab` ; do - fsdev="" - fsdev=`awk -v m=$m '$2==m {print $1}' /proc/mounts`; - if [ -z "$fsdev" -a "$m" != "/mnt" ] ; then - msg_="$msg_ $m(u)" - else - msg_="$msg_`awk -v m=$m '$2==m { if ( $4 ~ /^ro,/ ) {printf"%s(ro)",$2 } ; }' /proc/mounts`" - fi - done - - if [ -z "$msg_" ] ; then - echo "disks ok" ; exit 0 - else - echo "$msg_" ; exit 2 - fi - -} - -# Run all checks -for check in disks ; do - msg=`check_${check}` ; - if [ $? -eq 0 ] ; then - ok_msg="$ok_msg$msg," - else - err_msg="$err_msg$msg," - fi -done - -if [ ! -z "$err_msg" ] ; then - echo -n "ERROR $err_msg " -fi -if [ ! -z "$ok_msg" ] ; then - echo -n "OK: $ok_msg" -fi - -echo - -# Success! -exit 0 http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/include_hosts_list.j2 ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/include_hosts_list.j2 b/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/include_hosts_list.j2 deleted file mode 100755 index 4a9e713..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/include_hosts_list.j2 +++ /dev/null @@ -1,21 +0,0 @@ -{# -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -#} - -{% for host in slave_hosts %} -{{host}} -{% endfor %} http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/topology_mappings.data.j2 ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/topology_mappings.data.j2 b/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/topology_mappings.data.j2 deleted file mode 100755 index 15034d6..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/hooks/before-START/templates/topology_mappings.data.j2 +++ /dev/null @@ -1,24 +0,0 @@ -{# -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 - # -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -#} -[network_topology] -{% for host in all_hosts %} -{% if host in slave_hosts %} -{{host}}={{all_racks[loop.index-1]}} -{{all_ipv4_ips[loop.index-1]}}={{all_racks[loop.index-1]}} -{% endif %} -{% endfor %} http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/kerberos.json ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/kerberos.json b/bigtop-packages/src/common/ambari/ODPi/1.0/kerberos.json deleted file mode 100644 index 3aad080..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/kerberos.json +++ /dev/null @@ -1,60 +0,0 @@ -{ - "properties": { - "realm": "${kerberos-env/realm}", - "keytab_dir": "/etc/security/keytabs", - "additional_realms": "" - }, - "identities": [ - { - "name": "spnego", - "principal": { - "value": "HTTP/_HOST@${realm}", - "type" : "service" - }, - "keytab": { - "file": "${keytab_dir}/spnego.service.keytab", - "owner": { - "name": "root", - "access": "r" - }, - "group": { - "name": "${cluster-env/user_group}", - "access": "r" - } - } - }, - { - "name": "smokeuser", - "principal": { - "value": "${cluster-env/smokeuser}-${cluster_name|toLower()}@${realm}", - "type" : "user", - "configuration": "cluster-env/smokeuser_principal_name", - "local_username" : "${cluster-env/smokeuser}" - }, - "keytab": { - "file": "${keytab_dir}/smokeuser.headless.keytab", - "owner": { - "name": "${cluster-env/smokeuser}", - "access": "r" - }, - "group": { - "name": "${cluster-env/user_group}", - "access": "r" - }, - "configuration": "cluster-env/smokeuser_keytab" - } - }, - { - "name": "ambari-server", - "principal": { - "value": "ambari-server-${cluster_name|toLower()}@${realm}", - "type" : "user", - "configuration": "cluster-env/ambari_principal_name" - }, - "keytab": { - "file": "${keytab_dir}/ambari.server.keytab" - } - } - ] - -} http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/metainfo.xml ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/metainfo.xml b/bigtop-packages/src/common/ambari/ODPi/1.0/metainfo.xml deleted file mode 100755 index ca45822..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/metainfo.xml +++ /dev/null @@ -1,22 +0,0 @@ -<?xml version="1.0"?> -<!-- - Licensed to the Apache Software Foundation (ASF) under one or more - contributor license agreements. See the NOTICE file distributed with - this work for additional information regarding copyright ownership. - The ASF licenses this file to You under the Apache License, Version 2.0 - (the "License"); you may not use this file except in compliance with - the License. You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - - Unless required by applicable law or agreed to in writing, software - distributed under the License is distributed on an "AS IS" BASIS, - WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - See the License for the specific language governing permissions and - limitations under the License. ---> -<metainfo> - <versions> - <active>true</active> - </versions> -</metainfo> http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/properties/stack_features.json ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/properties/stack_features.json b/bigtop-packages/src/common/ambari/ODPi/1.0/properties/stack_features.json deleted file mode 100755 index 0c3e305..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/properties/stack_features.json +++ /dev/null @@ -1,51 +0,0 @@ -{ - "stack_features": [ - { - "name": "snappy", - "description": "Snappy compressor/decompressor support", - "min_version": "2.0.0.0", - "max_version": "2.2.0.0" - }, - { - "name": "lzo", - "description": "LZO libraries support", - "min_version": "2.2.1.0" - }, - { - "name": "copy_tarball_to_hdfs", - "description": "Copy tarball to HDFS support (AMBARI-12113)", - "min_version": "2.2.0.0" - }, - { - "name": "hive_metastore_upgrade_schema", - "description": "Hive metastore upgrade schema support (AMBARI-11176)", - "min_version": "2.3.0.0" - }, - { - "name": "hive_webhcat_specific_configs", - "description": "Hive webhcat specific configurations support (AMBARI-12364)", - "min_version": "2.3.0.0" - }, - { - "name": "hive_purge_table", - "description": "Hive purge table support (AMBARI-12260)", - "min_version": "2.3.0.0" - }, - { - "name": "hive_server2_kerberized_env", - "description": "Hive server2 working on kerberized environment (AMBARI-13749)", - "min_version": "2.2.3.0", - "max_version": "2.2.5.0" - }, - { - "name": "hive_env_heapsize", - "description": "Hive heapsize property defined in hive-env (AMBARI-12801)", - "min_version": "2.2.0.0" - }, - { - "name": "hive_metastore_site_support", - "description": "Hive Metastore site support", - "min_version": "2.5.0.0" - } - ] -} http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/properties/stack_tools.json ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/properties/stack_tools.json b/bigtop-packages/src/common/ambari/ODPi/1.0/properties/stack_tools.json deleted file mode 100755 index c3df235..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/properties/stack_tools.json +++ /dev/null @@ -1,4 +0,0 @@ -{ - "stack_selector": ["distro-select", "/usr/bin/distro-select", "distro-select"], - "conf_selector": ["conf-select", "/usr/bin/conf-select", "conf-select"] -} http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/repos/repoinfo.xml ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/repos/repoinfo.xml b/bigtop-packages/src/common/ambari/ODPi/1.0/repos/repoinfo.xml deleted file mode 100755 index 60eae65..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/repos/repoinfo.xml +++ /dev/null @@ -1,33 +0,0 @@ -<?xml version="1.0"?> -<!-- - Licensed to the Apache Software Foundation (ASF) under one or more - contributor license agreements. See the NOTICE file distributed with - this work for additional information regarding copyright ownership. - The ASF licenses this file to You under the Apache License, Version 2.0 - (the "License"); you may not use this file except in compliance with - the License. You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - - Unless required by applicable law or agreed to in writing, software - distributed under the License is distributed on an "AS IS" BASIS, - WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - See the License for the specific language governing permissions and - limitations under the License. ---> -<reposinfo> - <os family="redhat6"> - <repo> - <baseurl>http://repo.odpi.org/ODPi/trunk/centos-6/</baseurl> - <repoid>ODPi-trunk</repoid> - <reponame>ODPi</reponame> - </repo> - </os> - <os family="ubuntu14"> - <repo> - <baseurl>http://repo.odpi.org/ODPi/trunk/ubuntu-14.04/apt</baseurl> - <repoid>ODPi-trunk</repoid> - <reponame>odpi</reponame> - </repo> - </os> -</reposinfo> http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/role_command_order.json ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/role_command_order.json b/bigtop-packages/src/common/ambari/ODPi/1.0/role_command_order.json deleted file mode 100755 index 05beb76..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/role_command_order.json +++ /dev/null @@ -1,75 +0,0 @@ -{ - "_comment" : "Record format:", - "_comment" : "blockedRole-blockedCommand: [blockerRole1-blockerCommand1, blockerRole2-blockerCommand2, ...]", - "general_deps" : { - "_comment" : "dependencies for all cases", - "HBASE_MASTER-START": ["ZOOKEEPER_SERVER-START"], - "HBASE_REGIONSERVER-START": ["HBASE_MASTER-START"], - "APP_TIMELINE_SERVER-START": ["NAMENODE-START", "DATANODE-START"], - "OOZIE_SERVER-START": ["NODEMANAGER-START", "RESOURCEMANAGER-START"], - "WEBHCAT_SERVER-START": ["NODEMANAGER-START", "HIVE_SERVER-START"], - "WEBHCAT_SERVER-RESTART": ["NODEMANAGER-RESTART", "HIVE_SERVER-RESTART"], - "HIVE_METASTORE-START": ["MYSQL_SERVER-START", "NAMENODE-START"], - "HIVE_METASTORE-RESTART": ["MYSQL_SERVER-RESTART", "NAMENODE-RESTART"], - "HIVE_SERVER-START": ["NODEMANAGER-START", "MYSQL_SERVER-START"], - "HIVE_SERVER-RESTART": ["NODEMANAGER-RESTART", "MYSQL_SERVER-RESTART", "ZOOKEEPER_SERVER-RESTART"], - "HUE_SERVER-START": ["HIVE_SERVER-START", "HCAT-START", "OOZIE_SERVER-START"], - "FLUME_HANDLER-START": ["OOZIE_SERVER-START"], - "MAPREDUCE_SERVICE_CHECK-SERVICE_CHECK": ["NODEMANAGER-START", "RESOURCEMANAGER-START"], - "OOZIE_SERVICE_CHECK-SERVICE_CHECK": ["OOZIE_SERVER-START", "MAPREDUCE2_SERVICE_CHECK-SERVICE_CHECK"], - "HBASE_SERVICE_CHECK-SERVICE_CHECK": ["HBASE_MASTER-START", "HBASE_REGIONSERVER-START"], - "HIVE_SERVICE_CHECK-SERVICE_CHECK": ["HIVE_SERVER-START", "HIVE_METASTORE-START", "WEBHCAT_SERVER-START"], - "PIG_SERVICE_CHECK-SERVICE_CHECK": ["NODEMANAGER-START", "RESOURCEMANAGER-START"], - "SQOOP_SERVICE_CHECK-SERVICE_CHECK": ["NODEMANAGER-START", "RESOURCEMANAGER-START"], - "ZOOKEEPER_SERVICE_CHECK-SERVICE_CHECK": ["ZOOKEEPER_SERVER-START"], - "ZOOKEEPER_QUORUM_SERVICE_CHECK-SERVICE_CHECK": ["ZOOKEEPER_SERVER-START"], - "ZOOKEEPER_SERVER-STOP" : ["HBASE_MASTER-STOP", "HBASE_REGIONSERVER-STOP", "METRICS_COLLECTOR-STOP"], - "HBASE_MASTER-STOP": ["HBASE_REGIONSERVER-STOP"] - }, - "_comment" : "GLUSTERFS-specific dependencies", - "optional_glusterfs": { - "HBASE_MASTER-START": ["PEERSTATUS-START"], - "GLUSTERFS_SERVICE_CHECK-SERVICE_CHECK": ["PEERSTATUS-START"] - }, - "_comment" : "Dependencies that are used when GLUSTERFS is not present in cluster", - "optional_no_glusterfs": { - "METRICS_COLLECTOR-START": ["NAMENODE-START", "DATANODE-START", "SECONDARY_NAMENODE-START", "ZOOKEEPER_SERVER-START"], - "AMBARI_METRICS_SERVICE_CHECK-SERVICE_CHECK": ["METRICS_COLLECTOR-START", "HDFS_SERVICE_CHECK-SERVICE_CHECK"], - "SECONDARY_NAMENODE-START": ["NAMENODE-START"], - "SECONDARY_NAMENODE-RESTART": ["NAMENODE-RESTART"], - "RESOURCEMANAGER-START": ["NAMENODE-START", "DATANODE-START"], - "NODEMANAGER-START": ["NAMENODE-START", "DATANODE-START", "RESOURCEMANAGER-START"], - "HISTORYSERVER-START": ["NAMENODE-START", "DATANODE-START"], - "HBASE_MASTER-START": ["NAMENODE-START", "DATANODE-START"], - "HIVE_SERVER-START": ["DATANODE-START"], - "WEBHCAT_SERVER-START": ["DATANODE-START"], - "HISTORYSERVER-RESTART": ["NAMENODE-RESTART"], - "RESOURCEMANAGER-RESTART": ["NAMENODE-RESTART"], - "NODEMANAGER-RESTART": ["NAMENODE-RESTART"], - "OOZIE_SERVER-RESTART": ["NAMENODE-RESTART"], - "HDFS_SERVICE_CHECK-SERVICE_CHECK": ["NAMENODE-START", "DATANODE-START", - "SECONDARY_NAMENODE-START"], - "MAPREDUCE2_SERVICE_CHECK-SERVICE_CHECK": ["NODEMANAGER-START", - "RESOURCEMANAGER-START", "HISTORYSERVER-START", "YARN_SERVICE_CHECK-SERVICE_CHECK"], - "YARN_SERVICE_CHECK-SERVICE_CHECK": ["NODEMANAGER-START", "RESOURCEMANAGER-START"], - "RESOURCEMANAGER_SERVICE_CHECK-SERVICE_CHECK": ["RESOURCEMANAGER-START"], - "PIG_SERVICE_CHECK-SERVICE_CHECK": ["RESOURCEMANAGER-START", "NODEMANAGER-START"], - "NAMENODE-STOP": ["RESOURCEMANAGER-STOP", "NODEMANAGER-STOP", - "HISTORYSERVER-STOP", "HBASE_MASTER-STOP", "METRICS_COLLECTOR-STOP"], - "DATANODE-STOP": ["RESOURCEMANAGER-STOP", "NODEMANAGER-STOP", - "HISTORYSERVER-STOP", "HBASE_MASTER-STOP"], - "METRICS_GRAFANA-START": ["METRICS_COLLECTOR-START"], - "METRICS_COLLECTOR-STOP": ["METRICS_GRAFANA-STOP"] - }, - "_comment" : "Dependencies that are used in HA NameNode cluster", - "namenode_optional_ha": { - "NAMENODE-START": ["ZKFC-START", "JOURNALNODE-START", "ZOOKEEPER_SERVER-START"], - "ZKFC-START": ["ZOOKEEPER_SERVER-START"], - "ZKFC-STOP": ["NAMENODE-STOP"], - "JOURNALNODE-STOP": ["NAMENODE-STOP"] - }, - "_comment" : "Dependencies that are used in ResourceManager HA cluster", - "resourcemanager_optional_ha" : { - "RESOURCEMANAGER-START": ["ZOOKEEPER_SERVER-START"] - } -} http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/services/HDFS/metainfo.xml ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HDFS/metainfo.xml b/bigtop-packages/src/common/ambari/ODPi/1.0/services/HDFS/metainfo.xml deleted file mode 100755 index d6e30b7..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HDFS/metainfo.xml +++ /dev/null @@ -1,27 +0,0 @@ -<?xml version="1.0"?> -<!-- - Licensed to the Apache Software Foundation (ASF) under one or more - contributor license agreements. See the NOTICE file distributed with - this work for additional information regarding copyright ownership. - The ASF licenses this file to You under the Apache License, Version 2.0 - (the "License"); you may not use this file except in compliance with - the License. You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - - Unless required by applicable law or agreed to in writing, software - distributed under the License is distributed on an "AS IS" BASIS, - WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - See the License for the specific language governing permissions and - limitations under the License. ---> -<metainfo> - <schemaVersion>2.0</schemaVersion> - <services> - <service> - <name>HDFS</name> - <version>2.7.1+odpi</version> - <extends>common-services/HDFS/2.1.0.2.0</extends> - </service> - </services> -</metainfo> http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/alerts.json ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/alerts.json b/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/alerts.json deleted file mode 100755 index e2431c3..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/alerts.json +++ /dev/null @@ -1,232 +0,0 @@ -{ - "HIVE": { - "service": [], - "HIVE_METASTORE": [ - { - "name": "hive_metastore_process", - "label": "Hive Metastore Process", - "description": "This host-level alert is triggered if the Hive Metastore process cannot be determined to be up and listening on the network.", - "interval": 3, - "scope": "ANY", - "enabled": true, - "source": { - "type": "SCRIPT", - "path": "HIVE/0.12.0.2.0/package/alerts/alert_hive_metastore.py", - "parameters": [ - { - "name": "check.command.timeout", - "display_name": "Command Timeout", - "value": 60.0, - "type": "NUMERIC", - "description": "The maximum time before check command will be killed by timeout", - "units": "seconds", - "threshold": "CRITICAL" - }, - { - "name": "default.smoke.user", - "display_name": "Default Smoke User", - "value": "ambari-qa", - "type": "STRING", - "description": "The user that will run the Hive commands if not specified in cluster-env/smokeuser", - "visibility": "HIDDEN" - }, - { - "name": "default.smoke.principal", - "display_name": "Default Smoke Principal", - "value": "[email protected]", - "type": "STRING", - "description": "The principal to use when retrieving the kerberos ticket if not specified in cluster-env/smokeuser_principal_name", - "visibility": "HIDDEN" - }, - { - "name": "default.smoke.keytab", - "display_name": "Default Smoke Keytab", - "value": "/etc/security/keytabs/smokeuser.headless.keytab", - "type": "STRING", - "description": "The keytab to use when retrieving the kerberos ticket if not specified in cluster-env/smokeuser_keytab", - "visibility": "HIDDEN" - } - ] - } - } - ], - "HIVE_SERVER": [ - { - "name": "hive_server_process", - "label": "HiveServer2 Process", - "description": "This host-level alert is triggered if the HiveServer cannot be determined to be up and responding to client requests.", - "interval": 3, - "scope": "ANY", - "enabled": true, - "source": { - "type": "SCRIPT", - "path": "HIVE/0.12.0.2.0/package/alerts/alert_hive_thrift_port.py", - "parameters": [ - { - "name": "check.command.timeout", - "display_name": "Command Timeout", - "value": 60.0, - "type": "NUMERIC", - "description": "The maximum time before check command will be killed by timeout", - "units": "seconds", - "threshold": "CRITICAL" - }, - { - "name": "default.smoke.user", - "display_name": "Default Smoke User", - "value": "ambari-qa", - "type": "STRING", - "description": "The user that will run the Hive commands if not specified in cluster-env/smokeuser", - "visibility": "HIDDEN" - }, - { - "name": "default.smoke.principal", - "display_name": "Default Smoke Principal", - "value": "[email protected]", - "type": "STRING", - "description": "The principal to use when retrieving the kerberos ticket if not specified in cluster-env/smokeuser_principal_name", - "visibility": "HIDDEN" - }, - { - "name": "default.smoke.keytab", - "display_name": "Default Smoke Keytab", - "value": "/etc/security/keytabs/smokeuser.headless.keytab", - "type": "STRING", - "description": "The keytab to use when retrieving the kerberos ticket if not specified in cluster-env/smokeuser_keytab", - "visibility": "HIDDEN" - } - ] - } - } - ], - "HIVE_SERVER_INTERACTIVE": [ - { - "name": "hive_server_interactive_process", - "label": "HiveServer2 Interactive Process", - "description": "This host-level alert is triggered if the HiveServerInteractive cannot be determined to be up and responding to client requests.", - "interval": 3, - "scope": "ANY", - "enabled": true, - "source": { - "type": "SCRIPT", - "path": "HIVE/0.12.0.2.0/package/alerts/alert_hive_interactive_thrift_port.py", - "parameters": [ - { - "name": "check.command.timeout", - "display_name": "Command Timeout", - "value": 60.0, - "type": "NUMERIC", - "description": "The maximum time before check command will be killed by timeout", - "units": "seconds", - "threshold": "CRITICAL" - }, - { - "name": "default.smoke.user", - "display_name": "Default Smoke User", - "value": "ambari-qa", - "type": "STRING", - "description": "The user that will run the Hive commands if not specified in cluster-env/smokeuser", - "visibility": "HIDDEN" - }, - { - "name": "default.smoke.principal", - "display_name": "Default Smoke Principal", - "value": "[email protected]", - "type": "STRING", - "description": "The principal to use when retrieving the kerberos ticket if not specified in cluster-env/smokeuser_principal_name", - "visibility": "HIDDEN" - }, - { - "name": "default.smoke.keytab", - "display_name": "Default Smoke Keytab", - "value": "/etc/security/keytabs/smokeuser.headless.keytab", - "type": "STRING", - "description": "The keytab to use when retrieving the kerberos ticket if not specified in cluster-env/smokeuser_keytab", - "visibility": "HIDDEN" - } - ] - } - }, - { - "name": "llap_application", - "label": "LLAP Application", - "description": "This alert is triggered if the LLAP Application cannot be determined to be up and responding to requests.", - "interval": 3, - "scope": "ANY", - "enabled": true, - "source": { - "type": "SCRIPT", - "path": "HIVE/0.12.0.2.0/package/alerts/alert_llap_app_status.py", - "parameters": [ - { - "name": "check.command.timeout", - "display_name": "Command Timeout", - "value": 120.0, - "type": "NUMERIC", - "description": "The maximum time before check command will be killed by timeout", - "units": "seconds", - "threshold": "CRITICAL" - }, - { - "name": "default.hive.user", - "display_name": "Default HIVE User", - "value": "hive", - "type": "STRING", - "description": "The user that will run the Hive commands if not specified in cluster-env", - "visibility": "HIDDEN" - }, - { - "name": "default.hive.principal", - "display_name": "Default HIVE Principal", - "value": "hive/[email protected]", - "type": "STRING", - "description": "The principal to use when retrieving the kerberos ticket if not specified in cluster-env", - "visibility": "HIDDEN" - }, - { - "name": "default.hive.keytab", - "display_name": "Default HIVE Keytab", - "value": "/etc/security/keytabs/hive.llap.zk.sm.keytab", - "type": "STRING", - "description": "The keytab to use when retrieving the kerberos ticket if not specified in cluster-env.", - "visibility": "HIDDEN" - } - ] - } - } - ], - "WEBHCAT_SERVER": [ - { - "name": "hive_webhcat_server_status", - "label": "WebHCat Server Status", - "description": "This host-level alert is triggered if the templeton server status is not healthy.", - "interval": 1, - "scope": "ANY", - "enabled": true, - "source": { - "type": "SCRIPT", - "path": "HIVE/0.12.0.2.0/package/alerts/alert_webhcat_server.py", - "parameters": [ - { - "name": "default.smoke.user", - "display_name": "Default Smoke User", - "value": "ambari-qa", - "type": "STRING", - "description": "The user that will run the Hive commands if not specified in cluster-env/smokeuser", - "visibility": "HIDDEN" - }, - { - "name": "connection.timeout", - "display_name": "Connection Timeout", - "value": 5.0, - "type": "NUMERIC", - "description": "The maximum time before this alert is considered to be CRITICAL", - "units": "seconds", - "threshold": "CRITICAL" - } - ] - } - } - ] - } -} http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/configuration/hcat-env.xml ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/configuration/hcat-env.xml b/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/configuration/hcat-env.xml deleted file mode 100755 index 3908d61..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/configuration/hcat-env.xml +++ /dev/null @@ -1,41 +0,0 @@ -<configuration><property require-input="false"> - <name>content</name> - <value> - # Licensed to the Apache Software Foundation (ASF) under one - # or more contributor license agreements. See the NOTICE file - # distributed with this work for additional information - # regarding copyright ownership. The ASF licenses this file - # to you under the Apache License, Version 2.0 (the - # "License"); you may not use this file except in compliance - # with the License. You may obtain a copy of the License at - # - # http://www.apache.org/licenses/LICENSE-2.0 - # - # Unless required by applicable law or agreed to in writing, software - # distributed under the License is distributed on an "AS IS" BASIS, - # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - # See the License for the specific language governing permissions and - # limitations under the License. - - JAVA_HOME={{java64_home}} - HCAT_PID_DIR={{hcat_pid_dir}}/ - HCAT_LOG_DIR={{hcat_log_dir}}/ - HCAT_CONF_DIR={{hcat_conf_dir}} - HADOOP_HOME=${HADOOP_HOME:-{{hadoop_home}}} - #DBROOT is the path where the connector jars are downloaded - DBROOT={{hcat_dbroot}} - USER={{hcat_user}} - METASTORE_PORT={{hive_metastore_port}} - </value> - <description>This is the jinja template for hcat-env.sh file</description> - <display-name>hcat-env template</display-name> - <filename>hcat-env.xml</filename> - <deleted>false</deleted> - <on-ambari-upgrade add="true" delete="false" update="false"/> - <property-type></property-type> - <value-attributes> - <type>content</type> - </value-attributes> - <depends-on/> - <property_depended_by/> -</property></configuration> \ No newline at end of file http://git-wip-us.apache.org/repos/asf/bigtop/blob/0d3448b8/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/configuration/hive-env.xml ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/configuration/hive-env.xml b/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/configuration/hive-env.xml deleted file mode 100755 index e5ed319..0000000 --- a/bigtop-packages/src/common/ambari/ODPi/1.0/services/HIVE/configuration/hive-env.xml +++ /dev/null @@ -1,540 +0,0 @@ -<configuration><property require-input="false"> - <name>content</name> - <value> - export HADOOP_USER_CLASSPATH_FIRST=true #this prevents old metrics libs from mapreduce lib from bringing in old jar deps overriding HIVE_LIB - if [ "$SERVICE" = "cli" ]; then - if [ -z "$DEBUG" ]; then - export HADOOP_OPTS="$HADOOP_OPTS -XX:NewRatio=12 -XX:MaxHeapFreeRatio=40 -XX:MinHeapFreeRatio=15 -XX:+UseNUMA -XX:+UseParallelGC -XX:-UseGCOverheadLimit" - else - export HADOOP_OPTS="$HADOOP_OPTS -XX:NewRatio=12 -XX:MaxHeapFreeRatio=40 -XX:MinHeapFreeRatio=15 -XX:-UseGCOverheadLimit" - fi - fi - - # The heap size of the jvm stared by hive shell script can be controlled via: - - if [ "$SERVICE" = "metastore" ]; then - export HADOOP_HEAPSIZE={{hive_metastore_heapsize}} # Setting for HiveMetastore - else - export HADOOP_HEAPSIZE={{hive_heapsize}} # Setting for HiveServer2 and Client - fi - - export HADOOP_CLIENT_OPTS="$HADOOP_CLIENT_OPTS -Xmx${HADOOP_HEAPSIZE}m" - - # Larger heap size may be required when running queries over large number of files or partitions. - # By default hive shell scripts use a heap size of 256 (MB). Larger heap size would also be - # appropriate for hive server (hwi etc). - - - # Set HADOOP_HOME to point to a specific hadoop install directory - HADOOP_HOME=${HADOOP_HOME:-{{hadoop_home}}} - - export HIVE_HOME=${HIVE_HOME:-{{hive_home_dir}}} - - # Hive Configuration Directory can be controlled by: - export HIVE_CONF_DIR=${HIVE_CONF_DIR:-{{hive_config_dir}}} - - # Folder containing extra libraries required for hive compilation/execution can be controlled by: - if [ "${HIVE_AUX_JARS_PATH}" != "" ]; then - if [ -f "${HIVE_AUX_JARS_PATH}" ]; then - export HIVE_AUX_JARS_PATH=${HIVE_AUX_JARS_PATH} - elif [ -d "/usr/hdp/current/hive-webhcat/share/hcatalog" ]; then - export HIVE_AUX_JARS_PATH=/usr/hdp/current/hive-webhcat/share/hcatalog/hive-hcatalog-core.jar - fi - elif [ -d "/usr/hdp/current/hive-webhcat/share/hcatalog" ]; then - export HIVE_AUX_JARS_PATH=/usr/hdp/current/hive-webhcat/share/hcatalog/hive-hcatalog-core.jar - fi - - export METASTORE_PORT={{hive_metastore_port}} - - {% if sqla_db_used or lib_dir_available %} - export LD_LIBRARY_PATH="$LD_LIBRARY_PATH:{{jdbc_libs_dir}}" - export JAVA_LIBRARY_PATH="$JAVA_LIBRARY_PATH:{{jdbc_libs_dir}}" - {% endif %} - </value> - <description>This is the jinja template for hive-env.sh file</description> - <display-name>hive-env template</display-name> - <filename>hive-env.xml</filename> - <deleted>false</deleted> - <on-ambari-upgrade add="true" delete="false" update="false"/> - <property-type></property-type> - <value-attributes> - <type>content</type> - </value-attributes> - <depends-on/> - <property_depended_by/> -</property><property require-input="false"> - <name>hive.heapsize</name> - <value>512</value> - <description>Hive Java heap size</description> - <display-name>HiveServer2 Heap Size</display-name> - <filename>hive-env.xml</filename> - <deleted>false</deleted> - <on-ambari-upgrade add="true" delete="false" update="false"/> - <property-type></property-type> - <value-attributes> - <type>int</type> - <maximum>2048</maximum> - <minimum>512</minimum> - <unit>MB</unit> - <overridable>false</overridable> - <increment-step>512</increment-step> - </value-attributes> - <depends-on/> - <property_depended_by/> -</property><property require-input="false"> - <name>hive_exec_orc_storage_strategy</name> - <value>SPEED</value> - <description> - Define the encoding strategy to use while writing data. Changing this will only affect the light weight encoding for integers. - This flag will not change the compression level of higher level compression codec (like ZLIB). Possible options are SPEED and COMPRESSION. - </description> - <display-name>ORC Storage Strategy</display-name> - <filename>hive-env.xml</filename> - <deleted>false</deleted> - <on-ambari-upgrade add="true" delete="false" update="false"/> - <property-type></property-type> - <value-attributes> - <type>value-list</type> - <entries> - <entry> - <value>SPEED</value> - <label>Speed</label> - </entry> - <entry> - <value>COMPRESSION</value> - <label>Compression</label> - </entry> - </entries> - <selection-cardinality>1</selection-cardinality> - </value-attributes> - <depends-on/> - <property_depended_by> - <dependedByProperties> - <name>hive.exec.orc.encoding.strategy</name> - <type>hive-site</type> - </dependedByProperties> - <dependedByProperties> - <name>hive.exec.orc.compression.strategy</name> - <type>hive-site</type> - </dependedByProperties> - </property_depended_by> -</property><property require-input="false"> - <name>hive_txn_acid</name> - <value>off</value> - <display-name>ACID Transactions</display-name> - <filename>hive-env.xml</filename> - <deleted>false</deleted> - <on-ambari-upgrade add="true" delete="false" update="false"/> - <property-type></property-type> - <value-attributes> - <type>value-list</type> - <entries> - <entry> - <value>on</value> - <label>On</label> - </entry> - <entry> - <value>off</value> - <label>Off</label> - </entry> - </entries> - <selection-cardinality>1</selection-cardinality> - </value-attributes> - <depends-on/> - <property_depended_by> - <dependedByProperties> - <name>hive.enforce.bucketing</name> - <type>hive-site</type> - </dependedByProperties> - <dependedByProperties> - <name>hive.exec.dynamic.partition.mode</name> - <type>hive-site</type> - </dependedByProperties> - <dependedByProperties> - <name>hive.support.concurrency</name> - <type>hive-site</type> - </dependedByProperties> - <dependedByProperties> - <name>hive.txn.manager</name> - <type>hive-site</type> - </dependedByProperties> - <dependedByProperties> - <name>hive.compactor.initiator.on</name> - <type>hive-site</type> - </dependedByProperties> - <dependedByProperties> - <name>hive.compactor.worker.threads</name> - <type>hive-site</type> - </dependedByProperties> - </property_depended_by> -</property><property require-input="false"> - <name>hive_security_authorization</name> - <value>None</value> - <description> - Authorization mode, default NONE. Options are NONE, Ranger, SQLStdAuth. - SQL standard authorization provides grant/revoke functionality at database, table level. - Ranger provides a centralized authorization interface for Hive and provides more granular - access control at column level through the Hive plugin. - </description> - <display-name>Choose Authorization</display-name> - <filename>hive-env.xml</filename> - <deleted>false</deleted> - <on-ambari-upgrade add="true" delete="false" update="false"/> - <property-type></property-type> - <value-attributes> - <type>value-list</type> - <entries> - <entry> - <value>None</value> - <label>None</label> - </entry> - <entry> - <value>SQLStdAuth</value> - <label>SQLStdAuth</label> - </entry> - <entry> - <value>Ranger</value> - <label>Ranger</label> - </entry> - </entries> - </value-attributes> - <depends-on> - <property> - <name>ranger-hive-plugin-enabled</name> - <type>ranger-env</type> - </property> - </depends-on> - <property_depended_by> - <dependedByProperties> - <name>hive.server2.enable.doAs</name> - <type>hive-site</type> - </dependedByProperties> - <dependedByProperties> - <name>hive.security.authenticator.manager</name> - <type>hiveserver2-site</type> - </dependedByProperties> - <dependedByProperties> - <name>hive.server2.enable.doAs</name> - <type>hive-interactive-site</type> - </dependedByProperties> - <dependedByProperties> - <name>hive.security.authorization.enabled</name> - <type>hive-site</type> - </dependedByProperties> - <dependedByProperties> - <name>hive.conf.restricted.list</name> - <type>hiveserver2-site</type> - </dependedByProperties> - <dependedByProperties> - <name>hive.security.authenticator.manager</name> - <type>hive-site</type> - </dependedByProperties> - <dependedByProperties> - <name>hive.security.authorization.manager</name> - <type>hiveserver2-site</type> - </dependedByProperties> - <dependedByProperties> - <name>hive.security.authorization.manager</name> - <type>hive-site</type> - </dependedByProperties> - <dependedByProperties> - <name>hive.security.authorization.enabled</name> - <type>hiveserver2-site</type> - </dependedByProperties> - <dependedByProperties> - <name>hive.security.metastore.authorization.manager</name> - <type>hive-site</type> - </dependedByProperties> - </property_depended_by> -</property><property require-input="false"> - <name>hive_timeline_logging_enabled</name> - <value>true</value> - <display-name>Use ATS Logging</display-name> - <filename>hive-env.xml</filename> - <deleted>false</deleted> - <on-ambari-upgrade add="true" delete="false" update="false"/> - <property-type></property-type> - <value-attributes> - <type>value-list</type> - <entries> - <entry> - <value>true</value> - <label>True</label> - </entry> - <entry> - <value>false</value> - <label>False</label> - </entry> - </entries> - <selection-cardinality>1</selection-cardinality> - </value-attributes> - <depends-on/> - <property_depended_by> - <dependedByProperties> - <name>hive.exec.pre.hooks</name> - <type>hive-site</type> - </dependedByProperties> - <dependedByProperties> - <name>hive.exec.post.hooks</name> - <type>hive-site</type> - </dependedByProperties> - <dependedByProperties> - <name>hive.exec.failure.hooks</name> - <type>hive-site</type> - </dependedByProperties> - </property_depended_by> -</property><property require-input="false"> - <name>hive.client.heapsize</name> - <value>512</value> - <description>Hive Client Java heap size</description> - <display-name>Client Heap Size</display-name> - <filename>hive-env.xml</filename> - <deleted>false</deleted> - <on-ambari-upgrade add="true" delete="false" update="false"/> - <property-type></property-type> - <value-attributes> - <type>int</type> - <maximum>2048</maximum> - <minimum>512</minimum> - <unit>MB</unit> - <overridable>false</overridable> - <increment-step>512</increment-step> - </value-attributes> - <depends-on/> - <property_depended_by/> -</property><property require-input="false"> - <name>hive.metastore.heapsize</name> - <value>1024</value> - <description>Hive Metastore Java heap size</description> - <display-name>Metastore Heap Size</display-name> - <filename>hive-env.xml</filename> - <deleted>false</deleted> - <on-ambari-upgrade add="true" delete="false" update="false"/> - <property-type></property-type> - <value-attributes> - <type>int</type> - <maximum>2048</maximum> - <minimum>512</minimum> - <unit>MB</unit> - <increment-step>512</increment-step> - </value-attributes> - <depends-on/> - <property_depended_by/> -</property><property require-input="false"> - <name>hive_database_type</name> - <value>mysql</value> - <description>Default HIVE DB type.</description> - <display-name>Hive Database Type</display-name> - <filename>hive-env.xml</filename> - <deleted>false</deleted> - <on-ambari-upgrade add="true" delete="false" update="false"/> - <property-type></property-type> - <value-attributes> - <overridable>false</overridable> - </value-attributes> - <depends-on> - <property> - <name>hive_database</name> - <type>hive-env</type> - </property> - </depends-on> - <property_depended_by/> -</property><property require-input="false"> - <name>hive_database</name> - <value>New MySQL Database</value> - <description> - Property that determines whether the HIVE DB is managed by Ambari. - </description> - <display-name>Hive Database</display-name> - <filename>hive-env.xml</filename> - <deleted>false</deleted> - <on-ambari-upgrade add="true" delete="false" update="false"/> - <property-type></property-type> - <value-attributes> - <overridable>false</overridable> - </value-attributes> - <depends-on/> - <property_depended_by> - <dependedByProperties> - <name>hive_database_type</name> - <type>hive-env</type> - </dependedByProperties> - <dependedByProperties> - <name>javax.jdo.option.ConnectionURL</name> - <type>hive-site</type> - </dependedByProperties> - <dependedByProperties> - <name>javax.jdo.option.ConnectionDriverName</name> - <type>hive-site</type> - </dependedByProperties> - <dependedByProperties> - <name>datanucleus.rdbms.datastoreAdapterClassName</name> - <type>hive-site</type> - </dependedByProperties> - </property_depended_by> -</property><property require-input="false"> - <name>hive_ambari_database</name> - <value>MySQL</value> - <description>Database type.</description> - <filename>hive-env.xml</filename> - <deleted>false</deleted> - <on-ambari-upgrade add="true" delete="false" update="false"/> - <property-type></property-type> - <value-attributes/> - <depends-on/> - <property_depended_by/> -</property><property require-input="false"> - <name>hive_database_name</name> - <value>hive</value> - <description>Database name.</description> - <filename>hive-env.xml</filename> - <deleted>false</deleted> - <on-ambari-upgrade add="true" delete="false" update="false"/> - <property-type></property-type> - <value-attributes> - <type>database</type> - <visible>false</visible> - <overridable>false</overridable> - <editable-only-at-install>true</editable-only-at-install> - </value-attributes> - <depends-on/> - <property_depended_by/> -</property><property require-input="false"> - <name>hive_log_dir</name> - <value>/var/log/hive</value> - <description>Directory for Hive Log files.</description> - <display-name>Hive Log Dir</display-name> - <filename>hive-env.xml</filename> - <deleted>false</deleted> - <on-ambari-upgrade add="true" delete="false" update="false"/> - <property-type></property-type> - <value-attributes> - <type>directory</type> - <overridable>false</overridable> - </value-attributes> - <depends-on/> - <property_depended_by/> -</property><property require-input="false"> - <name>hive_pid_dir</name> - <value>/var/run/hive</value> - <description>Hive PID Dir.</description> - <display-name>Hive PID Dir</display-name> - <filename>hive-env.xml</filename> - <deleted>false</deleted> - <on-ambari-upgrade add="true" delete="false" update="false"/> - <property-type></property-type> - <value-attributes> - <type>directory</type> - <overridable>false</overridable> - <editable-only-at-install>true</editable-only-at-install> - </value-attributes> - <depends-on/> - <property_depended_by/> -</property><property require-input="false"> - <name>hive_user</name> - <value>hive</value> - <description>Hive User.</description> - <display-name>Hive User</display-name> - <filename>hive-env.xml</filename> - <deleted>false</deleted> - <on-ambari-upgrade add="true" delete="false" update="false"/> - <property-type>USER</property-type> - <value-attributes> - <type>user</type> - <overridable>false</overridable> - </value-attributes> - <depends-on/> - <property_depended_by> - <dependedByProperties> - <name>ranger.plugins.hive.serviceuser</name> - <type>ranger-admin-site</type> - </dependedByProperties> - <dependedByProperties> - <name>ranger.kms.service.user.hive</name> - <type>ranger-admin-site</type> - </dependedByProperties> - </property_depended_by> -</property><property require-input="false"> - <name>hcat_log_dir</name> - <value>/var/log/webhcat</value> - <description>WebHCat Log Dir.</description> - <display-name>WebHCat Log Dir</display-name> - <filename>hive-env.xml</filename> - <deleted>false</deleted> - <on-ambari-upgrade add="true" delete="false" update="false"/> - <property-type></property-type> - <value-attributes> - <type>directory</type> - <overridable>false</overridable> - </value-attributes> - <depends-on/> - <property_depended_by/> -</property><property require-input="false"> - <name>hcat_pid_dir</name> - <value>/var/run/webhcat</value> - <description>WebHCat Pid Dir.</description> - <display-name>WebHCat Pid Dir</display-name> - <filename>hive-env.xml</filename> - <deleted>false</deleted> - <on-ambari-upgrade add="true" delete="false" update="false"/> - <property-type></property-type> - <value-attributes> - <type>directory</type> - <overridable>false</overridable> - <editable-only-at-install>true</editable-only-at-install> - </value-attributes> - <depends-on/> - <property_depended_by/> -</property><property require-input="false"> - <name>hcat_user</name> - <value>hcat</value> - <description>HCat User.</description> - <display-name>HCat User</display-name> - <filename>hive-env.xml</filename> - <deleted>false</deleted> - <on-ambari-upgrade add="true" delete="false" update="false"/> - <property-type>USER</property-type> - <value-attributes> - <type>user</type> - <overridable>false</overridable> - </value-attributes> - <depends-on/> - <property_depended_by/> -</property><property require-input="false"> - <name>webhcat_user</name> - <value>hcat</value> - <description>WebHCat User.</description> - <display-name>WebHCat User</display-name> - <filename>hive-env.xml</filename> - <deleted>false</deleted> - <on-ambari-upgrade add="true" delete="false" update="false"/> - <property-type>USER</property-type> - <value-attributes> - <type>user</type> - <overridable>false</overridable> - </value-attributes> - <depends-on/> - <property_depended_by/> -</property><property require-input="false"> - <name>hive_user_nofile_limit</name> - <value>32000</value> - <description>Max open files limit setting for HIVE user.</description> - <filename>hive-env.xml</filename> - <deleted>false</deleted> - <on-ambari-upgrade add="true" delete="false" update="false"/> - <property-type></property-type> - <value-attributes/> - <depends-on/> - <property_depended_by/> -</property><property require-input="false"> - <name>hive_user_nproc_limit</name> - <value>16000</value> - <description>Max number of processes limit setting for HIVE user.</description> - <filename>hive-env.xml</filename> - <deleted>false</deleted> - <on-ambari-upgrade add="true" delete="false" update="false"/> - <property-type></property-type> - <value-attributes/> - <depends-on/> - <property_depended_by/> -</property></configuration> \ No newline at end of file
