Repository: ambari Updated Branches: refs/heads/branch-2.1 f559d1d50 -> 28428e2f6
AMBARI-13382. Lightweight bugfix to correct typo on configuration folder name & copy configuration folder from common-services to HDP stack so Ambari can find the new config files (Judy Nash via alejandro) Project: http://git-wip-us.apache.org/repos/asf/ambari/repo Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/28428e2f Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/28428e2f Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/28428e2f Branch: refs/heads/branch-2.1 Commit: 28428e2f683f29c82c56a1c333b379402476f5be Parents: f559d1d Author: Alejandro Fernandez <afernan...@hortonworks.com> Authored: Mon Oct 12 16:36:56 2015 -0700 Committer: Alejandro Fernandez <afernan...@hortonworks.com> Committed: Mon Oct 12 16:36:56 2015 -0700 ---------------------------------------------------------------------- .../configurations/spark-hive-site-override.xml | 36 ------ .../configurations/spark-thrift-sparkconf.xml | 125 ------------------- .../configuration/spark-hive-site-override.xml | 36 ++++++ .../configuration/spark-thrift-sparkconf.xml | 125 +++++++++++++++++++ 4 files changed, 161 insertions(+), 161 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/ambari/blob/28428e2f/ambari-server/src/main/resources/common-services/SPARK/1.4.1.2.3/configurations/spark-hive-site-override.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/SPARK/1.4.1.2.3/configurations/spark-hive-site-override.xml b/ambari-server/src/main/resources/common-services/SPARK/1.4.1.2.3/configurations/spark-hive-site-override.xml deleted file mode 100644 index 2de64c5..0000000 --- a/ambari-server/src/main/resources/common-services/SPARK/1.4.1.2.3/configurations/spark-hive-site-override.xml +++ /dev/null @@ -1,36 +0,0 @@ -<?xml version="1.0"?> -<?xml-stylesheet type="text/xsl" href="configuration.xsl"?> -<!-- -Licensed to the Apache Software Foundation (ASF) under one or more -contributor license agreements. See the NOTICE file distributed with -this work for additional information regarding copyright ownership. -The ASF licenses this file to You under the Apache License, Version 2.0 -(the "License"); you may not use this file except in compliance with -the License. You may obtain a copy of the License at - -http://www.apache.org/licenses/LICENSE-2.0 - -Unless required by applicable law or agreed to in writing, software -distributed under the License is distributed on an "AS IS" BASIS, -WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -See the License for the specific language governing permissions and -limitations under the License. ---> - -<configuration supports_final="true"> - <property> - <name>hive.server2.thrift.port</name> - <value>10000</value> - <description> - TCP port number to listen on, default 10000. - </description> - </property> - <property> - <name>hive.server2.transport.mode</name> - <value>binary</value> - <description> - Expects one of [binary, http]. - Transport mode of HiveServer2. - </description> - </property> -</configuration> http://git-wip-us.apache.org/repos/asf/ambari/blob/28428e2f/ambari-server/src/main/resources/common-services/SPARK/1.4.1.2.3/configurations/spark-thrift-sparkconf.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/SPARK/1.4.1.2.3/configurations/spark-thrift-sparkconf.xml b/ambari-server/src/main/resources/common-services/SPARK/1.4.1.2.3/configurations/spark-thrift-sparkconf.xml deleted file mode 100644 index c42841f..0000000 --- a/ambari-server/src/main/resources/common-services/SPARK/1.4.1.2.3/configurations/spark-thrift-sparkconf.xml +++ /dev/null @@ -1,125 +0,0 @@ -<?xml version="1.0" encoding="UTF-8"?> -<!-- -/** - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ ---> - -<configuration supports_final="true"> - - <property> - <name>spark.yarn.executor.memoryOverhead</name> - <value>384</value> - <description> - The amount of off heap memory (in megabytes) to be allocated per executor. - This is memory that accounts for things like VM overheads, interned strings, - other native overheads, etc. - </description> - </property> - - <property> - <name>spark.yarn.driver.memoryOverhead</name> - <value>384</value> - <description> - The amount of off heap memory (in megabytes) to be allocated per driver. - This is memory that accounts for things like VM overheads, interned strings, - other native overheads, etc. - </description> - </property> - - <property> - <name>spark.yarn.applicationMaster.waitTries</name> - <value>10</value> - <description> - Set the number of times the ApplicationMaster waits for the the Spark master and then - also the number of tries it waits for the SparkContext to be initialized. - </description> - </property> - - <property> - <name>spark.yarn.scheduler.heartbeat.interval-ms</name> - <value>5000</value> - <description> - The interval in ms in which the Spark application master heartbeats into the YARN ResourceManager. - </description> - </property> - - <property> - <name>spark.yarn.max.executor.failures</name> - <value>3</value> - <description> - The maximum number of executor failures before failing the application. - </description> - </property> - - <property> - <name>spark.yarn.queue</name> - <value>default</value> - <description> - The name of the YARN queue to which the application is submitted. - </description> - </property> - - <property> - <name>spark.yarn.containerLauncherMaxThreads</name> - <value>25</value> - <description> - The maximum number of threads to use in the application master for launching executor containers. - </description> - </property> - - <property> - <name>spark.yarn.submit.file.replication</name> - <value>3</value> - <description> - HDFS replication level for the files uploaded into HDFS for the application. - These include things like the Spark jar, the app jar, and any distributed cache files/archives. - </description> - </property> - - <property> - <name>spark.yarn.preserve.staging.files</name> - <value>false</value> - <description> - Set to true to preserve the staged files (Spark jar, app jar, distributed cache files) at the - end of the job rather then delete them. - </description> - </property> - - <property> - <name>spark.driver.extraJavaOptions</name> - <value>-Dhdp.version={{hdp_full_version}}</value> - <description> - Specifies parameters that are passed to the JVM of the Spark driver. - </description> - </property> - - <property> - <name>spark.yarn.am.extraJavaOptions</name> - <value>-Dhdp.version={{hdp_full_version}}</value> - <description> - Specifies the parameters that are passed to the JVM of the Spark Application Master. - </description> - </property> - - <property> - <name>spark.yarn.max.executor.failures</name> - <value>3</value> - <description>The maximum number of executor failures before failing the application.</description> - </property> - -</configuration> http://git-wip-us.apache.org/repos/asf/ambari/blob/28428e2f/ambari-server/src/main/resources/stacks/HDP/2.3/services/SPARK/configuration/spark-hive-site-override.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.3/services/SPARK/configuration/spark-hive-site-override.xml b/ambari-server/src/main/resources/stacks/HDP/2.3/services/SPARK/configuration/spark-hive-site-override.xml new file mode 100644 index 0000000..2de64c5 --- /dev/null +++ b/ambari-server/src/main/resources/stacks/HDP/2.3/services/SPARK/configuration/spark-hive-site-override.xml @@ -0,0 +1,36 @@ +<?xml version="1.0"?> +<?xml-stylesheet type="text/xsl" href="configuration.xsl"?> +<!-- +Licensed to the Apache Software Foundation (ASF) under one or more +contributor license agreements. See the NOTICE file distributed with +this work for additional information regarding copyright ownership. +The ASF licenses this file to You under the Apache License, Version 2.0 +(the "License"); you may not use this file except in compliance with +the License. You may obtain a copy of the License at + +http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +--> + +<configuration supports_final="true"> + <property> + <name>hive.server2.thrift.port</name> + <value>10000</value> + <description> + TCP port number to listen on, default 10000. + </description> + </property> + <property> + <name>hive.server2.transport.mode</name> + <value>binary</value> + <description> + Expects one of [binary, http]. + Transport mode of HiveServer2. + </description> + </property> +</configuration> http://git-wip-us.apache.org/repos/asf/ambari/blob/28428e2f/ambari-server/src/main/resources/stacks/HDP/2.3/services/SPARK/configuration/spark-thrift-sparkconf.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.3/services/SPARK/configuration/spark-thrift-sparkconf.xml b/ambari-server/src/main/resources/stacks/HDP/2.3/services/SPARK/configuration/spark-thrift-sparkconf.xml new file mode 100644 index 0000000..c42841f --- /dev/null +++ b/ambari-server/src/main/resources/stacks/HDP/2.3/services/SPARK/configuration/spark-thrift-sparkconf.xml @@ -0,0 +1,125 @@ +<?xml version="1.0" encoding="UTF-8"?> +<!-- +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +--> + +<configuration supports_final="true"> + + <property> + <name>spark.yarn.executor.memoryOverhead</name> + <value>384</value> + <description> + The amount of off heap memory (in megabytes) to be allocated per executor. + This is memory that accounts for things like VM overheads, interned strings, + other native overheads, etc. + </description> + </property> + + <property> + <name>spark.yarn.driver.memoryOverhead</name> + <value>384</value> + <description> + The amount of off heap memory (in megabytes) to be allocated per driver. + This is memory that accounts for things like VM overheads, interned strings, + other native overheads, etc. + </description> + </property> + + <property> + <name>spark.yarn.applicationMaster.waitTries</name> + <value>10</value> + <description> + Set the number of times the ApplicationMaster waits for the the Spark master and then + also the number of tries it waits for the SparkContext to be initialized. + </description> + </property> + + <property> + <name>spark.yarn.scheduler.heartbeat.interval-ms</name> + <value>5000</value> + <description> + The interval in ms in which the Spark application master heartbeats into the YARN ResourceManager. + </description> + </property> + + <property> + <name>spark.yarn.max.executor.failures</name> + <value>3</value> + <description> + The maximum number of executor failures before failing the application. + </description> + </property> + + <property> + <name>spark.yarn.queue</name> + <value>default</value> + <description> + The name of the YARN queue to which the application is submitted. + </description> + </property> + + <property> + <name>spark.yarn.containerLauncherMaxThreads</name> + <value>25</value> + <description> + The maximum number of threads to use in the application master for launching executor containers. + </description> + </property> + + <property> + <name>spark.yarn.submit.file.replication</name> + <value>3</value> + <description> + HDFS replication level for the files uploaded into HDFS for the application. + These include things like the Spark jar, the app jar, and any distributed cache files/archives. + </description> + </property> + + <property> + <name>spark.yarn.preserve.staging.files</name> + <value>false</value> + <description> + Set to true to preserve the staged files (Spark jar, app jar, distributed cache files) at the + end of the job rather then delete them. + </description> + </property> + + <property> + <name>spark.driver.extraJavaOptions</name> + <value>-Dhdp.version={{hdp_full_version}}</value> + <description> + Specifies parameters that are passed to the JVM of the Spark driver. + </description> + </property> + + <property> + <name>spark.yarn.am.extraJavaOptions</name> + <value>-Dhdp.version={{hdp_full_version}}</value> + <description> + Specifies the parameters that are passed to the JVM of the Spark Application Master. + </description> + </property> + + <property> + <name>spark.yarn.max.executor.failures</name> + <value>3</value> + <description>The maximum number of executor failures before failing the application.</description> + </property> + +</configuration>