http://git-wip-us.apache.org/repos/asf/ambari/blob/6919aa50/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration-mapred/mapred-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration-mapred/mapred-site.xml b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration-mapred/mapred-site.xml index 7955cb2..2655e51 100644 --- a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration-mapred/mapred-site.xml +++ b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration-mapred/mapred-site.xml @@ -1,6 +1,5 @@ <?xml version="1.0"?> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> - <!-- Licensed to the Apache Software Foundation (ASF) under one or more contributor license agreements. See the NOTICE file distributed with @@ -17,13 +16,9 @@ See the License for the specific language governing permissions and limitations under the License. --> - <!-- Put site-specific property overrides in this file. --> - -<configuration supports_final="true" xmlns:xi="http://www.w3.org/2001/XInclude"> - -<!-- i/o properties --> - +<configuration xmlns:xi="http://www.w3.org/2001/XInclude" supports_final="true"> + <!-- i/o properties --> <property> <name>mapreduce.task.io.sort.mb</name> <value>200</value> @@ -31,8 +26,9 @@ The total amount of buffer memory to use while sorting files, in megabytes. By default, gives each merge stream 1MB, which should minimize seeks. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.map.sort.spill.percent</name> <value>0.7</value> @@ -43,8 +39,9 @@ is already in progress, so spills may be larger than this threshold when it is set to less than .5 </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.task.io.sort.factor</name> <value>100</value> @@ -52,17 +49,19 @@ The number of streams to merge at once while sorting files. This determines the number of open file handles. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - -<!-- map/reduce properties --> + <!-- map/reduce properties --> <property> <name>mapreduce.cluster.administrators</name> <value> hadoop</value> <description> Administrators for MapReduce applications. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.reduce.shuffle.parallelcopies</name> <value>30</value> @@ -70,8 +69,9 @@ The default number of parallel transfers run by reduce during the copy(shuffle) phase. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.map.speculative</name> <value>false</value> @@ -79,8 +79,9 @@ If true, then multiple instances of some map tasks may be executed in parallel. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.reduce.speculative</name> <value>false</value> @@ -88,8 +89,9 @@ If true, then multiple instances of some reduce tasks may be executed in parallel. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.job.reduce.slowstart.completedmaps</name> <value>0.05</value> @@ -97,8 +99,9 @@ Fraction of the number of maps in the job which should be complete before reduces are scheduled for the job. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.reduce.shuffle.merge.percent</name> <value>0.66</value> @@ -108,8 +111,9 @@ storing in-memory map outputs, as defined by mapreduce.reduce.shuffle.input.buffer.percent. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.reduce.shuffle.input.buffer.percent</name> <value>0.7</value> @@ -117,8 +121,9 @@ The percentage of memory to be allocated from the maximum heap size to storing map outputs during the shuffle. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.output.fileoutputformat.compress.type</name> <value>BLOCK</value> @@ -126,8 +131,9 @@ If the job outputs are to compressed as SequenceFiles, how should they be compressed? Should be one of NONE, RECORD or BLOCK. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.reduce.input.buffer.percent</name> <value>0.0</value> @@ -137,8 +143,9 @@ remaining map outputs in memory must consume less than this threshold before the reduce can begin. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <!-- copied from kryptonite configuration --> <property> <name>mapreduce.map.output.compress</name> @@ -146,8 +153,9 @@ <description> Should the outputs of the maps be compressed before being sent across the network. Uses SequenceFile compression. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.task.timeout</name> <value>300000</value> @@ -156,20 +164,23 @@ terminated if it neither reads an input, writes an output, nor updates its status string. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.map.memory.mb</name> <value>1024</value> <description>Virtual memory for single Map task</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.reduce.memory.mb</name> <value>1024</value> <description>Virtual memory for single Reduce task</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.shuffle.port</name> <value>13562</value> @@ -178,36 +189,41 @@ ShuffleHandler is a service run at the NodeManager to facilitate transfers of intermediate Map outputs to requesting Reducers. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.jobhistory.intermediate-done-dir</name> <value>/mr-history/tmp</value> <description> Directory where history files are written by MapReduce jobs. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.jobhistory.done-dir</name> <value>/mr-history/done</value> <description> Directory where history files are managed by the MR JobHistory Server. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - - <property> + <property> <name>mapreduce.jobhistory.address</name> <value>localhost:10020</value> <description>Enter your JobHistoryServer hostname.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - - <property> + <property> <name>mapreduce.jobhistory.webapp.address</name> <value>localhost:19888</value> <description>Enter your JobHistoryServer hostname.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.framework.name</name> <value>yarn</value> @@ -215,22 +231,25 @@ The runtime framework for executing MapReduce jobs. Can be one of local, classic or yarn. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.app.mapreduce.am.staging-dir</name> <value>/user</value> <description> The staging dir used while submitting jobs. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.app.mapreduce.am.resource.mb</name> <value>512</value> <description>The amount of memory the MR AppMaster needs.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.app.mapreduce.am.command-opts</name> <value>-Xmx312m</value> @@ -247,8 +266,9 @@ of LD_LIBRARY_PATH in the map / reduce JVM env using the mapreduce.map.env and mapreduce.reduce.env config settings. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.app.mapreduce.am.admin-command-opts</name> <value>-server -XX:NewRatio=8 -Djava.net.preferIPv4Stack=true -Dhadoop.metrics.log.level=WARN</value> @@ -262,26 +282,30 @@ of LD_LIBRARY_PATH in the map / reduce JVM env using the mapreduce.map.env and mapreduce.reduce.env config settings. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.app.mapreduce.am.log.level</name> <value>INFO</value> <description>MR App Master process log level.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.admin.map.child.java.opts</name> <value>-server -XX:NewRatio=8 -Djava.net.preferIPv4Stack=true -Dhadoop.metrics.log.level=WARN</value> <description>This property stores Java options for map tasks.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.admin.reduce.child.java.opts</name> <value>-server -XX:NewRatio=8 -Djava.net.preferIPv4Stack=true -Dhadoop.metrics.log.level=WARN</value> <description>This property stores Java options for reduce tasks.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.application.classpath</name> <value>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/*,$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/lib/*</value> @@ -289,8 +313,9 @@ CLASSPATH for MR applications. A comma-separated list of CLASSPATH entries. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.am.max-attempts</name> <value>2</value> @@ -300,27 +325,27 @@ set by resourcemanager. Otherwise, it will be override. The default number is set to 2, to allow at least one retry for AM. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - - - <property> <name>mapreduce.map.java.opts</name> <value>-Xmx756m</value> <description> Larger heap-size for child jvms of maps. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - - <property> <name>mapreduce.reduce.java.opts</name> <value>-Xmx756m</value> <description> Larger heap-size for child jvms of reduces. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.map.log.level</name> <value>INFO</value> @@ -328,8 +353,9 @@ The logging level for the map task. The allowed levels are: OFF, FATAL, ERROR, WARN, INFO, DEBUG, TRACE and ALL. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.reduce.log.level</name> <value>INFO</value> @@ -337,8 +363,9 @@ The logging level for the reduce task. The allowed levels are: OFF, FATAL, ERROR, WARN, INFO, DEBUG, TRACE and ALL. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.admin.user.env</name> <value>LD_LIBRARY_PATH=/usr/lib/hadoop/lib/native:/usr/lib/hadoop/lib/native/Linux-amd64-64</value> @@ -347,14 +374,16 @@ This is not an additive property. You must preserve the original value if you want your map and reduce tasks to have access to native libraries (compression, etc) </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>mapreduce.output.fileoutputformat.compress</name> <value>false</value> <description> Should the job outputs be compressed? </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - </configuration>
http://git-wip-us.apache.org/repos/asf/ambari/blob/6919aa50/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration/capacity-scheduler.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration/capacity-scheduler.xml b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration/capacity-scheduler.xml index 6c21848..80d0e91 100644 --- a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration/capacity-scheduler.xml +++ b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration/capacity-scheduler.xml @@ -1,3 +1,4 @@ +<?xml version="1.0"?> <!-- Licensed to the Apache Software Foundation (ASF) under one or more contributor license agreements. See the NOTICE file distributed with @@ -14,17 +15,16 @@ See the License for the specific language governing permissions and limitations under the License. --> - <configuration supports_final="false" supports_adding_forbidden="true"> - <property> <name>yarn.scheduler.capacity.maximum-applications</name> <value>10000</value> <description> Maximum number of applications that can be pending and running. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.scheduler.capacity.maximum-am-resource-percent</name> <value>0.2</value> @@ -33,16 +33,18 @@ application masters i.e. controls number of concurrent running applications. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.scheduler.capacity.root.queues</name> <value>default</value> <description> The queues at the this level (root is the root queue). </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.scheduler.capacity.root.capacity</name> <value>100</value> @@ -52,54 +54,61 @@ The child queues capacity should add up to their parent queue's capacity or less. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.scheduler.capacity.root.default.capacity</name> <value>100</value> <description>Default queue target capacity.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.scheduler.capacity.root.default.user-limit-factor</name> <value>1</value> <description> Default queue user limit a percentage from 0.0 to 1.0. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.scheduler.capacity.root.default.maximum-capacity</name> <value>100</value> <description> The maximum capacity of the default queue. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.scheduler.capacity.root.default.state</name> <value>RUNNING</value> <description> The state of the default queue. State can be one of RUNNING or STOPPED. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.scheduler.capacity.root.default.acl_submit_applications</name> <value>*</value> <description> The ACL of who can submit jobs to the default queue. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.scheduler.capacity.root.default.acl_administer_jobs</name> <value>*</value> <description> The ACL of who can administer jobs on the default queue. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.scheduler.capacity.root.acl_administer_queue</name> <value>*</value> @@ -107,8 +116,9 @@ The ACL for who can administer this queue i.e. change sub-queue allocations. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.scheduler.capacity.node-locality-delay</name> <value>40</value> @@ -118,15 +128,16 @@ Typically this should be set to number of nodes in the cluster, By default is setting approximately number of nodes in one rack which is 40. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.scheduler.capacity.default.minimum-user-limit-percent</name> <value>100</value> <description> Default minimum queue resource limit depends on the number of users who have submitted applications. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - - </configuration> http://git-wip-us.apache.org/repos/asf/ambari/blob/6919aa50/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration/yarn-env.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration/yarn-env.xml b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration/yarn-env.xml index eea221e..c595f3a 100644 --- a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration/yarn-env.xml +++ b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration/yarn-env.xml @@ -19,17 +19,20 @@ * limitations under the License. */ --> - <configuration supports_adding_forbidden="true"> <property> <name>yarn_log_dir_prefix</name> <value>/var/log/hadoop-yarn</value> <description>YARN Log Dir Prefix</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>yarn_pid_dir_prefix</name> <value>/var/run/hadoop-yarn</value> <description>YARN PID Dir Prefix</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>yarn_user</name> @@ -41,33 +44,44 @@ <type>user</type> <overridable>false</overridable> </value-attributes> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>yarn_heapsize</name> <value>1024</value> <description>Max heapsize for all YARN components using a numerical value in the scale of MB</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>resourcemanager_heapsize</name> <value>1024</value> <description>Max heapsize for ResourceManager using a numerical value in the scale of MB</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>nodemanager_heapsize</name> <value>1024</value> <description>Max heapsize for NodeManager using a numerical value in the scale of MB</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>min_user_id</name> <value>1000</value> <description>Set to 0 to disallow root from submitting jobs. Set to 1000 to disallow all superusers from submitting jobs</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>apptimelineserver_heapsize</name> <value>1024</value> <description>Max heapsize for AppTimelineServer using a numerical value in the scale of MB</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <!-- yarn-env.sh --> <property> <name>content</name> @@ -187,6 +201,7 @@ YARN_OPTS="$YARN_OPTS -Dyarn.policy.file=$YARN_POLICYFILE" <value-attributes> <type>content</type> </value-attributes> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - </configuration> http://git-wip-us.apache.org/repos/asf/ambari/blob/6919aa50/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration/yarn-log4j.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration/yarn-log4j.xml b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration/yarn-log4j.xml index bbbe5f5..cece269 100644 --- a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration/yarn-log4j.xml +++ b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration/yarn-log4j.xml @@ -19,9 +19,7 @@ * limitations under the License. */ --> - <configuration supports_final="false" supports_adding_forbidden="true"> - <property> <name>content</name> <description>Custom log4j.properties</description> @@ -69,6 +67,7 @@ log4j.additivity.org.apache.hadoop.yarn.server.resourcemanager.RMAppManager$Appl <value-attributes> <type>content</type> </value-attributes> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - </configuration> http://git-wip-us.apache.org/repos/asf/ambari/blob/6919aa50/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration/yarn-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration/yarn-site.xml b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration/yarn-site.xml index ec29f28..fc6708f 100644 --- a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration/yarn-site.xml +++ b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/YARN/configuration/yarn-site.xml @@ -16,31 +16,30 @@ See the License for the specific language governing permissions and limitations under the License. --> - <!-- Put site-specific property overrides in this file. --> - -<configuration supports_final="true" xmlns:xi="http://www.w3.org/2001/XInclude"> - +<configuration xmlns:xi="http://www.w3.org/2001/XInclude" supports_final="true"> <!-- ResourceManager --> - <property> <name>yarn.resourcemanager.hostname</name> <value>localhost</value> <description>The hostname of the RM.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.resourcemanager.resource-tracker.address</name> <value>localhost:8025</value> <description> The address of ResourceManager. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.resourcemanager.scheduler.address</name> <value>localhost:8030</value> <description>The address of the scheduler interface.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.resourcemanager.address</name> <value>localhost:8050</value> @@ -48,20 +47,23 @@ The address of the applications manager interface in the RM. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.resourcemanager.admin.address</name> <value>localhost:8141</value> <description>The address of the RM admin interface.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.resourcemanager.scheduler.class</name> <value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler</value> <description>The class to use as the resource scheduler.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.scheduler.minimum-allocation-mb</name> <value>512</value> @@ -70,8 +72,9 @@ in MBs. Memory requests lower than this won't take effect, and the specified value will get allocated at minimum. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.scheduler.maximum-allocation-mb</name> <value>2048</value> @@ -80,41 +83,46 @@ in MBs. Memory requests higher than this won't take effect, and will get capped to this value. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.acl.enable</name> <value>false</value> <description> Are acls enabled. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.admin.acl</name> - <value></value> + <value/> <description> ACL of who can be admin of the YARN cluster. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <!-- NodeManager --> - <property> <name>yarn.nodemanager.address</name> <value>0.0.0.0:45454</value> <description>The address of the container manager in the NM.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.nodemanager.resource.memory-mb</name> <value>5120</value> <description>Amount of physical memory, in MB, that can be allocated for containers.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.application.classpath</name> <value>/etc/hadoop/conf,/usr/lib/hadoop/*,/usr/lib/hadoop/lib/*,/usr/lib/hadoop-hdfs/*,/usr/lib/hadoop-hdfs/lib/*,/usr/lib/hadoop-yarn/*,/usr/lib/hadoop-yarn/lib/*,/usr/lib/hadoop-mapreduce/*,/usr/lib/hadoop-mapreduce/lib/*</value> <description>Classpath for typical applications.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.nodemanager.vmem-pmem-ratio</name> <value>2.1</value> @@ -123,33 +131,38 @@ expressed in terms of physical memory, and virtual memory usage is allowed to exceed this allocation by this ratio. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.nodemanager.container-executor.class</name> <value>org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor</value> <description>ContainerExecutor for launching containers</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.nodemanager.linux-container-executor.group</name> <value>hadoop</value> <description>Unix group of the NodeManager</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.nodemanager.aux-services</name> <value>mapreduce_shuffle</value> <description>Auxilliary services of NodeManager. A valid service name should only contain a-zA-Z0-9_ and can not start with numbers</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.nodemanager.aux-services.mapreduce_shuffle.class</name> <value>org.apache.hadoop.mapred.ShuffleHandler</value> <description>The auxiliary service class to use </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.nodemanager.log-dirs</name> <value>/hadoop/yarn/log</value> @@ -160,8 +173,9 @@ named container_{$contid}. Each container directory will contain the files stderr, stdin, and syslog generated by that container. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.nodemanager.local-dirs</name> <value>/hadoop/yarn/local</value> @@ -172,8 +186,9 @@ Individual containers' work directories, called container_${contid}, will be subdirectories of this. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.nodemanager.container-monitor.interval-ms</name> <value>3000</value> @@ -181,8 +196,9 @@ The interval, in milliseconds, for which the node manager waits between two cycles of monitoring its containers' memory usage. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <!-- <property> <name>yarn.nodemanager.health-checker.script.path</name> @@ -190,19 +206,20 @@ <description>The health check script to run.</description> </property> --> - <property> <name>yarn.nodemanager.health-checker.interval-ms</name> <value>135000</value> <description>Frequency of running node health script.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.nodemanager.health-checker.script.timeout-ms</name> <value>60000</value> <description>Script time out period.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.nodemanager.log.retain-second</name> <value>604800</value> @@ -210,20 +227,23 @@ Time in seconds to retain user logs. Only applicable if log aggregation is disabled. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.log-aggregation-enable</name> <value>true</value> <description>Whether to enable log aggregation. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.nodemanager.remote-app-log-dir</name> <value>/app-logs</value> <description>Location to aggregate logs to. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.nodemanager.remote-app-log-dir-suffix</name> <value>logs</value> @@ -231,16 +251,18 @@ The remote log dir will be created at {yarn.nodemanager.remote-app-log-dir}/${user}/{thisParam}. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.nodemanager.log-aggregation.compression-type</name> <value>gz</value> <description> T-file compression types used to compress aggregated logs. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.nodemanager.delete.debug-delay-sec</name> <value>0</value> @@ -259,8 +281,9 @@ of the Yarn applications' log directories is configurable with the yarn.nodemanager.log-dirs property (see also below). </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.log-aggregation.retain-seconds</name> <value>2592000</value> @@ -268,8 +291,9 @@ How long to keep aggregation logs before deleting them. -1 disables. Be careful set this too small and you will spam the name node. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.nodemanager.admin-env</name> <value>MALLOC_ARENA_MAX=$MALLOC_ARENA_MAX</value> @@ -277,8 +301,9 @@ Environment variables that should be forwarded from the NodeManager's environment to the container's. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.nodemanager.disk-health-checker.min-healthy-disks</name> <value>0.25</value> @@ -289,8 +314,9 @@ If there are less number of healthy local-dirs (or log-dirs) available, then new containers will not be launched on this node. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.resourcemanager.am.max-attempts</name> <value>2</value> @@ -302,32 +328,36 @@ the resourcemanager will override it. The default number is set to 2, to allow at least one retry for AM. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.resourcemanager.webapp.address</name> <value>localhost:8088</value> <description> The address of the RM web application. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.nodemanager.vmem-check-enabled</name> <value>false</value> <description> Whether virtual memory limits will be enforced for containers. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.log.server.url</name> <value>http://localhost:19888/jobhistory/logs</value> <description> URI for the HistoryServer's log resource </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.resourcemanager.nodes.exclude-path</name> <value>/etc/hadoop/conf/yarn.exclude</value> @@ -337,8 +367,9 @@ file must be specified. If the value is empty, no hosts are excluded. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.timeline-service.enabled</name> <value>true</value> @@ -348,24 +379,27 @@ <value-attributes> <type>boolean</type> </value-attributes> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.timeline-service.store-class</name> <value>org.apache.hadoop.yarn.server.applicationhistoryservice.timeline.LeveldbTimelineStore</value> <description> Store class name for timeline store </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.timeline-service.generic-application-history.store-class</name> <value>org.apache.hadoop.yarn.server.applicationhistoryservice.NullApplicationHistoryStore</value> <description> Store class name for history store, defaulting to file system store </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.timeline-service.leveldb-timeline-store.path</name> <value>/var/log/hadoop-yarn/timeline</value> @@ -375,24 +409,27 @@ <value-attributes> <type>directory</type> </value-attributes> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.timeline-service.webapp.address</name> <value>0.0.0.0:8188</value> <description> The http address of the timeline service web application. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.timeline-service.webapp.https.address</name> <value>0.0.0.0:8190</value> <description> The http address of the timeline service web application. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>yarn.timeline-service.address</name> <value>0.0.0.0:10200</value> @@ -400,6 +437,8 @@ This is default address for the timeline server to start the RPC server. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <description>Enable age off of timeline store data.</description> @@ -408,6 +447,8 @@ <value-attributes> <type>boolean</type> </value-attributes> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <description>Time to live for timeline store data in milliseconds.</description> @@ -416,6 +457,8 @@ <value-attributes> <type>int</type> </value-attributes> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <description>Length of time to wait between deletion cycles of leveldb timeline store in milliseconds.</description> @@ -424,5 +467,7 @@ <value-attributes> <type>int</type> </value-attributes> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> </configuration> http://git-wip-us.apache.org/repos/asf/ambari/blob/6919aa50/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/ZOOKEEPER/configuration/zookeeper-env.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/ZOOKEEPER/configuration/zookeeper-env.xml b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/ZOOKEEPER/configuration/zookeeper-env.xml index 74e6cae..30f32ac 100644 --- a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/ZOOKEEPER/configuration/zookeeper-env.xml +++ b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/ZOOKEEPER/configuration/zookeeper-env.xml @@ -19,7 +19,6 @@ * limitations under the License. */ --> - <configuration supports_adding_forbidden="true"> <property> <name>zk_user</name> @@ -31,43 +30,58 @@ <type>user</type> <overridable>false</overridable> </value-attributes> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>zk_data_dir</name> <value>/hadoop/zookeeper</value> <description>Data directory for ZooKeeper.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>zk_log_dir</name> <value>/var/log/zookeeper</value> <description>ZooKeeper Log Dir</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>zk_pid_dir</name> <value>/var/run/zookeeper</value> <description>ZooKeeper Pid Dir</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>tickTime</name> <value>2000</value> <description>The length of a single tick in milliseconds, which is the basic time unit used by ZooKeeper</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>initLimit</name> <value>10</value> <description>Ticks to allow for sync at Init.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>syncLimit</name> <value>5</value> <description>Ticks to allow for sync at Runtime.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>clientPort</name> <value>2181</value> <description>Port for running ZK Server.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <!-- zookeeper-env.sh --> <property> <name>content</name> @@ -89,5 +103,7 @@ export CLIENT_JVMFLAGS="$CLIENT_JVMFLAGS -Djava.security.auth.login.config={{zk_ <value-attributes> <type>content</type> </value-attributes> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> </configuration> http://git-wip-us.apache.org/repos/asf/ambari/blob/6919aa50/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/ZOOKEEPER/configuration/zookeeper-log4j.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/ZOOKEEPER/configuration/zookeeper-log4j.xml b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/ZOOKEEPER/configuration/zookeeper-log4j.xml index 4ed8a01..1f5031f 100644 --- a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/ZOOKEEPER/configuration/zookeeper-log4j.xml +++ b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/ZOOKEEPER/configuration/zookeeper-log4j.xml @@ -19,9 +19,7 @@ * limitations under the License. */ --> - <configuration supports_final="false" supports_adding_forbidden="true"> - <property> <name>content</name> <description>Custom log4j.properties</description> @@ -99,6 +97,7 @@ log4j.appender.TRACEFILE.layout.ConversionPattern=%d{ISO8601} - %-5p [%t:%C{1}@% <value-attributes> <type>content</type> </value-attributes> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - </configuration> http://git-wip-us.apache.org/repos/asf/ambari/blob/6919aa50/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/GLUSTERFS/configuration/core-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/GLUSTERFS/configuration/core-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/GLUSTERFS/configuration/core-site.xml index 00dd4b3..b06ea45 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/GLUSTERFS/configuration/core-site.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/GLUSTERFS/configuration/core-site.xml @@ -1,7 +1,6 @@ <?xml version="1.0"?> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> - - <!-- +<!-- Licensed to the Apache Software Foundation (ASF) under one or more contributor license agreements. See the NOTICE file distributed with this work for additional information regarding copyright ownership. @@ -17,37 +16,39 @@ See the License for the specific language governing permissions and limitations under the License. --> - <!-- Put site-specific property overrides in this file. --> - -<configuration supports_final="true" xmlns:xi="http://www.w3.org/2001/XInclude"> - -<!-- file system properties --> - +<configuration xmlns:xi="http://www.w3.org/2001/XInclude" supports_final="true"> + <!-- file system properties --> <property> - <name>fs.AbstractFileSystem.glusterfs.impl</name> - <value>org.apache.hadoop.fs.local.GlusterFs</value> + <name>fs.AbstractFileSystem.glusterfs.impl</name> + <value>org.apache.hadoop.fs.local.GlusterFs</value> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> - <name>fs.glusterfs.impl</name> - <value>org.apache.hadoop.fs.glusterfs.GlusterFileSystem</value> + <name>fs.glusterfs.impl</name> + <value>org.apache.hadoop.fs.glusterfs.GlusterFileSystem</value> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> - <name>fs.glusterfs.volumes</name> + <name>fs.glusterfs.volumes</name> <description>The name of the gluster volume(s) you would like Hadoop to use. Values should be seperated by commas i.e. gv0, gv1</description> - <value>gv0</value> + <value>gv0</value> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> - <name>fs.glusterfs.volume.fuse.gv0</name> - <description>The mount point that corresponds to the fs.glusterfs.volumes value</description> + <name>fs.glusterfs.volume.fuse.gv0</name> + <description>The mount point that corresponds to the fs.glusterfs.volumes value</description> <value>/mnt/gv0</value> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>fs.defaultFS</name> <value>glusterfs:///localhost:8020</value> - </property> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> + </property> </configuration> http://git-wip-us.apache.org/repos/asf/ambari/blob/6919aa50/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/GLUSTERFS/configuration/hadoop-env.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/GLUSTERFS/configuration/hadoop-env.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/GLUSTERFS/configuration/hadoop-env.xml index 9759aa8..d7978c3 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/GLUSTERFS/configuration/hadoop-env.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/GLUSTERFS/configuration/hadoop-env.xml @@ -19,42 +19,55 @@ * limitations under the License. */ --> - <configuration supports_final="false"> <property> <name>hadoop_pid_dir_prefix</name> <value>/var/run/hadoop</value> <description>Hadoop PID Dir Prefix</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> + <property> <name>hadoop_heapsize</name> <value>1024</value> <description>Hadoop maximum Java heap size</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>glusterfs_user</name> <value>root</value> - <description></description> + <description/> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hdfs_log_dir_prefix</name> <value>/var/log/hadoop</value> <description>Hadoop Log Dir Prefix</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>namenode_heapsize</name> <value>1024</value> <description>NameNode Java heap size</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>namenode_host</name> - <value></value> + <value/> <description>NameNode Host.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>snamenode_host</name> - <value></value> + <value/> <description>Secondary NameNode.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>proxyuser_group</name> @@ -64,12 +77,16 @@ <value-attributes> <overridable>false</overridable> </value-attributes> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hdfs_user</name> <display-name>HDFS User</display-name> <value>hdfs</value> <description>User to run HDFS as</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <!-- <property> @@ -78,7 +95,7 @@ <description>Proxy user group.</description> </property> --> - <!-- hadoop-env.sh --> + <!-- hadoop-env.sh --> <property> <name>content</name> <display-name>hadoop-env template</display-name> @@ -172,12 +189,12 @@ export HADOOP_IDENT_STRING=$USER # Use libraries from standard classpath JAVA_JDBC_LIBS="" #Add libraries required by mysql connector -for jarFile in `ls /usr/share/java/*mysql* 2>/dev/null` +for jarFile in `ls /usr/share/java/*mysql* 2>/dev/null` do JAVA_JDBC_LIBS=${JAVA_JDBC_LIBS}:$jarFile done #Add libraries required by oracle connector -for jarFile in `ls /usr/share/java/*ojdbc* 2>/dev/null` +for jarFile in `ls /usr/share/java/*ojdbc* 2>/dev/null` do JAVA_JDBC_LIBS=${JAVA_JDBC_LIBS}:$jarFile done @@ -196,8 +213,9 @@ export HADOOP_LIBEXEC_DIR={{hadoop_libexec_dir}} export JAVA_LIBRARY_PATH=${JAVA_LIBRARY_PATH}:/usr/lib/hadoop/lib/native/Linux-amd64-64 </value> <value-attributes> - <type>content</type> + <type>content</type> </value-attributes> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - </configuration> http://git-wip-us.apache.org/repos/asf/ambari/blob/6919aa50/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/HBASE/configuration/hbase-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/HBASE/configuration/hbase-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/HBASE/configuration/hbase-site.xml index c9a5220..cb447e3 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/HBASE/configuration/hbase-site.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/HBASE/configuration/hbase-site.xml @@ -32,6 +32,8 @@ into /tmp. Change this configuration else all data will be lost on machine restart. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hbase.cluster.distributed</name> @@ -41,11 +43,15 @@ false, startup will run all HBase and ZooKeeper daemons together in the one JVM. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hbase.master.port</name> <value>60000</value> <description>The port the HBase Master should bind to.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hbase.tmp.dir</name> @@ -58,28 +64,38 @@ <value-attributes> <type>directory</type> </value-attributes> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hbase.local.dir</name> <value>${hbase.tmp.dir}/local</value> <description>Directory on the local filesystem to be used as a local storage </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hbase.master.info.bindAddress</name> <value>0.0.0.0</value> <description>The bind address for the HBase Master web UI </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hbase.master.info.port</name> <value>60010</value> <description>The port for the HBase Master web UI.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hbase.regionserver.info.port</name> <value>60030</value> <description>The port for the HBase RegionServer web UI.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hbase.regionserver.global.memstore.upperLimit</name> @@ -87,6 +103,8 @@ <description>Maximum size of all memstores in a region server before new updates are blocked and flushes are forced. Defaults to 40% of heap </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hbase.regionserver.handler.count</name> @@ -95,6 +113,8 @@ Same property is used by the Master for count of master handlers. Default is 10. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hbase.hregion.majorcompaction</name> @@ -103,8 +123,9 @@ HStoreFiles in a region. Default: 1 day. Set to 0 to disable automated major compactions. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>hbase.regionserver.global.memstore.lowerLimit</name> <value>0.38</value> @@ -114,6 +135,8 @@ the minimum possible flushing to occur when updates are blocked due to memstore limiting. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hbase.hregion.memstore.block.multiplier</name> @@ -125,6 +148,8 @@ resultant flush files take a long time to compact or split, or worse, we OOME </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hbase.hregion.memstore.flush.size</name> @@ -134,6 +159,8 @@ exceeds this number of bytes. Value is checked by a thread that runs every hbase.server.thread.wakefrequency. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hbase.hregion.memstore.mslab.enabled</name> @@ -144,6 +171,8 @@ heavy write loads. This can reduce the frequency of stop-the-world GC pauses on large heaps. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hbase.hregion.max.filesize</name> @@ -153,6 +182,8 @@ grown to exceed this value, the hosting HRegion is split in two. Default: 1G. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hbase.client.scanner.caching</name> @@ -164,6 +195,8 @@ Do not set this value such that the time between invocations is greater than the scanner timeout; i.e. hbase.regionserver.lease.period </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>zookeeper.session.timeout</name> @@ -175,6 +208,8 @@ "The client sends a requested timeout, the server responds with the timeout that it can give the client. " In milliseconds. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hbase.client.keyvalue.maxsize</name> @@ -186,6 +221,8 @@ to set this to a fraction of the maximum region size. Setting it to zero or less disables the check. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hbase.hstore.compactionThreshold</name> @@ -196,6 +233,8 @@ is run to rewrite all HStoreFiles files as one. Larger numbers put off compaction but when it runs, it takes longer to complete. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hbase.hstore.flush.retries.number</name> @@ -203,8 +242,9 @@ <description> The number of times the region flush operation will be retried. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>hbase.hstore.blockingStoreFiles</name> <value>10</value> @@ -214,6 +254,8 @@ blocked for this HRegion until a compaction is completed, or until hbase.hstore.blockingWaitTime has been exceeded. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hfile.block.cache.size</name> @@ -223,8 +265,9 @@ used by HFile/StoreFile. Default of 0.25 means allocate 25%. Set to 0 to disable but it's not recommended. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <!-- The following properties configure authentication information for HBase processes when using Kerberos security. There are no default values, included here for documentation purposes --> @@ -234,6 +277,8 @@ <description>Full path to the kerberos keytab file to use for logging in the configured HMaster server principal. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hbase.master.kerberos.principal</name> @@ -244,6 +289,8 @@ portion, it will be replaced with the actual hostname of the running instance. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hbase.regionserver.keytab.file</name> @@ -251,6 +298,8 @@ <description>Full path to the kerberos keytab file to use for logging in the configured HRegionServer server principal. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>hbase.regionserver.kerberos.principal</name> @@ -262,8 +311,9 @@ running instance. An entry for this principal must exist in the file specified in hbase.regionserver.keytab.file </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <!-- Additional configuration specific to HBase security --> <property> <name>hbase.superuser</name> @@ -272,37 +322,41 @@ full privileges, regardless of stored ACLs, across the cluster. Only used when HBase security is enabled. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>hbase.security.authentication</name> <value>simple</value> <description> Controls whether or not secure authentication is enabled for HBase. Possible values are 'simple' (no authentication), and 'kerberos'. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>hbase.security.authorization</name> <value>false</value> <description>Enables HBase authorization. Set the value of this property to false to disable HBase authorization. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>hbase.coprocessor.region.classes</name> - <value></value> + <value/> <description>A comma-separated list of Coprocessors that are loaded by default on all tables. For any override coprocessor method, these classes will be called in order. After implementing your own Coprocessor, just put it in HBase's classpath and add the fully qualified class name here. A coprocessor can also be loaded on demand by setting HTableDescriptor. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>hbase.coprocessor.master.classes</name> - <value></value> + <value/> <description>A comma-separated list of org.apache.hadoop.hbase.coprocessor.MasterObserver coprocessors that are loaded by default on the active HMaster process. For any implemented @@ -310,16 +364,18 @@ implementing your own MasterObserver, just put it in HBase's classpath and add the fully qualified class name here. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>hbase.zookeeper.property.clientPort</name> <value>2181</value> <description>Property from ZooKeeper's config zoo.cfg. The port at which the clients will connect. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <!-- The following three properties are used together to create the list of host:peer_port:leader_port quorum servers for ZooKeeper. @@ -334,19 +390,22 @@ list of ZooKeeper quorum servers. If HBASE_MANAGES_ZK is set in hbase-env.sh this is the list of servers which we will start/stop ZooKeeper on. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <!-- End of properties used to generate ZooKeeper host:port quorum list. --> - <property> <name>hbase.zookeeper.useMulti</name> <value>true</value> <description>Instructs HBase to make use of ZooKeeper's multi-update functionality. This allows certain ZooKeeper operations to complete more quickly and prevents some issues - with rare Replication failure scenarios (see the release note of HBASE-2611 for an example).· + with rare Replication failure scenarios (see the release note of HBASE-2611 for an example).· IMPORTANT: only set this to true if all ZooKeeper servers in the cluster are on version 3.4+ and will not be downgraded. ZooKeeper versions before 3.4 do not support multi-update and will not fail gracefully if multi-update is invoked (see ZOOKEEPER-1495). </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> <property> <name>zookeeper.znode.parent</name> @@ -356,18 +415,21 @@ By default, all of HBase's ZooKeeper file path are configured with a relative path, so they will all go under this directory unless changed. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>hbase.defaults.for.version.skip</name> <value>true</value> <description>Disables version verification.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>dfs.domain.socket.path</name> <value>/var/lib/hadoop-hdfs/dn_socket</value> <description>Path to domain socket.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - </configuration> http://git-wip-us.apache.org/repos/asf/ambari/blob/6919aa50/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/HDFS/configuration/core-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/HDFS/configuration/core-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/HDFS/configuration/core-site.xml index db665bd..3ba05ce 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/HDFS/configuration/core-site.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/HDFS/configuration/core-site.xml @@ -1,7 +1,6 @@ <?xml version="1.0"?> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> - - <!-- +<!-- Licensed to the Apache Software Foundation (ASF) under one or more contributor license agreements. See the NOTICE file distributed with this work for additional information regarding copyright ownership. @@ -17,13 +16,9 @@ See the License for the specific language governing permissions and limitations under the License. --> - <!-- Put site-specific property overrides in this file. --> - -<configuration supports_final="true" xmlns:xi="http://www.w3.org/2001/XInclude"> - -<!-- i/o properties --> - +<configuration xmlns:xi="http://www.w3.org/2001/XInclude" supports_final="true"> + <!-- i/o properties --> <property> <name>io.file.buffer.size</name> <value>131072</value> @@ -31,22 +26,24 @@ The size of this buffer should probably be a multiple of hardware page size (4096 on Intel x86), and it determines how much data is buffered during read and write operations.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>io.serializations</name> <value>org.apache.hadoop.io.serializer.WritableSerialization</value> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>io.compression.codecs</name> <value>org.apache.hadoop.io.compress.GzipCodec,org.apache.hadoop.io.compress.DefaultCodec</value> <description>A list of the compression codec classes that can be used for compression/decompression.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - -<!-- file system properties --> - + <!-- file system properties --> <property> <name>fs.defaultFS</name> <!-- cluster variant --> @@ -54,16 +51,18 @@ <description>The name of the default file system. Either the literal string "local" or a host:port for HDFS.</description> <final>true</final> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>fs.trash.interval</name> <value>360</value> <description>Number of minutes between trash checkpoints. If zero, the trash feature is disabled. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <!-- ipc properties: copied from kryptonite configuration --> <property> <name>ipc.client.idlethreshold</name> @@ -71,22 +70,25 @@ <description>Defines the threshold number of connections after which connections will be inspected for idleness. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>ipc.client.connection.maxidletime</name> <value>30000</value> <description>The maximum time after which a client will bring down the connection to the server. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <property> <name>ipc.client.connect.max.retries</name> <value>50</value> <description>Defines the maximum number of retries for IPC connections.</description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - <!-- Web Interface Configuration --> <property> <name>mapreduce.jobtracker.webinterface.trusted</name> @@ -96,24 +98,28 @@ not be exposed to public. Enable this option if the interfaces are only reachable by those who have the right authorization. </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> - - <property> - <name>hadoop.security.authentication</name> - <value>simple</value> - <description> + <property> + <name>hadoop.security.authentication</name> + <value>simple</value> + <description> Set the authentication for the cluster. Valid values are: simple or kerberos. </description> - </property> -<property> - <name>hadoop.security.authorization</name> - <value>false</value> - <description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> + </property> + <property> + <name>hadoop.security.authorization</name> + <value>false</value> + <description> Enable authorization for different protocols. </description> -</property> - + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> + </property> <property> <name>hadoop.security.auth_to_local</name> <value> @@ -124,7 +130,7 @@ RULE:[2:$1@$0](rs@.*)s/.*/hbase/ DEFAULT </value> -<description>The mapping from kerberos principal names to local OS mapreduce.job.user.names. + <description>The mapping from kerberos principal names to local OS mapreduce.job.user.names. So the default rule is just "DEFAULT" which takes all principals in your default domain to their first component. "[email protected]" and "omalley/[email protected]" to "omalley", if your default domain is APACHE.ORG. The translations rules have 3 sections: @@ -162,5 +168,7 @@ If you want to treat all principals from APACHE.ORG with /admin as "admin", your RULE[2:$1%$2@$0](.%[email protected])s/./admin/ DEFAULT </description> + <on-ambari-upgrade add="false" change="true" delete="true"/> + <on-stack-upgrade add="true" change="true" delete="false"/> </property> </configuration>
