http://git-wip-us.apache.org/repos/asf/ambari/blob/ffb4d3b8/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/MAPREDUCE2/configuration/mapred-site.xml ---------------------------------------------------------------------- diff --git a/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/MAPREDUCE2/configuration/mapred-site.xml b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/MAPREDUCE2/configuration/mapred-site.xml new file mode 100644 index 0000000..5896f13 --- /dev/null +++ b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/MAPREDUCE2/configuration/mapred-site.xml @@ -0,0 +1,300 @@ +<?xml version="1.0"?> +<?xml-stylesheet type="text/xsl" href="configuration.xsl"?> + +<!-- + Licensed to the Apache Software Foundation (ASF) under one or more + contributor license agreements. See the NOTICE file distributed with + this work for additional information regarding copyright ownership. + The ASF licenses this file to You under the Apache License, Version 2.0 + (the "License"); you may not use this file except in compliance with + the License. You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +--> + +<!-- Put site-specific property overrides in this file. --> + +<configuration xmlns:xi="http://www.w3.org/2001/XInclude"> + +<!-- i/o properties --> + + <property> + <name>io.sort.mb</name> + <value>100</value> + <description>No description</description> + </property> + + <property> + <name>io.sort.record.percent</name> + <value>.2</value> + <description>No description</description> + </property> + + <property> + <name>io.sort.spill.percent</name> + <value>0.1</value> + <description>No description</description> + </property> + + <property> + <name>io.sort.factor</name> + <value>100</value> + <description>No description</description> + </property> + +<!-- map/reduce properties --> + + <property> + <name>mapred.system.dir</name> + <value></value> + <description>No description</description> + <final>true</final> + </property> + + <property> + <!-- cluster specific --> + <name>mapred.local.dir</name> + <value></value> + <description>No description</description> + <final>true</final> + </property> + + <property> + <name>mapred.reduce.parallel.copies</name> + <value>30</value> + <description>No description</description> + </property> + + <property> + <name>mapred.tasktracker.map.tasks.maximum</name> + <value></value> + <description>No description</description> + </property> + + <property> + <name>mapred.map.tasks.speculative.execution</name> + <value>false</value> + <description>If true, then multiple instances of some map tasks + may be executed in parallel.</description> + </property> + + <property> + <name>mapred.reduce.tasks.speculative.execution</name> + <value>false</value> + <description>If true, then multiple instances of some reduce tasks + may be executed in parallel.</description> + </property> + + <property> + <name>mapred.reduce.slowstart.completed.maps</name> + <value>0.05</value> + </property> + + <property> + <name>mapred.inmem.merge.threshold</name> + <value>1000</value> + <description>The threshold, in terms of the number of files + for the in-memory merge process. When we accumulate threshold number of files + we initiate the in-memory merge and spill to disk. A value of 0 or less than + 0 indicates we want to DON'T have any threshold and instead depend only on + the ramfs's memory consumption to trigger the merge. + </description> + </property> + + <property> + <name>mapred.job.shuffle.merge.percent</name> + <value>0.66</value> + <description>The usage threshold at which an in-memory merge will be + initiated, expressed as a percentage of the total memory allocated to + storing in-memory map outputs, as defined by + mapred.job.shuffle.input.buffer.percent. + </description> + </property> + + <property> + <name>mapred.job.shuffle.input.buffer.percent</name> + <value>0.7</value> + <description>The percentage of memory to be allocated from the maximum heap + size to storing map outputs during the shuffle. + </description> + </property> + + <property> + <name>mapred.map.output.compression.codec</name> + <value></value> + <description>If the map outputs are compressed, how should they be + compressed + </description> + </property> + +<property> + <name>mapred.output.compression.type</name> + <value>BLOCK</value> + <description>If the job outputs are to compressed as SequenceFiles, how should + they be compressed? Should be one of NONE, RECORD or BLOCK. + </description> +</property> + + <property> + <name>mapred.job.reduce.input.buffer.percent</name> + <value>0.0</value> + <description>The percentage of memory- relative to the maximum heap size- to + retain map outputs during the reduce. When the shuffle is concluded, any + remaining map outputs in memory must consume less than this threshold before + the reduce can begin. + </description> + </property> + + <property> + <name>mapreduce.reduce.input.limit</name> + <value>10737418240</value> + <description>The limit on the input size of the reduce. (This value + is 10 Gb.) If the estimated input size of the reduce is greater than + this value, job is failed. A value of -1 means that there is no limit + set. </description> +</property> + + + <!-- copied from kryptonite configuration --> + <property> + <name>mapred.compress.map.output</name> + <value></value> + </property> + + + <property> + <name>mapred.task.timeout</name> + <value>600000</value> + <description>The number of milliseconds before a task will be + terminated if it neither reads an input, writes an output, nor + updates its status string. + </description> + </property> + + <property> + <name>jetty.connector</name> + <value>org.mortbay.jetty.nio.SelectChannelConnector</value> + <description>No description</description> + </property> + + <property> + <name>mapred.child.root.logger</name> + <value>INFO,TLA</value> + </property> + + <property> + <name>mapred.child.java.opts</name> + <value>-Xmx512m</value> + <description>No description</description> + </property> + + <property> + <name>mapred.cluster.reduce.memory.mb</name> + <value></value> + </property> + + <property> + <name>mapred.job.map.memory.mb</name> + <value>1024</value> + </property> + + <property> + <name>mapred.job.reduce.memory.mb</name> + <value>1024</value> + </property> + +<property> + <name>mapred.max.tracker.blacklists</name> + <value>16</value> + <description> + if node is reported blacklisted by 16 successful jobs within timeout-window, it will be graylisted + </description> +</property> + +<property> + <name>mapred.healthChecker.script.path</name> + <value></value> +</property> + +<property> + <name>mapred.healthChecker.script.timeout</name> + <value>60000</value> +</property> + +<property> + <name>mapred.task.maxvmem</name> + <value></value> + <final>true</final> + <description>No description</description> +</property> + +<property> + <name>mapreduce.fileoutputcommitter.marksuccessfuljobs</name> + <value>false</value> +</property> + + <property> + <name>mapreduce.tasktracker.keytab.file</name> + <value></value> + <description>The filename of the keytab for the task tracker</description> + </property> + + <property> + <name>mapreduce.jobtracker.split.metainfo.maxsize</name> + <value>50000000</value> + <final>true</final> + <description>If the size of the split metainfo file is larger than this, the JobTracker will fail the job during + initialize. + </description> + </property> + + <property> + <name>mapreduce.jobhistory.keytab.file</name> + <!-- cluster variant --> + <value></value> + <description>The keytab for the job history server principal.</description> + </property> + +<property> + <name>mapreduce.shuffle.port</name> + <value>8081</value> + <description>Default port that the ShuffleHandler will run on. ShuffleHandler is a service run at the NodeManager to facilitate transfers of intermediate Map outputs to requesting Reducers.</description> +</property> + +<property> + <name>mapreduce.jobhistory.intermediate-done-dir</name> + <value>/mr-history/tmp</value> + <description>Directory where history files are written by MapReduce jobs.</description> +</property> + +<property> + <name>mapreduce.jobhistory.done-dir</name> + <value>/mr-history/done</value> + <description>Directory where history files are managed by the MR JobHistory Server.</description> +</property> + +<property> Â Â Â Â + <name>mapreduce.jobhistory.address</name> Â Â Â Â + <value>localhost:10020</value> + <description>Enter your JobHistoryServer hostname.</description> +</property> + +<property> Â Â Â Â + <name>mapreduce.jobhistory.webapp.address</name> Â Â Â Â + <value>localhost:19888</value> + <description>Enter your JobHistoryServer hostname.</description> +</property> + +<property> + <name>mapreduce.framework.name</name> + <value>yarn</value> + <description>No description</description> +</property> + +</configuration>
http://git-wip-us.apache.org/repos/asf/ambari/blob/ffb4d3b8/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/MAPREDUCE2/metainfo.xml ---------------------------------------------------------------------- diff --git a/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/MAPREDUCE2/metainfo.xml b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/MAPREDUCE2/metainfo.xml new file mode 100644 index 0000000..d6f92d1 --- /dev/null +++ b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/MAPREDUCE2/metainfo.xml @@ -0,0 +1,87 @@ +<?xml version="1.0"?> +<!-- + Licensed to the Apache Software Foundation (ASF) under one or more + contributor license agreements. See the NOTICE file distributed with + this work for additional information regarding copyright ownership. + The ASF licenses this file to You under the Apache License, Version 2.0 + (the "License"); you may not use this file except in compliance with + the License. You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +--> +<metainfo> + <schemaVersion>2.0</schemaVersion> + <services> + <service> + <name>MAPREDUCE2</name> + <comment>Apache Hadoop NextGen MapReduce (client libraries)</comment> + <version>2.0.3.22-1</version> + <components> + <component> + <name>HISTORYSERVER</name> + <category>MASTER</category> + <cardinality>1</cardinality> + <auto-deploy> + <enabled>true</enabled> + <co-locate>YARN/RESOURCEMANAGER</co-locate> + </auto-deploy> + <dependencies> + <dependency> + <name>HDFS/HDFS_CLIENT</name> + <scope>host</scope> + <auto-deploy> + <enabled>true</enabled> + </auto-deploy> + </dependency> + </dependencies> + <commandScript> + <script>scripts/historyserver.py</script> + <scriptType>PYTHON</scriptType> + <timeout>600</timeout> + </commandScript> + </component> + + <component> + <name>MAPREDUCE2_CLIENT</name> + <category>CLIENT</category> + <cardinality>0+</cardinality> + <commandScript> + <script>scripts/mapreduce2_client.py</script> + <scriptType>PYTHON</scriptType> + <timeout>600</timeout> + </commandScript> + </component> + </components> + + <osSpecifics> + <osSpecific> + <osFamily>any</osFamily> + <packages> + <package> + <name>hadoop-mapreduce</name> + </package> + </packages> + </osSpecific> + </osSpecifics> + + <commandScript> + <script>scripts/mapred_service_check.py</script> + <scriptType>PYTHON</scriptType> + <timeout>300</timeout> + </commandScript> + + <configuration-dependencies> + <config-type>core-site</config-type> + <config-type>global</config-type> + <config-type>mapred-site</config-type> + <config-type>mapred-queue-acls</config-type> + </configuration-dependencies> + </service> + </services> +</metainfo> http://git-wip-us.apache.org/repos/asf/ambari/blob/ffb4d3b8/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/OOZIE/configuration/oozie-site.xml ---------------------------------------------------------------------- diff --git a/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/OOZIE/configuration/oozie-site.xml b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/OOZIE/configuration/oozie-site.xml new file mode 100644 index 0000000..69dfe30 --- /dev/null +++ b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/OOZIE/configuration/oozie-site.xml @@ -0,0 +1,245 @@ +<?xml version="1.0"?> +<!-- + Licensed to the Apache Software Foundation (ASF) under one + or more contributor license agreements. See the NOTICE file + distributed with this work for additional information + regarding copyright ownership. The ASF licenses this file + to you under the Apache License, Version 2.0 (the + "License"); you may not use this file except in compliance + with the License. You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +--> + +<configuration> + +<!-- + Refer to the oozie-default.xml file for the complete list of + Oozie configuration properties and their default values. +--> + <property> + <name>oozie.base.url</name> + <value>http://localhost:11000/oozie</value> + <description>Base Oozie URL.</description> + </property> + + <property> + <name>oozie.system.id</name> + <value>oozie-${user.name}</value> + <description> + The Oozie system ID. + </description> + </property> + + <property> + <name>oozie.systemmode</name> + <value>NORMAL</value> + <description> + System mode for Oozie at startup. + </description> + </property> + + <property> + <name>oozie.service.AuthorizationService.security.enabled</name> + <value>true</value> + <description> + Specifies whether security (user name/admin role) is enabled or not. + If disabled any user can manage Oozie system and manage any job. + </description> + </property> + + <property> + <name>oozie.service.PurgeService.older.than</name> + <value>30</value> + <description> + Jobs older than this value, in days, will be purged by the PurgeService. + </description> + </property> + + <property> + <name>oozie.service.PurgeService.purge.interval</name> + <value>3600</value> + <description> + Interval at which the purge service will run, in seconds. + </description> + </property> + + <property> + <name>oozie.service.CallableQueueService.queue.size</name> + <value>1000</value> + <description>Max callable queue size</description> + </property> + + <property> + <name>oozie.service.CallableQueueService.threads</name> + <value>10</value> + <description>Number of threads used for executing callables</description> + </property> + + <property> + <name>oozie.service.CallableQueueService.callable.concurrency</name> + <value>3</value> + <description> + Maximum concurrency for a given callable type. + Each command is a callable type (submit, start, run, signal, job, jobs, suspend,resume, etc). + Each action type is a callable type (Map-Reduce, Pig, SSH, FS, sub-workflow, etc). + All commands that use action executors (action-start, action-end, action-kill and action-check) use + the action type as the callable type. + </description> + </property> + + <property> + <name>oozie.service.coord.normal.default.timeout</name> + <value>120</value> + <description>Default timeout for a coordinator action input check (in minutes) for normal job. + -1 means infinite timeout</description> + </property> + + <property> + <name>oozie.db.schema.name</name> + <value>oozie</value> + <description> + Oozie DataBase Name + </description> + </property> + + <property> + <name>oozie.service.HadoopAccessorService.jobTracker.whitelist</name> + <value> </value> + <description> + Whitelisted job tracker for Oozie service. + </description> + </property> + + <property> + <name>oozie.authentication.type</name> + <value>simple</value> + <description> + </description> + </property> + + <property> + <name>oozie.service.HadoopAccessorService.nameNode.whitelist</name> + <value> </value> + <description> + </description> + </property> + + <property> + <name>oozie.service.WorkflowAppService.system.libpath</name> + <value>/user/${user.name}/share/lib</value> + <description> + System library path to use for workflow applications. + This path is added to workflow application if their job properties sets + the property 'oozie.use.system.libpath' to true. + </description> + </property> + + <property> + <name>use.system.libpath.for.mapreduce.and.pig.jobs</name> + <value>false</value> + <description> + If set to true, submissions of MapReduce and Pig jobs will include + automatically the system library path, thus not requiring users to + specify where the Pig JAR files are. Instead, the ones from the system + library path are used. + </description> + </property> + <property> + <name>oozie.authentication.kerberos.name.rules</name> + <value> + + + + + + </value> + <description>The mapping from kerberos principal names to local OS user names.</description> + </property> + <property> + <name>oozie.service.HadoopAccessorService.hadoop.configurations</name> + <value>*=/etc/hadoop/conf</value> + <description> + Comma separated AUTHORITY=HADOOP_CONF_DIR, where AUTHORITY is the HOST:PORT of + the Hadoop service (JobTracker, HDFS). The wildcard '*' configuration is + used when there is no exact match for an authority. The HADOOP_CONF_DIR contains + the relevant Hadoop *-site.xml files. If the path is relative is looked within + the Oozie configuration directory; though the path can be absolute (i.e. to point + to Hadoop client conf/ directories in the local filesystem. + </description> + </property> + <property> + <name>oozie.service.ActionService.executor.ext.classes</name> + <value> + org.apache.oozie.action.email.EmailActionExecutor, + org.apache.oozie.action.hadoop.HiveActionExecutor, + org.apache.oozie.action.hadoop.ShellActionExecutor, + org.apache.oozie.action.hadoop.SqoopActionExecutor, + org.apache.oozie.action.hadoop.DistcpActionExecutor + </value> + </property> + + <property> + <name>oozie.service.SchemaService.wf.ext.schemas</name> + <value>shell-action-0.1.xsd,email-action-0.1.xsd,hive-action-0.2.xsd,sqoop-action-0.2.xsd,ssh-action-0.1.xsd,distcp-action-0.1.xsd,hive-action-0.3.xsd</value> + </property> + <property> + <name>oozie.service.JPAService.create.db.schema</name> + <value>false</value> + <description> + Creates Oozie DB. + + If set to true, it creates the DB schema if it does not exist. If the DB schema exists is a NOP. + If set to false, it does not create the DB schema. If the DB schema does not exist it fails start up. + </description> + </property> + + <property> + <name>oozie.service.JPAService.jdbc.driver</name> + <value>org.apache.derby.jdbc.EmbeddedDriver</value> + <description> + JDBC driver class. + </description> + </property> + + <property> + <name>oozie.service.JPAService.jdbc.url</name> + <value>jdbc:derby:${oozie.data.dir}/${oozie.db.schema.name}-db;create=true</value> + <description> + JDBC URL. + </description> + </property> + + <property> + <name>oozie.service.JPAService.jdbc.username</name> + <value>sa</value> + <description> + DB user name. + </description> + </property> + + <property> + <name>oozie.service.JPAService.jdbc.password</name> + <value> </value> + <description> + DB user password. + + IMPORTANT: if password is emtpy leave a 1 space string, the service trims the value, + if empty Configuration assumes it is NULL. + </description> + </property> + + <property> + <name>oozie.service.JPAService.pool.max.active.conn</name> + <value>10</value> + <description> + Max number of connections. + </description> + </property> +</configuration> http://git-wip-us.apache.org/repos/asf/ambari/blob/ffb4d3b8/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/OOZIE/metainfo.xml ---------------------------------------------------------------------- diff --git a/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/OOZIE/metainfo.xml b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/OOZIE/metainfo.xml new file mode 100644 index 0000000..d76c7ed --- /dev/null +++ b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/OOZIE/metainfo.xml @@ -0,0 +1,140 @@ +<?xml version="1.0"?> +<!-- + Licensed to the Apache Software Foundation (ASF) under one or more + contributor license agreements. See the NOTICE file distributed with + this work for additional information regarding copyright ownership. + The ASF licenses this file to You under the Apache License, Version 2.0 + (the "License"); you may not use this file except in compliance with + the License. You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +--> +<metainfo> + <schemaVersion>2.0</schemaVersion> + <services> + <service> + <name>OOZIE</name> + <comment>System for workflow coordination and execution of Apache Hadoop jobs</comment> + <version>3.3.1</version> + + <components> + <component> + <name>OOZIE_SERVER</name> + <category>MASTER</category> + <cardinality>1</cardinality> + <dependencies> + <dependency> + <name>HDFS/HDFS_CLIENT</name> + <scope>host</scope> + <auto-deploy> + <enabled>true</enabled> + </auto-deploy> + </dependency> + <dependency> + <name>MAPREDUCE2/MAPREDUCE2_CLIENT</name> + <scope>host</scope> + <auto-deploy> + <enabled>true</enabled> + </auto-deploy> + </dependency> + <dependency> + <name>YARN/YARN_CLIENT</name> + <scope>host</scope> + <auto-deploy> + <enabled>true</enabled> + </auto-deploy> + </dependency> + </dependencies> + <commandScript> + <script>scripts/oozie_server.py</script> + <scriptType>PYTHON</scriptType> + <timeout>600</timeout> + </commandScript> + </component> + + <component> + <name>OOZIE_CLIENT</name> + <category>CLIENT</category> + <cardinality>0+</cardinality> + <dependencies> + <dependency> + <name>HDFS/HDFS_CLIENT</name> + <scope>host</scope> + <auto-deploy> + <enabled>true</enabled> + </auto-deploy> + </dependency> + <dependency> + <name>MAPREDUCE2/MAPREDUCE2_CLIENT</name> + <scope>host</scope> + <auto-deploy> + <enabled>true</enabled> + </auto-deploy> + </dependency> + </dependencies> + <commandScript> + <script>scripts/oozie_client.py</script> + <scriptType>PYTHON</scriptType> + </commandScript> + </component> + </components> + + <osSpecifics> + <osSpecific> + <osFamily>any</osFamily> + <packages> + <package> + <name>oozie</name> + </package> + <package> + <name>oozie-client</name> + </package> + <package> + <name>mysql-connector-java</name> + </package> + </packages> + </osSpecific> + + <osSpecific> + <osFamily>redhat5,redhat6,suse11</osFamily> + <packages> + <package> + <name>extjs-2.2-1</name> + </package> + </packages> + </osSpecific> + + <osSpecific> + <osFamily>ubuntu12</osFamily> + <packages> + <package> + <name>extjs</name> + </package> + <package> + <name>libxml2-utils</name> + </package> + </packages> + </osSpecific> + + </osSpecifics> + + <commandScript> + <script>scripts/service_check.py</script> + <scriptType>PYTHON</scriptType> + <timeout>300</timeout> + </commandScript> + + <configuration-dependencies> + <config-type>global</config-type> + <config-type>oozie-site</config-type> + <config-type>oozie-log4j</config-type> + </configuration-dependencies> + </service> + </services> +</metainfo> http://git-wip-us.apache.org/repos/asf/ambari/blob/ffb4d3b8/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/PIG/configuration/pig.properties ---------------------------------------------------------------------- diff --git a/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/PIG/configuration/pig.properties b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/PIG/configuration/pig.properties new file mode 100644 index 0000000..01000b5 --- /dev/null +++ b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/PIG/configuration/pig.properties @@ -0,0 +1,52 @@ +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +# Pig default configuration file. All values can be overwritten by pig.properties and command line arguments. +# see bin/pig -help + +# brief logging (no timestamps) +brief=false + +#debug level, INFO is default +debug=INFO + +#verbose print all log messages to screen (default to print only INFO and above to screen) +verbose=false + +#exectype local|mapreduce, mapreduce is default +exectype=mapreduce + +#Enable insertion of information about script into hadoop job conf +pig.script.info.enabled=true + +#Do not spill temp files smaller than this size (bytes) +pig.spill.size.threshold=5000000 +#EXPERIMENT: Activate garbage collection when spilling a file bigger than this size (bytes) +#This should help reduce the number of files being spilled. +pig.spill.gc.activation.size=40000000 + +#the following two parameters are to help estimate the reducer number +pig.exec.reducers.bytes.per.reducer=1000000000 +pig.exec.reducers.max=999 + +#Temporary location to store the intermediate data. +pig.temp.dir=/tmp/ + +#Threshold for merging FRJoin fragment files +pig.files.concatenation.threshold=100 +pig.optimistic.files.concatenation=false; + +pig.disable.counter=false http://git-wip-us.apache.org/repos/asf/ambari/blob/ffb4d3b8/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/PIG/metainfo.xml ---------------------------------------------------------------------- diff --git a/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/PIG/metainfo.xml b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/PIG/metainfo.xml new file mode 100644 index 0000000..a727179 --- /dev/null +++ b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/PIG/metainfo.xml @@ -0,0 +1,62 @@ +<?xml version="1.0"?> +<!-- + Licensed to the Apache Software Foundation (ASF) under one or more + contributor license agreements. See the NOTICE file distributed with + this work for additional information regarding copyright ownership. + The ASF licenses this file to You under the Apache License, Version 2.0 + (the "License"); you may not use this file except in compliance with + the License. You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +--> +<metainfo> + <schemaVersion>2.0</schemaVersion> + <services> + <service> + <name>PIG</name> + <comment>Scripting platform for analyzing large datasets</comment> + <version>0.10.1.22-1</version> + + <components> + <component> + <name>PIG</name> + <category>CLIENT</category> + <cardinality>0+</cardinality> + <commandScript> + <script>scripts/pig_client.py</script> + <scriptType>PYTHON</scriptType> + <timeout>600</timeout> + </commandScript> + </component> + </components> + <osSpecifics> + <osSpecific> + <osFamily>any</osFamily> + <packages> + <package> + <name>pig</name> + </package> + </packages> + </osSpecific> + </osSpecifics> + + <commandScript> + <script>scripts/service_check.py</script> + <scriptType>PYTHON</scriptType> + <timeout>300</timeout> + </commandScript> + + <configuration-dependencies> + <config-type>global</config-type> + <config-type>pig-log4j</config-type> + <config-type>pig-properties</config-type> + </configuration-dependencies> + </service> + </services> +</metainfo> http://git-wip-us.apache.org/repos/asf/ambari/blob/ffb4d3b8/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/TEZ/metainfo.xml ---------------------------------------------------------------------- diff --git a/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/TEZ/metainfo.xml b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/TEZ/metainfo.xml new file mode 100644 index 0000000..4ba1371 --- /dev/null +++ b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/TEZ/metainfo.xml @@ -0,0 +1,34 @@ +<?xml version="1.0"?> +<!-- + Licensed to the Apache Software Foundation (ASF) under one or more + contributor license agreements. See the NOTICE file distributed with + this work for additional information regarding copyright ownership. + The ASF licenses this file to You under the Apache License, Version 2.0 + (the "License"); you may not use this file except in compliance with + the License. You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +--> +<metainfo> + <schemaVersion>2.0</schemaVersion> + <services> + <service> + <name>TEZ</name> + <comment>Tez is the next generation Hadoop Query Processing framework written on top of YARN</comment> + <version>0.1.0.22-1</version> + + <components> + <component> + <name>TEZ_CLIENT</name> + <category>CLIENT</category> + </component> + </components> + </service> + </services> +</metainfo> http://git-wip-us.apache.org/repos/asf/ambari/blob/ffb4d3b8/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/WEBHCAT/configuration/webhcat-site.xml ---------------------------------------------------------------------- diff --git a/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/WEBHCAT/configuration/webhcat-site.xml b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/WEBHCAT/configuration/webhcat-site.xml new file mode 100644 index 0000000..31d0113 --- /dev/null +++ b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/WEBHCAT/configuration/webhcat-site.xml @@ -0,0 +1,126 @@ +<?xml version="1.0" encoding="UTF-8"?> +<!-- +Licensed to the Apache Software Foundation (ASF) under one +or more contributor license agreements. See the NOTICE file +distributed with this work for additional information +regarding copyright ownership. The ASF licenses this file +to you under the Apache License, Version 2.0 (the +"License"); you may not use this file except in compliance +with the License. You may obtain a copy of the License at + +http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +--> + +<!-- The default settings for Templeton. --> +<!-- Edit templeton-site.xml to change settings for your local --> +<!-- install. --> + +<configuration> + + <property> + <name>templeton.port</name> + <value>50111</value> + <description>The HTTP port for the main server.</description> + </property> + + <property> + <name>templeton.hadoop.conf.dir</name> + <value>/etc/hadoop/conf</value> + <description>The path to the Hadoop configuration.</description> + </property> + + <property> + <name>templeton.jar</name> + <value>/usr/lib/hcatalog/share/webhcat/svr/webhcat.jar</value> + <description>The path to the Templeton jar file.</description> + </property> + + <property> + <name>templeton.libjars</name> + <value>/usr/lib/zookeeper/zookeeper.jar</value> + <description>Jars to add the the classpath.</description> + </property> + + + <property> + <name>templeton.hadoop</name> + <value>/usr/bin/hadoop</value> + <description>The path to the Hadoop executable.</description> + </property> + + <property> + <name>templeton.pig.archive</name> + <value>hdfs:///apps/webhcat/pig.tar.gz</value> + <description>The path to the Pig archive.</description> + </property> + + <property> + <name>templeton.pig.path</name> + <value>pig.tar.gz/pig/bin/pig</value> + <description>The path to the Pig executable.</description> + </property> + + <property> + <name>templeton.hcat</name> + <value>/usr/bin/hcat</value> + <description>The path to the hcatalog executable.</description> + </property> + + <property> + <name>templeton.hive.archive</name> + <value>hdfs:///apps/webhcat/hive.tar.gz</value> + <description>The path to the Hive archive.</description> + </property> + + <property> + <name>templeton.hive.path</name> + <value>hive.tar.gz/hive/bin/hive</value> + <description>The path to the Hive executable.</description> + </property> + + <property> + <name>templeton.hive.properties</name> + <value></value> + <description>Properties to set when running hive.</description> + </property> + + + <property> + <name>templeton.zookeeper.hosts</name> + <value></value> + <description>ZooKeeper servers, as comma separated host:port pairs</description> + </property> + + <property> + <name>templeton.storage.class</name> + <value>org.apache.hcatalog.templeton.tool.ZooKeeperStorage</value> + <description>The class to use as storage</description> + </property> + + <property> + <name>templeton.override.enabled</name> + <value>false</value> + <description> + Enable the override path in templeton.override.jars + </description> + </property> + + <property> + <name>templeton.streaming.jar</name> + <value>hdfs:///apps/webhcat/hadoop-streaming.jar</value> + <description>The hdfs path to the Hadoop streaming jar file.</description> + </property> + + <property> + <name>templeton.exec.timeout</name> + <value>60000</value> + <description>Time out for templeton api</description> + </property> + +</configuration> http://git-wip-us.apache.org/repos/asf/ambari/blob/ffb4d3b8/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/WEBHCAT/metainfo.xml ---------------------------------------------------------------------- diff --git a/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/WEBHCAT/metainfo.xml b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/WEBHCAT/metainfo.xml new file mode 100644 index 0000000..f704159 --- /dev/null +++ b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/WEBHCAT/metainfo.xml @@ -0,0 +1,102 @@ +<?xml version="1.0"?> +<!-- + Licensed to the Apache Software Foundation (ASF) under one or more + contributor license agreements. See the NOTICE file distributed with + this work for additional information regarding copyright ownership. + The ASF licenses this file to You under the Apache License, Version 2.0 + (the "License"); you may not use this file except in compliance with + the License. You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +--> +<metainfo> + <schemaVersion>2.0</schemaVersion> + <services> + <service> + <name>WEBHCAT</name> + <comment>This is comment for WEBHCAT service</comment> + <version>0.5.0</version> + + <components> + <component> + <name>WEBHCAT_SERVER</name> + <category>MASTER</category> + <cardinality>1</cardinality> + <dependencies> + <dependency> + <name>HDFS/HDFS_CLIENT</name> + <scope>host</scope> + <auto-deploy> + <enabled>true</enabled> + </auto-deploy> + </dependency> + <dependency> + <name>MAPREDUCE2/MAPREDUCE2_CLIENT</name> + <scope>host</scope> + <auto-deploy> + <enabled>true</enabled> + </auto-deploy> + </dependency> + <dependency> + <name>ZOOKEEPER/ZOOKEEPER_SERVER</name> + <scope>cluster</scope> + <auto-deploy> + <enabled>true</enabled> + <co-locate>WEBHCAT/WEBHCAT_SERVER</co-locate> + </auto-deploy> + </dependency> + <dependency> + <name>ZOOKEEPER/ZOOKEEPER_CLIENT</name> + <scope>host</scope> + <auto-deploy> + <enabled>true</enabled> + </auto-deploy> + </dependency> + <dependency> + <name>YARN/YARN_CLIENT</name> + <scope>host</scope> + <auto-deploy> + <enabled>true</enabled> + </auto-deploy> + </dependency> + </dependencies> + <commandScript> + <script>scripts/webhcat_server.py</script> + <scriptType>PYTHON</scriptType> + <timeout>600</timeout> + </commandScript> + </component> + </components> + <osSpecifics> + <osSpecific> + <osFamily>any</osFamily> + <packages> + <package> + <name>hcatalog</name> + </package> + <package> + <name>webhcat-tar-hive</name> + </package> + <package> + <name>webhcat-tar-pig</name> + </package> + </packages> + </osSpecific> + </osSpecifics> + <commandScript> + <script>scripts/service_check.py</script> + <scriptType>PYTHON</scriptType> + <timeout>300</timeout> + </commandScript> + <configuration-dependencies> + <config-type>webhcat-site</config-type> + </configuration-dependencies> + </service> + </services> +</metainfo> http://git-wip-us.apache.org/repos/asf/ambari/blob/ffb4d3b8/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/configuration/capacity-scheduler.xml ---------------------------------------------------------------------- diff --git a/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/configuration/capacity-scheduler.xml b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/configuration/capacity-scheduler.xml new file mode 100644 index 0000000..cc64305 --- /dev/null +++ b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/configuration/capacity-scheduler.xml @@ -0,0 +1,112 @@ +<!-- + Licensed to the Apache Software Foundation (ASF) under one or more + contributor license agreements. See the NOTICE file distributed with + this work for additional information regarding copyright ownership. + The ASF licenses this file to You under the Apache License, Version 2.0 + (the "License"); you may not use this file except in compliance with + the License. You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +--> + +<configuration> + + <property> + <name>yarn.scheduler.capacity.maximum-applications</name> + <value>10000</value> + <description> + Maximum number of applications that can be pending and running. + </description> + </property> + + <property> + <name>yarn.scheduler.capacity.maximum-am-resource-percent</name> + <value>0.1</value> + <description> + Maximum percent of resources in the cluster which can be used to run + application masters i.e. controls number of concurrent running + applications. + </description> + </property> + + <property> + <name>yarn.scheduler.capacity.root.queues</name> + <value>default</value> + <description> + The queues at the this level (root is the root queue). + </description> + </property> + + <property> + <name>yarn.scheduler.capacity.root.capacity</name> + <value>100</value> + <description> + The total capacity as a percentage out of 100 for this queue. + If it has child queues then this includes their capacity as well. + The child queues capacity should add up to their parent queue's capacity + or less. + </description> + </property> + + <property> + <name>yarn.scheduler.capacity.root.default.capacity</name> + <value>100</value> + <description>Default queue target capacity.</description> + </property> + + <property> + <name>yarn.scheduler.capacity.root.default.user-limit-factor</name> + <value>1</value> + <description> + Default queue user limit a percentage from 0.0 to 1.0. + </description> + </property> + + <property> + <name>yarn.scheduler.capacity.root.default.maximum-capacity</name> + <value>100</value> + <description> + The maximum capacity of the default queue. + </description> + </property> + + <property> + <name>yarn.scheduler.capacity.root.default.state</name> + <value>RUNNING</value> + <description> + The state of the default queue. State can be one of RUNNING or STOPPED. + </description> + </property> + + <property> + <name>yarn.scheduler.capacity.root.default.acl_submit_jobs</name> + <value>*</value> + <description> + The ACL of who can submit jobs to the default queue. + </description> + </property> + + <property> + <name>yarn.scheduler.capacity.root.default.acl_administer_jobs</name> + <value>*</value> + <description> + The ACL of who can administer jobs on the default queue. + </description> + </property> + + <property> + <name>yarn.scheduler.capacity.root.acl_administer_queues</name> + <value>*</value> + <description> + The ACL for who can administer this queue i.e. change sub-queue + allocations. + </description> + </property> + +</configuration> http://git-wip-us.apache.org/repos/asf/ambari/blob/ffb4d3b8/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/configuration/container-executor.cfg ---------------------------------------------------------------------- diff --git a/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/configuration/container-executor.cfg b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/configuration/container-executor.cfg new file mode 100644 index 0000000..d07b3db --- /dev/null +++ b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/configuration/container-executor.cfg @@ -0,0 +1,20 @@ +# +# Licensed to the Apache Software Foundation (ASF) under one or more +# contributor license agreements. See the NOTICE file distributed with +# this work for additional information regarding copyright ownership. +# The ASF licenses this file to You under the Apache License, Version 2.0 +# (the "License"); you may not use this file except in compliance with +# the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# +yarn.nodemanager.local-dirs=TODO-YARN-LOCAL-DIR +yarn.nodemanager.linux-container-executor.group=hadoop +yarn.nodemanager.log-dirs=TODO-YARN-LOG-DIR +banned.users=hdfs,bin,0 http://git-wip-us.apache.org/repos/asf/ambari/blob/ffb4d3b8/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/configuration/core-site.xml ---------------------------------------------------------------------- diff --git a/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/configuration/core-site.xml b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/configuration/core-site.xml new file mode 100644 index 0000000..3a2af49 --- /dev/null +++ b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/configuration/core-site.xml @@ -0,0 +1,20 @@ +<?xml version="1.0"?> +<?xml-stylesheet type="text/xsl" href="configuration.xsl"?> +<!-- + Licensed to the Apache Software Foundation (ASF) under one or more + contributor license agreements. See the NOTICE file distributed with + this work for additional information regarding copyright ownership. + The ASF licenses this file to You under the Apache License, Version 2.0 + (the "License"); you may not use this file except in compliance with + the License. You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +--> +<configuration> +</configuration> http://git-wip-us.apache.org/repos/asf/ambari/blob/ffb4d3b8/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/configuration/global.xml ---------------------------------------------------------------------- diff --git a/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/configuration/global.xml b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/configuration/global.xml new file mode 100644 index 0000000..7ddd6e3 --- /dev/null +++ b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/configuration/global.xml @@ -0,0 +1,49 @@ +<?xml version="1.0"?> +<?xml-stylesheet type="text/xsl" href="configuration.xsl"?> +<!-- +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +--> + +<configuration> + <property> + <name>rm_host</name> + <value></value> + <description>ResourceManager.</description> + </property> + <property> + <name>nm_hosts</name> + <value></value> + <description>List of NodeManager Hosts.</description> + </property> + <property> + <name>yarn_log_dir_prefix</name> + <value>/var/log/hadoop-yarn</value> + <description>YARN Log Dir Prefix</description> + </property> + <property> + <name>yarn_pid_dir_prefix</name> + <value>/var/run/hadoop-yarn</value> + <description>YARN PID Dir Prefix</description> + </property> + <property> + <name>yarn_user</name> + <value>yarn</value> + <description>YARN User</description> + </property> +</configuration> http://git-wip-us.apache.org/repos/asf/ambari/blob/ffb4d3b8/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/configuration/yarn-site.xml ---------------------------------------------------------------------- diff --git a/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/configuration/yarn-site.xml b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/configuration/yarn-site.xml new file mode 100644 index 0000000..da6641f --- /dev/null +++ b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/configuration/yarn-site.xml @@ -0,0 +1,172 @@ +<?xml version="1.0"?> +<?xml-stylesheet type="text/xsl" href="configuration.xsl"?> +<!-- + Licensed to the Apache Software Foundation (ASF) under one or more + contributor license agreements. See the NOTICE file distributed with + this work for additional information regarding copyright ownership. + The ASF licenses this file to You under the Apache License, Version 2.0 + (the "License"); you may not use this file except in compliance with + the License. You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +--> + +<!-- Put site-specific property overrides in this file. --> + +<configuration xmlns:xi="http://www.w3.org/2001/XInclude"> + +<!-- ResourceManager --> + + <property> + <name>yarn.resourcemanager.resource-tracker.address</name> + <value>localhost:8025</value> + </property> + + <property> + <name>yarn.resourcemanager.scheduler.address</name> + <value>localhost:8030</value> + </property> + + <property> + <name>yarn.resourcemanager.address</name> + <value>localhost:8050</value> + </property> + + <property> + <name>yarn.resourcemanager.admin.address</name> + <value>localhost:8141</value> + </property> + + <property> + <name>yarn.resourcemanager.scheduler.class</name> + <value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler</value> + </property> + + <property> + <name>yarn.scheduler.minimum-allocation-mb</name> + <value>1024</value> + </property> + + <property> + <name>yarn.scheduler.maximum-allocation-mb</name> + <value>8192</value> + </property> + +<!-- NodeManager --> + + <property> + <name>yarn.nodemanager.address</name> + <value>0.0.0.0:45454</value> + </property> + + <property> + <name>yarn.nodemanager.resource.memory-mb</name> + <value>8192</value> + <description>Amount of physical memory, in MB, that can be allocated + for containers.</description> + </property> + + <property> + <name>yarn.application.classpath</name> + <value>/etc/hadoop/conf,/usr/lib/hadoop/*,/usr/lib/hadoop/lib/*,/usr/lib/hadoop-hdfs/*,/usr/lib/hadoop-hdfs/lib/*,/usr/lib/hadoop-yarn/*,/usr/lib/hadoop-yarn/lib/*,/usr/lib/hadoop-mapreduce/*,/usr/lib/hadoop-mapreduce/lib/*</value> + <description>Classpath for typical applications.</description> + </property> + + <property> + <name>yarn.nodemanager.vmem-pmem-ratio</name> + <value>2.1</value> + <description>Ratio between virtual memory to physical memory when + setting memory limits for containers. Container allocations are + expressed in terms of physical memory, and virtual memory usage + is allowed to exceed this allocation by this ratio. + </description> + </property> + + <property> + <name>yarn.nodemanager.container-executor.class</name> + <value>org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor</value> + <description>ContainerExecutor for launching containers</description> + </property> + + <property> + <name>yarn.nodemanager.aux-services</name> + <value>mapreduce.shuffle</value> + <description>Auxilliary services of NodeManager</description> + </property> + + <property> + <name>yarn.nodemanager.aux-services.class</name> + <value>org.apache.hadoop.mapred.ShuffleHandler</value> + </property> + + <property> + <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name> + <value>org.apache.hadoop.mapred.ShuffleHandler</value> + </property> + + <property> + <name>yarn.nodemanager.log-dirs</name> + <value>/var/log/hadoop/yarn</value> + </property> + + <property> + <name>yarn.nodemanager.container-monitor.interval-ms</name> + <value>3000</value> + <description>The interval, in milliseconds, for which the node manager + waits between two cycles of monitoring its containers' memory usage. + </description> + </property> + + <property> + <name>yarn.nodemanager.health-checker.script.path</name> + <value>/etc/hadoop/conf/health_check</value> + </property> + + <property> + <name>yarn.nodemanager.health-checker.interval-ms</name> + <value>135000</value> + </property> + + <property> + <name>yarn.nodemanager.health-checker.script.timeout-ms</name> + <value>60000</value> + </property> + + <property> + <name>yarn.nodemanager.log.retain-second</name> + <value>604800</value> + </property> + + <property> + <name>yarn.log-aggregation-enable</name> + <value>true</value> + </property> + + <property> + <name>yarn.nodemanager.remote-app-log-dir</name> + <value>/app-logs</value> + </property> + + <property> + <name>yarn.nodemanager.remote-app-log-dir-suffix</name> + <value>logs</value> + </property> + + <property> + <name>yarn.nodemanager.log-aggregation.compression-type</name> + <value>gz</value> + </property> + + <property> + <name>yarn.nodemanager.delete.debug-delay-sec</name> + <value>36000</value> + </property> + + +</configuration> http://git-wip-us.apache.org/repos/asf/ambari/blob/ffb4d3b8/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/metainfo.xml ---------------------------------------------------------------------- diff --git a/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/metainfo.xml b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/metainfo.xml new file mode 100644 index 0000000..6098637 --- /dev/null +++ b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/YARN/metainfo.xml @@ -0,0 +1,110 @@ +<?xml version="1.0"?> +<!-- + Licensed to the Apache Software Foundation (ASF) under one or more + contributor license agreements. See the NOTICE file distributed with + this work for additional information regarding copyright ownership. + The ASF licenses this file to You under the Apache License, Version 2.0 + (the "License"); you may not use this file except in compliance with + the License. You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +--> +<metainfo> + <schemaVersion>2.0</schemaVersion> + <services> + <service> + <name>YARN</name> + <comment>Apache Hadoop NextGen MapReduce (YARN)</comment> + <version>2.0.3.22-1</version> + <components> + + <component> + <name>RESOURCEMANAGER</name> + <category>MASTER</category> + <cardinality>1</cardinality> + <commandScript> + <script>scripts/resourcemanager.py</script> + <scriptType>PYTHON</scriptType> + <timeout>600</timeout> + </commandScript> + <customCommands> + <customCommand> + <name>DECOMMISSION</name> + <commandScript> + <script>scripts/resourcemanager.py</script> + <scriptType>PYTHON</scriptType> + <timeout>600</timeout> + </commandScript> + </customCommand> + <customCommand> + <name>REFRESHQUEUES</name> + <commandScript> + <script>scripts/resourcemanager.py</script> + <scriptType>PYTHON</scriptType> + <timeout>600</timeout> + </commandScript> + </customCommand> + </customCommands> + </component> + + <component> + <name>NODEMANAGER</name> + <category>SLAVE</category> + <cardinality>1+</cardinality> + <commandScript> + <script>scripts/nodemanager.py</script> + <scriptType>PYTHON</scriptType> + <timeout>600</timeout> + </commandScript> + </component> + + <component> + <name>YARN_CLIENT</name> + <category>CLIENT</category> + <cardinality>0+</cardinality> + <commandScript> + <script>scripts/yarn_client.py</script> + <scriptType>PYTHON</scriptType> + <timeout>600</timeout> + </commandScript> + </component> + </components> + + <osSpecifics> + <osSpecific> + <osFamily>any</osFamily> + <packages> + <package> + <name>hadoop-yarn</name> + </package> + <package> + <name>hadoop-mapreduce</name> + </package> + </packages> + </osSpecific> + </osSpecifics> + + <commandScript> + <script>scripts/service_check.py</script> + <scriptType>PYTHON</scriptType> + <timeout>300</timeout> + </commandScript> + + <configuration-dependencies> + <config-type>yarn-site</config-type> + <config-type>capacity-scheduler</config-type> + <config-type>core-site</config-type> + <config-type>global</config-type> + <config-type>mapred-site</config-type> + <config-type>mapred-queue-acls</config-type> + <config-type>yarn-log4j</config-type> + </configuration-dependencies> + </service> + </services> +</metainfo> http://git-wip-us.apache.org/repos/asf/ambari/blob/ffb4d3b8/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/ZOOKEEPER/metainfo.xml ---------------------------------------------------------------------- diff --git a/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/ZOOKEEPER/metainfo.xml b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/ZOOKEEPER/metainfo.xml new file mode 100644 index 0000000..ca13ed7 --- /dev/null +++ b/ambari-funtest/src/test/resources/stacks/HDP/2.0.1/services/ZOOKEEPER/metainfo.xml @@ -0,0 +1,75 @@ +<?xml version="1.0"?> +<!-- + Licensed to the Apache Software Foundation (ASF) under one or more + contributor license agreements. See the NOTICE file distributed with + this work for additional information regarding copyright ownership. + The ASF licenses this file to You under the Apache License, Version 2.0 + (the "License"); you may not use this file except in compliance with + the License. You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +--> +<metainfo> + <schemaVersion>2.0</schemaVersion> + <services> + <service> + <name>ZOOKEEPER</name> + <comment>Centralized service which provides highly reliable distributed coordination</comment> + <version>3.4.5.22-1</version> + + <components> + + <component> + <name>ZOOKEEPER_SERVER</name> + <category>MASTER</category> + <cardinality>1+</cardinality> + <commandScript> + <script>scripts/zookeeper_server.py</script> + <scriptType>PYTHON</scriptType> + <timeout>600</timeout> + </commandScript> + </component> + + <component> + <name>ZOOKEEPER_CLIENT</name> + <category>CLIENT</category> + <cardinality>0+</cardinality> + <commandScript> + <script>scripts/zookeeper_client.py</script> + <scriptType>PYTHON</scriptType> + </commandScript> + </component> + </components> + + <osSpecifics> + <osSpecific> + <osFamily>any</osFamily> + <packages> + <package> + <name>zookeeper</name> + </package> + </packages> + </osSpecific> + </osSpecifics> + + <commandScript> + <script>scripts/service_check.py</script> + <scriptType>PYTHON</scriptType> + <timeout>300</timeout> + </commandScript> + + <configuration-dependencies> + <config-type>global</config-type> + <config-type>zookeeper-log4j</config-type> + <config-type>zoo.cfg</config-type> + </configuration-dependencies> + <restartRequiredAfterChange>true</restartRequiredAfterChange> + </service> + </services> +</metainfo> http://git-wip-us.apache.org/repos/asf/ambari/blob/ffb4d3b8/ambari-funtest/src/test/resources/stacks/HDP/2.0.5/metainfo.xml ---------------------------------------------------------------------- diff --git a/ambari-funtest/src/test/resources/stacks/HDP/2.0.5/metainfo.xml b/ambari-funtest/src/test/resources/stacks/HDP/2.0.5/metainfo.xml new file mode 100644 index 0000000..ca45822 --- /dev/null +++ b/ambari-funtest/src/test/resources/stacks/HDP/2.0.5/metainfo.xml @@ -0,0 +1,22 @@ +<?xml version="1.0"?> +<!-- + Licensed to the Apache Software Foundation (ASF) under one or more + contributor license agreements. See the NOTICE file distributed with + this work for additional information regarding copyright ownership. + The ASF licenses this file to You under the Apache License, Version 2.0 + (the "License"); you may not use this file except in compliance with + the License. You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +--> +<metainfo> + <versions> + <active>true</active> + </versions> +</metainfo> http://git-wip-us.apache.org/repos/asf/ambari/blob/ffb4d3b8/ambari-funtest/src/test/resources/stacks/HDP/2.0.5/repos/repoinfo.xml ---------------------------------------------------------------------- diff --git a/ambari-funtest/src/test/resources/stacks/HDP/2.0.5/repos/repoinfo.xml b/ambari-funtest/src/test/resources/stacks/HDP/2.0.5/repos/repoinfo.xml new file mode 100644 index 0000000..91f0f02 --- /dev/null +++ b/ambari-funtest/src/test/resources/stacks/HDP/2.0.5/repos/repoinfo.xml @@ -0,0 +1,61 @@ +<?xml version="1.0"?> +<!-- + Licensed to the Apache Software Foundation (ASF) under one or more + contributor license agreements. See the NOTICE file distributed with + this work for additional information regarding copyright ownership. + The ASF licenses this file to You under the Apache License, Version 2.0 + (the "License"); you may not use this file except in compliance with + the License. You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +--> +<reposinfo> + <os family="centos6"> + <repo> + <baseurl>http://dev.hortonworks.com.s3.amazonaws.com/HDP-2.0.5.0/repos/centos6</baseurl> + <repoid>HDP-2.0.5</repoid> + <reponame>HDP</reponame> + </repo> + </os> + <os family="centos5"> + <repo> + <baseurl>http://dev.hortonworks.com.s3.amazonaws.com/HDP-2.0.5.0/repos/centos5</baseurl> + <repoid>HDP-2.0.5</repoid> + <reponame>HDP</reponame> + </repo> + </os> + <os family="redhat6"> + <repo> + <baseurl>http://dev.hortonworks.com.s3.amazonaws.com/HDP-2.0.5.0/repos/centos6</baseurl> + <repoid>HDP-2.0.5</repoid> + <reponame>HDP</reponame> + </repo> + </os> + <os family="redhat5"> + <repo> + <baseurl>http://dev.hortonworks.com.s3.amazonaws.com/HDP-2.0.5.0/repos/centos5</baseurl> + <repoid>HDP-2.0.5</repoid> + <reponame>HDP</reponame> + </repo> + </os> + <os family="suse11"> + <repo> + <baseurl>http://dev.hortonworks.com.s3.amazonaws.com/HDP-2.0.5.0/repos/suse11</baseurl> + <repoid>HDP-2.0.5</repoid> + <reponame>HDP</reponame> + </repo> + </os> + <os family="sles11"> + <repo> + <baseurl>http://dev.hortonworks.com.s3.amazonaws.com/HDP-2.0.5.0/repos/suse11</baseurl> + <repoid>HDP-2.0.5</repoid> + <reponame>HDP</reponame> + </repo> + </os> +</reposinfo> http://git-wip-us.apache.org/repos/asf/ambari/blob/ffb4d3b8/ambari-funtest/src/test/resources/stacks/HDP/2.0.5/services/GANGLIA/metainfo.xml ---------------------------------------------------------------------- diff --git a/ambari-funtest/src/test/resources/stacks/HDP/2.0.5/services/GANGLIA/metainfo.xml b/ambari-funtest/src/test/resources/stacks/HDP/2.0.5/services/GANGLIA/metainfo.xml new file mode 100644 index 0000000..9167f87 --- /dev/null +++ b/ambari-funtest/src/test/resources/stacks/HDP/2.0.5/services/GANGLIA/metainfo.xml @@ -0,0 +1,130 @@ +<?xml version="1.0"?> +<!-- + Licensed to the Apache Software Foundation (ASF) under one or more + contributor license agreements. See the NOTICE file distributed with + this work for additional information regarding copyright ownership. + The ASF licenses this file to You under the Apache License, Version 2.0 + (the "License"); you may not use this file except in compliance with + the License. You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +--> +<metainfo> + <schemaVersion>2.0</schemaVersion> + <services> + <service> + <name>GANGLIA</name> + <comment>Ganglia Metrics Collection system (<a href="http://oss.oetiker.ch/rrdtool/" target="_blank">RRDTool</a> will be installed too)</comment> + <version>3.5.0</version> + <components> + <component> + <name>GANGLIA_SERVER</name> + <category>MASTER</category> + <cardinality>1</cardinality> + <versionAdvertised>false</versionAdvertised> + <commandScript> + <script>scripts/ganglia_server.py</script> + <scriptType>PYTHON</scriptType> + <timeout>600</timeout> + </commandScript> + </component> + + <component> + <name>GANGLIA_MONITOR</name> + <category>SLAVE</category> + <cardinality>ALL</cardinality> + <versionAdvertised>false</versionAdvertised> + <auto-deploy> + <enabled>true</enabled> + </auto-deploy> + <commandScript> + <script>scripts/ganglia_monitor.py</script> + <scriptType>PYTHON</scriptType> + <timeout>600</timeout> + </commandScript> + </component> + </components> + <osSpecifics> + <osSpecific> + <osFamily>any</osFamily> + <packages> + <package> + <name>python-rrdtool</name> + </package> + </packages> + </osSpecific> + <osSpecific> + <osFamily>redhat5,redhat6,suse11</osFamily> + <packages> + <package> + <name>libganglia-3.5.0-99</name> + </package> + <package> + <name>ganglia-devel-3.5.0-99</name> + </package> + <package> + <name>ganglia-gmetad-3.5.0-99</name> + </package> + <package> + <name>ganglia-web-3.5.7-99.noarch</name> + </package> + <package> + <name>python-rrdtool.x86_64</name> + </package> + <package> + <name>ganglia-gmond-3.5.0-99</name> + </package> + <package> + <name>ganglia-gmond-modules-python-3.5.0-99</name> + </package> + </packages> + </osSpecific> + <osSpecific> + <osFamily>ubuntu12</osFamily> + <packages> + <package> + <name>gmetad</name> + </package> + <package> + <name>ganglia-webfrontend</name> + </package> + <package> + <name>ganglia-monitor-python</name> + </package> + <package> + <name>rrdcached</name> + </package> + </packages> + </osSpecific> + <osSpecific> + <osFamily>suse11</osFamily> + <packages> + <package> + <name>apache2</name> + </package> + <package> + <name>apache2?mod_php*</name> + </package> + </packages> + </osSpecific> + <osSpecific> + <osFamily>redhat5,redhat6</osFamily> + <packages> + <package> + <name>httpd</name> + </package> + </packages> + </osSpecific> + </osSpecifics> + <configuration-dependencies> + <config-type>global</config-type> + </configuration-dependencies> + </service> + </services> +</metainfo> http://git-wip-us.apache.org/repos/asf/ambari/blob/ffb4d3b8/ambari-funtest/src/test/resources/stacks/HDP/2.0.5/services/HBASE/configuration/global.xml ---------------------------------------------------------------------- diff --git a/ambari-funtest/src/test/resources/stacks/HDP/2.0.5/services/HBASE/configuration/global.xml b/ambari-funtest/src/test/resources/stacks/HDP/2.0.5/services/HBASE/configuration/global.xml new file mode 100644 index 0000000..c0d1fa9 --- /dev/null +++ b/ambari-funtest/src/test/resources/stacks/HDP/2.0.5/services/HBASE/configuration/global.xml @@ -0,0 +1,179 @@ +<?xml version="1.0"?> +<?xml-stylesheet type="text/xsl" href="configuration.xsl"?> +<!-- +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +--> + +<configuration> + <property> + <name>hbasemaster_host</name> + <value></value> + <description>HBase Master Host.</description> + </property> + <property> + <name>regionserver_hosts</name> + <value></value> + <description>Region Server Hosts</description> + </property> + <property> + <name>hbase_log_dir</name> + <value>/var/log/hbase</value> + <description>Log Directories for HBase.</description> + </property> + <property> + <name>hbase_pid_dir</name> + <value>/var/run/hbase</value> + <description>Log Directories for HBase.</description> + </property> + <property> + <name>hbase_log_dir</name> + <value>/var/log/hbase</value> + <description>Log Directories for HBase.</description> + </property> + <property> + <name>hbase_regionserver_heapsize</name> + <value>1024</value> + <description>RegionServers heap size.</description> + </property> + <property> + <name>hbase_regionserver_xmn_max</name> + <value>512</value> + <description> +Sets the upper bound on HBase RegionServers' young generation size. +This value is used in case the young generation size (-Xmn) calculated based on the max heapsize (hbase_regionserver_heapsize) +and the -Xmn ratio (hbase_regionserver_xmn_ratio) exceeds this value. + </description> + </property> + <property> + <name>hbase_regionserver_xmn_ratio</name> + <value>0.2</value> + <description>Percentage of max heap size which used for young generation heap (-Xmx).</description> + </property> + <property> + <name>hbase_master_heapsize</name> + <value>1024</value> + <description>HBase Master Heap Size</description> + </property> + <property> + <name>hstore_compactionthreshold</name> + <value>3</value> + <description>HBase HStore compaction threshold.</description> + </property> + <property> + <name>hfile_blockcache_size</name> + <value>0.25</value> + <description>HFile block cache size.</description> + </property> + <property> + <name>hstorefile_maxsize</name> + <value>1073741824</value> + <description>Maximum HStoreFile Size</description> + </property> + <property> + <name>regionserver_handlers</name> + <value>30</value> + <description>HBase RegionServer Handler</description> + </property> + <property> + <name>hregion_majorcompaction</name> + <value>86400000</value> + <description>HBase Major Compaction.</description> + </property> + <property> + <name>hregion_blockmultiplier</name> + <value>2</value> + <description>HBase Region Block Multiplier</description> + </property> + <property> + <name>hregion_memstoreflushsize</name> + <value></value> + <description>HBase Region MemStore Flush Size.</description> + </property> + <property> + <name>client_scannercaching</name> + <value>100</value> + <description>Base Client Scanner Caching</description> + </property> + <property> + <name>zookeeper_sessiontimeout</name> + <value>60000</value> + <description>ZooKeeper Session Timeout</description> + </property> + <property> + <name>hfile_max_keyvalue_size</name> + <value>10485760</value> + <description>HBase Client Maximum key-value Size</description> + </property> + <property> + <name>hbase_hdfs_root_dir</name> + <value>/apps/hbase/data</value> + <description>HBase Relative Path to HDFS.</description> + </property> + <property> + <name>hbase_tmp_dir</name> + <value>/var/log/hbase</value> + <description>Hbase temp directory</description> + </property> + <property> + <name>hbase_conf_dir</name> + <value>/etc/hbase</value> + <description>Config Directory for HBase.</description> + </property> + <property> + <name>hdfs_enable_shortcircuit_read</name> + <value>true</value> + <description>HDFS Short Circuit Read</description> + </property> + <property> + <name>hdfs_support_append</name> + <value>true</value> + <description>HDFS append support</description> + </property> + <property> + <name>hstore_blockingstorefiles</name> + <value>7</value> + <description>HStore blocking storefiles.</description> + </property> + <property> + <name>regionserver_memstore_lab</name> + <value>true</value> + <description>Region Server memstore.</description> + </property> + <property> + <name>regionserver_memstore_lowerlimit</name> + <value>0.35</value> + <description>Region Server memstore lower limit.</description> + </property> + <property> + <name>regionserver_memstore_upperlimit</name> + <value>0.4</value> + <description>Region Server memstore upper limit.</description> + </property> + <property> + <name>hbase_conf_dir</name> + <value>/etc/hbase</value> + <description>HBase conf dir.</description> + </property> + <property> + <name>hbase_user</name> + <value>hbase</value> + <description>HBase User Name.</description> + </property> + +</configuration> http://git-wip-us.apache.org/repos/asf/ambari/blob/ffb4d3b8/ambari-funtest/src/test/resources/stacks/HDP/2.0.5/services/HBASE/configuration/hbase-policy.xml ---------------------------------------------------------------------- diff --git a/ambari-funtest/src/test/resources/stacks/HDP/2.0.5/services/HBASE/configuration/hbase-policy.xml b/ambari-funtest/src/test/resources/stacks/HDP/2.0.5/services/HBASE/configuration/hbase-policy.xml new file mode 100644 index 0000000..e45f23c --- /dev/null +++ b/ambari-funtest/src/test/resources/stacks/HDP/2.0.5/services/HBASE/configuration/hbase-policy.xml @@ -0,0 +1,53 @@ +<?xml version="1.0"?> +<?xml-stylesheet type="text/xsl" href="configuration.xsl"?> +<!-- +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +--> + +<configuration> + <property> + <name>security.client.protocol.acl</name> + <value>*</value> + <description>ACL for HRegionInterface protocol implementations (ie. + clients talking to HRegionServers) + The ACL is a comma-separated list of user and group names. The user and + group list is separated by a blank. For e.g. "alice,bob users,wheel". + A special value of "*" means all users are allowed.</description> + </property> + + <property> + <name>security.admin.protocol.acl</name> + <value>*</value> + <description>ACL for HMasterInterface protocol implementation (ie. + clients talking to HMaster for admin operations). + The ACL is a comma-separated list of user and group names. The user and + group list is separated by a blank. For e.g. "alice,bob users,wheel". + A special value of "*" means all users are allowed.</description> + </property> + + <property> + <name>security.masterregion.protocol.acl</name> + <value>*</value> + <description>ACL for HMasterRegionInterface protocol implementations + (for HRegionServers communicating with HMaster) + The ACL is a comma-separated list of user and group names. The user and + group list is separated by a blank. For e.g. "alice,bob users,wheel". + A special value of "*" means all users are allowed.</description> + </property> +</configuration>
