Updated Branches: refs/heads/trunk 8871874f0 -> 06d53ed97
AMBARI-2888. Cleaned up Hadoop2 configs (ncole) Project: http://git-wip-us.apache.org/repos/asf/incubator-ambari/repo Commit: http://git-wip-us.apache.org/repos/asf/incubator-ambari/commit/06d53ed9 Tree: http://git-wip-us.apache.org/repos/asf/incubator-ambari/tree/06d53ed9 Diff: http://git-wip-us.apache.org/repos/asf/incubator-ambari/diff/06d53ed9 Branch: refs/heads/trunk Commit: 06d53ed972639d415704764a6ab86fd89d040950 Parents: 8871874 Author: Nate Cole <[email protected]> Authored: Wed Aug 14 10:32:50 2013 -0400 Committer: Nate Cole <[email protected]> Committed: Wed Aug 14 13:08:54 2013 -0400 ---------------------------------------------------------------------- .../services/HDFS/configuration/core-site.xml | 63 ------------------ .../services/HDFS/configuration/hdfs-site.xml | 67 +------------------- .../MAPREDUCE2/configuration/mapred-site.xml | 42 +----------- .../services/YARN/configuration/yarn-site.xml | 11 ---- .../services/HDFS/configuration/core-site.xml | 63 ------------------ .../services/HDFS/configuration/hdfs-site.xml | 67 +------------------- .../MAPREDUCE2/configuration/mapred-site.xml | 38 ----------- .../services/YARN/configuration/yarn-site.xml | 5 -- 8 files changed, 4 insertions(+), 352 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/06d53ed9/ambari-server/src/main/resources/stacks/HDP/2.0.5/services/HDFS/configuration/core-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.5/services/HDFS/configuration/core-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.5/services/HDFS/configuration/core-site.xml index ceab4ed..d9ca5e8 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.0.5/services/HDFS/configuration/core-site.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.0.5/services/HDFS/configuration/core-site.xml @@ -45,12 +45,6 @@ for compression/decompression.</description> </property> - <property> - <name>io.compression.codec.lzo.class</name> - <value>com.hadoop.compression.lzo.LzoCodec</value> - <description>The implementation for lzo codec.</description> - </property> - <!-- file system properties --> <property> @@ -98,14 +92,6 @@ </description> </property> - <property> - <name>fs.checkpoint.size</name> - <value>536870912</value> - <description>The size of the current edit log (in bytes) that triggers - a periodic checkpoint even if the dfs.namenode.checkpoint.period hasn't expired. - </description> - </property> - <!-- ipc properties: copied from kryptonite configuration --> <property> <name>ipc.client.idlethreshold</name> @@ -206,53 +192,4 @@ DEFAULT </description> </property> -<!-- -<property> - <name>hadoop.proxyuser.scope.function_hdp_user("hcat_user").groups</name> - <value></value> - <description> - Proxy group for Hadoop. - </description> -</property> - -<property> - <name>hadoop.proxyuser.scope.function_hdp_user("hcat_user").hosts</name> - <value></value> - <description> - Proxy host for Hadoop. - </description> -</property> - -<property> - <name>hadoop.proxyuser.scope.function_hdp_user("oozie_user").groups</name> - <value></value> - <description> - Proxy group for Hadoop. - </description> -</property> - -<property> - <name>hadoop.proxyuser.scope.function_hdp_user("oozie_user").hosts</name> - <value></value> - <description> - Proxy host for Hadoop. - </description> -</property> - -<property> - <name>hadoop.proxyuser.scope.function_hdp_user("templeton_user").groups</name> - <value></value> - <description> - Proxy group for templeton. - </description> -</property> - -<property> - <name>hadoop.proxyuser.scope.function_hdp_user("templeton_user").hosts</name> - <value></value> - <description> - Proxy host for templeton. - </description> -</property> ---> </configuration> http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/06d53ed9/ambari-server/src/main/resources/stacks/HDP/2.0.5/services/HDFS/configuration/hdfs-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.5/services/HDFS/configuration/hdfs-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.5/services/HDFS/configuration/hdfs-site.xml index d5d0106..ecb5d6f 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.0.5/services/HDFS/configuration/hdfs-site.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.0.5/services/HDFS/configuration/hdfs-site.xml @@ -49,14 +49,6 @@ <final>true</final> </property> -<!-- - <property> - <name>dfs.datanode.socket.write.timeout</name> - <value>0</value> - <description>DFS Client write socket timeout</description> - </property> ---> - <property> <name>dfs.datanode.failed.volumes.tolerated</name> <value>0</value> @@ -65,15 +57,6 @@ </property> <property> - <name>dfs.block.local-path-access.user</name> - <value>hbase</value> - <description>the user who is allowed to perform short - circuit reads. - </description> - <final>true</final> - </property> - - <property> <name>dfs.datanode.data.dir</name> <value></value> <description>Determines where on the local filesystem an DFS data node @@ -106,15 +89,6 @@ --> <property> - <name>dfs.checksum.type</name> - <value>CRC32</value> - <description>The checksum method to be used by default. To maintain - compatibility, it is being set to CRC32. Once all migration steps - are complete, we can change it to CRC32C and take advantage of the - additional performance benefit.</description> - </property> - - <property> <name>dfs.replication.max</name> <value>50</value> <description>Maximal block replication. @@ -209,13 +183,6 @@ If the port is 0 then the server will start on a free port. </property> <property> -<name>dfs.datanode.du.pct</name> -<value>0.85f</value> -<description>When calculating remaining space, only use this percentage of the real available space -</description> -</property> - -<property> <name>dfs.namenode.handler.count</name> <value>40</value> <description>The number of server threads for the namenode.</description> @@ -238,15 +205,6 @@ The octal umask used when creating files and directories. </property> <property> -<name>dfs.web.ugi</name> -<!-- cluster variant --> -<value>gopher,gopher</value> -<description>The user account used by the web interface. -Syntax: USERNAME,GROUP1,GROUP2, ... -</description> -</property> - -<property> <name>dfs.permissions.enabled</name> <value>true</value> <description> @@ -259,7 +217,7 @@ owner or group of files or directories. </property> <property> -<name>dfs.permissions.enabled.superusergroup</name> +<name>dfs.permissions.superusergroup</name> <value>hdfs</value> <description>The name of the group of super-users.</description> </property> @@ -271,10 +229,6 @@ owner or group of files or directories. </property> <property> -<name>ipc.server.max.response.size</name> -<value>5242880</value> -</property> -<property> <name>dfs.block.access.token.enable</name> <value>true</value> <description> @@ -325,12 +279,6 @@ Kerberos principal name for the NameNode </property> <property> - <name>dfs.secondary.https.port</name> - <value>50490</value> - <description>The https port where secondary-namenode binds</description> - </property> - - <property> <name>dfs.web.authentication.kerberos.principal</name> <value></value> <description> @@ -382,13 +330,6 @@ Kerberos principal name for the NameNode </property> <property> - <name>dfs.https.port</name> - <value>50470</value> - <description>The https port where namenode binds</description> - - </property> - - <property> <name>dfs.namenode.https-address</name> <value></value> <description>The https address where namenode binds</description> @@ -420,12 +361,6 @@ don't exist, they will be created with this permission.</description> </property> <property> - <name>ipc.server.read.threadpool.size</name> - <value>5</value> - <description></description> - </property> - - <property> <name>dfs.namenode.avoid.read.stale.datanode</name> <value>true</value> <description> http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/06d53ed9/ambari-server/src/main/resources/stacks/HDP/2.0.5/services/MAPREDUCE2/configuration/mapred-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.5/services/MAPREDUCE2/configuration/mapred-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.5/services/MAPREDUCE2/configuration/mapred-site.xml index 53b7fb7..51e3e4d 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.0.5/services/MAPREDUCE2/configuration/mapred-site.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.0.5/services/MAPREDUCE2/configuration/mapred-site.xml @@ -31,12 +31,6 @@ </property> <property> - <name>io.sort.record.percent</name> - <value>.2</value> - <description>No description</description> - </property> - - <property> <name>mapreduce.map.sort.spill.percent</name> <value>0.1</value> <description>No description</description> @@ -178,17 +172,6 @@ </property> <property> - <name>jetty.connector</name> - <value>org.mortbay.jetty.nio.SelectChannelConnector</value> - <description>No description</description> - </property> - - <property> - <name>mapred.child.root.logger</name> - <value>INFO,TLA</value> - </property> - - <property> <name>mapred.child.java.opts</name> <value>-Xmx512m</value> <description>No description</description> @@ -228,32 +211,11 @@ </property> <property> - <name>mapred.task.maxvmem</name> - <value></value> - <final>true</final> - <description>No description</description> - </property> - - <property> - <name>mapreduce.fileoutputcommitter.marksuccessfuljobs</name> - <value>false</value> - </property> - - <property> <name>mapreduce.tasktracker.keytab.file</name> <value></value> <description>The filename of the keytab for the task tracker</description> </property> - <property> - <name>mapreduce.jobtracker.split.metainfo.maxsize</name> - <value>50000000</value> - <final>true</final> - <description>If the size of the split metainfo file is larger than this, the JobTracker will fail the job during - initialize. - </description> - </property> - <property> <name>mapreduce.jobhistory.keytab.file</name> <!-- cluster variant --> @@ -280,13 +242,13 @@ </property> <property> Â Â Â Â - <name>mapreduce.jobhistory.address</name> Â Â Â Â + <name>mapreduce.jobhistory.address</name> <value>localhost:10020</value> <description>Enter your JobHistoryServer hostname.</description> </property> <property> Â Â Â Â - <name>mapreduce.jobhistory.webapp.address</name> Â Â Â Â + <name>mapreduce.jobhistory.webapp.address</name> <value>localhost:19888</value> <description>Enter your JobHistoryServer hostname.</description> </property> http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/06d53ed9/ambari-server/src/main/resources/stacks/HDP/2.0.5/services/YARN/configuration/yarn-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.5/services/YARN/configuration/yarn-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.5/services/YARN/configuration/yarn-site.xml index 630c95e..8c7e5f6 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.0.5/services/YARN/configuration/yarn-site.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.0.5/services/YARN/configuration/yarn-site.xml @@ -101,11 +101,6 @@ </property> <property> - <name>yarn.nodemanager.aux-services.class</name> - <value>org.apache.hadoop.mapred.ShuffleHandler</value> - </property> - - <property> <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name> <value>org.apache.hadoop.mapred.ShuffleHandler</value> </property> @@ -121,12 +116,6 @@ </property> <property> - <name>yarn.nodemanager.linux-container-executor.group</name> - <value>hadoop</value> - <description>Unix group of the NodeManager</description> - </property> - - <property> <name>yarn.nodemanager.container-executor.class</name> <value>org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor</value> <description>Executor(launcher) of the containers</description> http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/06d53ed9/ambari-server/src/main/resources/stacks/HDPLocal/2.0.5/services/HDFS/configuration/core-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDPLocal/2.0.5/services/HDFS/configuration/core-site.xml b/ambari-server/src/main/resources/stacks/HDPLocal/2.0.5/services/HDFS/configuration/core-site.xml index e6742a5..7e6514d 100644 --- a/ambari-server/src/main/resources/stacks/HDPLocal/2.0.5/services/HDFS/configuration/core-site.xml +++ b/ambari-server/src/main/resources/stacks/HDPLocal/2.0.5/services/HDFS/configuration/core-site.xml @@ -45,12 +45,6 @@ for compression/decompression.</description> </property> - <property> - <name>io.compression.codec.lzo.class</name> - <value>com.hadoop.compression.lzo.LzoCodec</value> - <description>The implementation for lzo codec.</description> - </property> - <!-- file system properties --> <property> @@ -98,14 +92,6 @@ </description> </property> - <property> - <name>fs.checkpoint.size</name> - <value>536870912</value> - <description>The size of the current edit log (in bytes) that triggers - a periodic checkpoint even if the fs.checkpoint.period hasn't expired. - </description> - </property> - <!-- ipc properties: copied from kryptonite configuration --> <property> <name>ipc.client.idlethreshold</name> @@ -205,53 +191,4 @@ DEFAULT </description> </property> -<!-- -<property> - <name>hadoop.proxyuser.scope.function_hdp_user("hcat_user").groups</name> - <value></value> - <description> - Proxy group for Hadoop. - </description> -</property> - -<property> - <name>hadoop.proxyuser.scope.function_hdp_user("hcat_user").hosts</name> - <value></value> - <description> - Proxy host for Hadoop. - </description> -</property> - -<property> - <name>hadoop.proxyuser.scope.function_hdp_user("oozie_user").groups</name> - <value></value> - <description> - Proxy group for Hadoop. - </description> -</property> - -<property> - <name>hadoop.proxyuser.scope.function_hdp_user("oozie_user").hosts</name> - <value></value> - <description> - Proxy host for Hadoop. - </description> -</property> - -<property> - <name>hadoop.proxyuser.scope.function_hdp_user("templeton_user").groups</name> - <value></value> - <description> - Proxy group for templeton. - </description> -</property> - -<property> - <name>hadoop.proxyuser.scope.function_hdp_user("templeton_user").hosts</name> - <value></value> - <description> - Proxy host for templeton. - </description> -</property> ---> </configuration> http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/06d53ed9/ambari-server/src/main/resources/stacks/HDPLocal/2.0.5/services/HDFS/configuration/hdfs-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDPLocal/2.0.5/services/HDFS/configuration/hdfs-site.xml b/ambari-server/src/main/resources/stacks/HDPLocal/2.0.5/services/HDFS/configuration/hdfs-site.xml index 820a164..06709ac 100644 --- a/ambari-server/src/main/resources/stacks/HDPLocal/2.0.5/services/HDFS/configuration/hdfs-site.xml +++ b/ambari-server/src/main/resources/stacks/HDPLocal/2.0.5/services/HDFS/configuration/hdfs-site.xml @@ -49,14 +49,6 @@ <final>true</final> </property> -<!-- - <property> - <name>dfs.datanode.socket.write.timeout</name> - <value>0</value> - <description>DFS Client write socket timeout</description> - </property> ---> - <property> <name>dfs.datanode.failed.volumes.tolerated</name> <value>0</value> @@ -65,15 +57,6 @@ </property> <property> - <name>dfs.block.local-path-access.user</name> - <value>hbase</value> - <description>the user who is allowed to perform short - circuit reads. - </description> - <final>true</final> - </property> - - <property> <name>dfs.datanode.data.dir</name> <value></value> <description>Determines where on the local filesystem an DFS data node @@ -106,15 +89,6 @@ --> <property> - <name>dfs.checksum.type</name> - <value>CRC32</value> - <description>The checksum method to be used by default. To maintain - compatibility, it is being set to CRC32. Once all migration steps - are complete, we can change it to CRC32C and take advantage of the - additional performance benefit.</description> - </property> - - <property> <name>dfs.replication.max</name> <value>50</value> <description>Maximal block replication. @@ -209,13 +183,6 @@ If the port is 0 then the server will start on a free port. </property> <property> -<name>dfs.datanode.du.pct</name> -<value>0.85f</value> -<description>When calculating remaining space, only use this percentage of the real available space -</description> -</property> - -<property> <name>dfs.namenode.handler.count</name> <value>40</value> <description>The number of server threads for the namenode.</description> @@ -238,15 +205,6 @@ The octal umask used when creating files and directories. </property> <property> -<name>dfs.web.ugi</name> -<!-- cluster variant --> -<value>gopher,gopher</value> -<description>The user account used by the web interface. -Syntax: USERNAME,GROUP1,GROUP2, ... -</description> -</property> - -<property> <name>dfs.permissions.enabled</name> <value>true</value> <description> @@ -259,7 +217,7 @@ owner or group of files or directories. </property> <property> -<name>dfs.permissions.enabled.superusergroup</name> +<name>dfs.permissions.superusergroup</name> <value>hdfs</value> <description>The name of the group of super-users.</description> </property> @@ -271,10 +229,6 @@ owner or group of files or directories. </property> <property> -<name>ipc.server.max.response.size</name> -<value>5242880</value> -</property> -<property> <name>dfs.block.access.token.enable</name> <value>true</value> <description> @@ -325,12 +279,6 @@ Kerberos principal name for the NameNode </property> <property> - <name>dfs.secondary.https.port</name> - <value>50490</value> - <description>The https port where secondary-namenode binds</description> - </property> - - <property> <name>dfs.web.authentication.kerberos.principal</name> <value></value> <description> @@ -382,13 +330,6 @@ Kerberos principal name for the NameNode </property> <property> - <name>dfs.https.port</name> - <value>50470</value> - <description>The https port where namenode binds</description> - - </property> - - <property> <name>dfs.namenode.https-address</name> <value></value> <description>The https address where namenode binds</description> @@ -420,12 +361,6 @@ don't exist, they will be created with this permission.</description> </property> <property> - <name>ipc.server.read.threadpool.size</name> - <value>5</value> - <description></description> - </property> - - <property> <name>dfs.namenode.avoid.read.stale.datanode</name> <value>true</value> <description> http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/06d53ed9/ambari-server/src/main/resources/stacks/HDPLocal/2.0.5/services/MAPREDUCE2/configuration/mapred-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDPLocal/2.0.5/services/MAPREDUCE2/configuration/mapred-site.xml b/ambari-server/src/main/resources/stacks/HDPLocal/2.0.5/services/MAPREDUCE2/configuration/mapred-site.xml index 53b7fb7..28072f7 100644 --- a/ambari-server/src/main/resources/stacks/HDPLocal/2.0.5/services/MAPREDUCE2/configuration/mapred-site.xml +++ b/ambari-server/src/main/resources/stacks/HDPLocal/2.0.5/services/MAPREDUCE2/configuration/mapred-site.xml @@ -31,12 +31,6 @@ </property> <property> - <name>io.sort.record.percent</name> - <value>.2</value> - <description>No description</description> - </property> - - <property> <name>mapreduce.map.sort.spill.percent</name> <value>0.1</value> <description>No description</description> @@ -178,17 +172,6 @@ </property> <property> - <name>jetty.connector</name> - <value>org.mortbay.jetty.nio.SelectChannelConnector</value> - <description>No description</description> - </property> - - <property> - <name>mapred.child.root.logger</name> - <value>INFO,TLA</value> - </property> - - <property> <name>mapred.child.java.opts</name> <value>-Xmx512m</value> <description>No description</description> @@ -228,32 +211,11 @@ </property> <property> - <name>mapred.task.maxvmem</name> - <value></value> - <final>true</final> - <description>No description</description> - </property> - - <property> - <name>mapreduce.fileoutputcommitter.marksuccessfuljobs</name> - <value>false</value> - </property> - - <property> <name>mapreduce.tasktracker.keytab.file</name> <value></value> <description>The filename of the keytab for the task tracker</description> </property> - <property> - <name>mapreduce.jobtracker.split.metainfo.maxsize</name> - <value>50000000</value> - <final>true</final> - <description>If the size of the split metainfo file is larger than this, the JobTracker will fail the job during - initialize. - </description> - </property> - <property> <name>mapreduce.jobhistory.keytab.file</name> <!-- cluster variant --> http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/06d53ed9/ambari-server/src/main/resources/stacks/HDPLocal/2.0.5/services/YARN/configuration/yarn-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDPLocal/2.0.5/services/YARN/configuration/yarn-site.xml b/ambari-server/src/main/resources/stacks/HDPLocal/2.0.5/services/YARN/configuration/yarn-site.xml index fed4933..ca49869 100644 --- a/ambari-server/src/main/resources/stacks/HDPLocal/2.0.5/services/YARN/configuration/yarn-site.xml +++ b/ambari-server/src/main/resources/stacks/HDPLocal/2.0.5/services/YARN/configuration/yarn-site.xml @@ -101,11 +101,6 @@ </property> <property> - <name>yarn.nodemanager.aux-services.class</name> - <value>org.apache.hadoop.mapred.ShuffleHandler</value> - </property> - - <property> <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name> <value>org.apache.hadoop.mapred.ShuffleHandler</value> </property>
