http://git-wip-us.apache.org/repos/asf/ambari/blob/d845449a/ambari-server/src/main/resources/stacks/HDP/3.0/services/HDFS/configuration/widgets.json ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/3.0/services/HDFS/configuration/widgets.json b/ambari-server/src/main/resources/stacks/HDP/3.0/services/HDFS/configuration/widgets.json deleted file mode 100644 index 4a645b0..0000000 --- a/ambari-server/src/main/resources/stacks/HDP/3.0/services/HDFS/configuration/widgets.json +++ /dev/null @@ -1,649 +0,0 @@ -{ - "layouts": [ - { - "layout_name": "default_hdfs_dashboard", - "display_name": "Standard HDFS Dashboard", - "section_name": "HDFS_SUMMARY", - "widgetLayoutInfo": [ - { - "widget_name": "NameNode GC count", - "description": "Count of total garbage collections and count of major type garbage collections of the JVM.", - "widget_type": "GRAPH", - "is_visible": true, - "metrics": [ - { - "name": "jvm.JvmMetrics.GcCount._rate", - "metric_path": "metrics/jvm/gcCount._rate", - "service_name": "HDFS", - "component_name": "NAMENODE", - "host_component_criteria": "host_components/metrics/dfs/FSNamesystem/HAState=active" - }, - { - "name": "jvm.JvmMetrics.GcCountConcurrentMarkSweep._rate", - "metric_path": "metrics/jvm/GcCountConcurrentMarkSweep._rate", - "service_name": "HDFS", - "component_name": "NAMENODE", - "host_component_criteria": "host_components/metrics/dfs/FSNamesystem/HAState=active" - } - ], - "values": [ - { - "name": "GC total count", - "value": "${jvm.JvmMetrics.GcCount._rate}" - }, - { - "name": "GC count of type major collection", - "value": "${jvm.JvmMetrics.GcCountConcurrentMarkSweep._rate}" - } - ], - "properties": { - "graph_type": "LINE", - "time_range": "1" - } - }, - { - "widget_name": "NameNode GC time", - "description": "Total time taken by major type garbage collections in milliseconds.", - "widget_type": "GRAPH", - "is_visible": true, - "metrics": [ - { - "name": "jvm.JvmMetrics.GcTimeMillisConcurrentMarkSweep._rate", - "metric_path": "metrics/jvm/GcTimeMillisConcurrentMarkSweep._rate", - "service_name": "HDFS", - "component_name": "NAMENODE", - "host_component_criteria": "host_components/metrics/dfs/FSNamesystem/HAState=active" - } - ], - "values": [ - { - "name": "GC time in major collection", - "value": "${jvm.JvmMetrics.GcTimeMillisConcurrentMarkSweep._rate}" - } - ], - "properties": { - "display_unit": "ms", - "graph_type": "LINE", - "time_range": "1" - } - }, - { - "widget_name": "NN Connection Load", - "description": "Number of open RPC connections being managed by NameNode.", - "widget_type": "GRAPH", - "is_visible": true, - "metrics": [ - { - "name": "rpc.rpc.client.NumOpenConnections", - "metric_path": "metrics/rpc/client/NumOpenConnections", - "category": "", - "service_name": "HDFS", - "component_name": "NAMENODE", - "host_component_criteria": "host_components/metrics/dfs/FSNamesystem/HAState=active" - }, - { - "name": "rpc.rpc.datanode.NumOpenConnections", - "metric_path": "metrics/rpc/datanode/NumOpenConnections", - "category": "", - "service_name": "HDFS", - "component_name": "NAMENODE", - "host_component_criteria": "host_components/metrics/dfs/FSNamesystem/HAState=active" - } - ], - "values": [ - { - "name": "Open Client Connections", - "value": "${rpc.rpc.client.NumOpenConnections}" - }, - { - "name": "Open Datanode Connections", - "value": "${rpc.rpc.datanode.NumOpenConnections}" - } - ], - "properties": { - "graph_type": "LINE", - "time_range": "1" - } - }, - { - "widget_name": "NameNode Heap", - "description": "Heap memory committed and Heap memory used with respect to time.", - "widget_type": "GRAPH", - "is_visible": true, - "metrics": [ - { - "name": "jvm.JvmMetrics.MemHeapCommittedM", - "metric_path": "metrics/jvm/memHeapCommittedM", - "service_name": "HDFS", - "component_name": "NAMENODE", - "host_component_criteria": "host_components/metrics/dfs/FSNamesystem/HAState=active" - }, - { - "name": "jvm.JvmMetrics.MemHeapUsedM", - "metric_path": "metrics/jvm/memHeapUsedM", - "service_name": "HDFS", - "component_name": "NAMENODE", - "host_component_criteria": "host_components/metrics/dfs/FSNamesystem/HAState=active" - } - ], - "values": [ - { - "name": "JVM heap committed", - "value": "${jvm.JvmMetrics.MemHeapCommittedM}" - }, - { - "name": "JVM heap used", - "value": "${jvm.JvmMetrics.MemHeapUsedM}" - } - ], - "properties": { - "display_unit": "MB", - "graph_type": "LINE", - "time_range": "1" - } - }, - { - "widget_name": "NameNode Host Load", - "description": "Percentage of CPU and Memory resources being consumed on NameNode host.", - "widget_type": "GRAPH", - "is_visible": true, - "metrics": [ - { - "name": "cpu_system", - "metric_path": "metrics/cpu/cpu_system", - "service_name": "HDFS", - "component_name": "NAMENODE", - "host_component_criteria": "host_components/metrics/dfs/FSNamesystem/HAState=active" - }, - { - "name": "cpu_user", - "metric_path": "metrics/cpu/cpu_user", - "service_name": "HDFS", - "component_name": "NAMENODE", - "host_component_criteria": "host_components/metrics/dfs/FSNamesystem/HAState=active" - }, - { - "name": "cpu_nice", - "metric_path": "metrics/cpu/cpu_nice", - "service_name": "HDFS", - "component_name": "NAMENODE", - "host_component_criteria": "host_components/metrics/dfs/FSNamesystem/HAState=active" - }, - { - "name": "cpu_idle", - "metric_path": "metrics/cpu/cpu_idle", - "service_name": "HDFS", - "component_name": "NAMENODE", - "host_component_criteria": "host_components/metrics/dfs/FSNamesystem/HAState=active" - }, - { - "name": "cpu_wio", - "metric_path": "metrics/cpu/cpu_wio", - "service_name": "HDFS", - "component_name": "NAMENODE", - "host_component_criteria": "host_components/metrics/dfs/FSNamesystem/HAState=active" - }, - { - "name": "mem_total", - "metric_path": "metrics/memory/mem_total", - "service_name": "HDFS", - "component_name": "NAMENODE", - "host_component_criteria": "host_components/metrics/dfs/FSNamesystem/HAState=active" - }, - { - "name": "mem_free", - "metric_path": "metrics/memory/mem_free", - "service_name": "HDFS", - "component_name": "NAMENODE", - "host_component_criteria": "host_components/metrics/dfs/FSNamesystem/HAState=active" - } - ], - "values": [ - { - "name": "CPU utilization", - "value": "${((cpu_system + cpu_user + cpu_nice)/(cpu_system + cpu_user + cpu_nice + cpu_idle + cpu_wio)) * 100}" - }, - { - "name": "Memory utilization", - "value": "${((mem_total - mem_free)/mem_total) * 100}" - } - ], - "properties": { - "graph_type": "LINE", - "time_range": "1", - "display_unit": "%" - } - }, - { - "widget_name": "NameNode RPC", - "description": "Compares the average time spent for RPC request in a queue and RPC request being processed.", - "widget_type": "GRAPH", - "is_visible": true, - "metrics": [ - { - "name": "rpc.rpc.client.RpcQueueTimeAvgTime", - "metric_path": "metrics/rpc/client/RpcQueueTime_avg_time", - "service_name": "HDFS", - "component_name": "NAMENODE", - "host_component_criteria": "host_components/metrics/dfs/FSNamesystem/HAState=active" - }, - { - "name": "rpc.rpc.client.RpcProcessingTimeAvgTime", - "metric_path": "metrics/rpc/client/RpcProcessingTime_avg_time", - "service_name": "HDFS", - "component_name": "NAMENODE", - "host_component_criteria": "host_components/metrics/dfs/FSNamesystem/HAState=active" - }, - { - "name": "rpc.rpc.datanode.RpcQueueTimeAvgTime", - "metric_path": "metrics/rpc/datanode/RpcQueueTime_avg_time", - "service_name": "HDFS", - "component_name": "NAMENODE", - "host_component_criteria": "host_components/metrics/dfs/FSNamesystem/HAState=active" - }, - { - "name": "rpc.rpc.datanode.RpcProcessingTimeAvgTime", - "metric_path": "metrics/rpc/datanode/RpcProcessingTime_avg_time", - "service_name": "HDFS", - "component_name": "NAMENODE", - "host_component_criteria": "host_components/metrics/dfs/FSNamesystem/HAState=active" - } - ], - "values": [ - { - "name": "Client RPC Queue Wait time", - "value": "${rpc.rpc.client.RpcQueueTimeAvgTime}" - }, - { - "name": "Client RPC Processing time", - "value": "${rpc.rpc.client.RpcProcessingTimeAvgTime}" - }, - { - "name": "Datanode RPC Queue Wait time", - "value": "${rpc.rpc.datanode.RpcQueueTimeAvgTime}" - }, - { - "name": "Datanode RPC Processing time", - "value": "${rpc.rpc.datanode.RpcProcessingTimeAvgTime}" - } - ], - "properties": { - "graph_type": "LINE", - "time_range": "1", - "display_unit": "ms" - } - }, - { - "widget_name": "NameNode Operations", - "description": "Rate per second of number of file operation over time.", - "widget_type": "GRAPH", - "is_visible": false, - "metrics": [ - { - "name": "dfs.namenode.TotalFileOps._rate", - "metric_path": "metrics/dfs/namenode/TotalFileOps._rate", - "service_name": "HDFS", - "component_name": "NAMENODE", - "host_component_criteria": "host_components/metrics/dfs/FSNamesystem/HAState=active" - } - ], - "values": [ - { - "name": "NameNode File Operations", - "value": "${dfs.namenode.TotalFileOps._rate}" - } - ], - "properties": { - "graph_type": "LINE", - "time_range": "1" - } - }, - { - "widget_name": "Failed disk volumes", - "description": "Number of Failed disk volumes across all DataNodes. Its indicative of HDFS bad health.", - "widget_type": "NUMBER", - "is_visible": true, - "metrics": [ - { - "name": "FSDatasetState.org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.NumFailedVolumes._sum", - "metric_path": "metrics/dfs/datanode/NumFailedVolumes", - "service_name": "HDFS", - "component_name": "DATANODE" - } - ], - "values": [ - { - "name": "Failed disk volumes", - "value": "${FSDatasetState.org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.NumFailedVolumes._sum}" - } - ], - "properties": { - "display_unit": "" - } - }, - { - "widget_name": "Blocks With Corrupted Replicas", - "description": "Number represents data blocks with at least one corrupted replica (but not all of them). Its indicative of HDFS bad health.", - "widget_type": "NUMBER", - "is_visible": true, - "metrics": [ - { - "name": "Hadoop:service=NameNode,name=FSNamesystem.CorruptBlocks", - "metric_path": "metrics/dfs/FSNamesystem/CorruptBlocks", - "service_name": "HDFS", - "component_name": "NAMENODE", - "host_component_criteria": "host_components/metrics/dfs/FSNamesystem/HAState=active" - } - ], - "values": [ - { - "name": "Blocks With Corrupted Replicas", - "value": "${Hadoop:service=NameNode,name=FSNamesystem.CorruptBlocks}" - } - ], - "properties": { - "warning_threshold": "0", - "error_threshold": "50" - } - }, - { - "widget_name": "Under Replicated Blocks", - "description": "Number represents file blocks that does not meet the replication factor criteria. Its indicative of HDFS bad health.", - "widget_type": "NUMBER", - "is_visible": true, - "metrics": [ - { - "name": "Hadoop:service=NameNode,name=FSNamesystem.UnderReplicatedBlocks", - "metric_path": "metrics/dfs/FSNamesystem/UnderReplicatedBlocks", - "service_name": "HDFS", - "component_name": "NAMENODE", - "host_component_criteria": "host_components/metrics/dfs/FSNamesystem/HAState=active" - } - ], - "values": [ - { - "name": "Under Replicated Blocks", - "value": "${Hadoop:service=NameNode,name=FSNamesystem.UnderReplicatedBlocks}" - } - ], - "properties": { - "warning_threshold": "0", - "error_threshold": "50" - } - }, - { - "widget_name": "HDFS Space Utilization", - "description": "Percentage of available space used in the DFS.", - "widget_type": "GAUGE", - "is_visible": true, - "metrics": [ - { - "name": "FSDatasetState.org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.Remaining", - "metric_path": "metrics/FSDatasetState/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl/Remaining", - "service_name": "HDFS", - "component_name": "DATANODE" - }, - { - "name": "FSDatasetState.org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.Capacity", - "metric_path": "metrics/dfs/datanode/Capacity", - "service_name": "HDFS", - "component_name": "DATANODE" - } - ], - "values": [ - { - "name": "HDFS Space Utilization", - "value": "${(FSDatasetState.org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.Capacity - FSDatasetState.org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.Remaining)/FSDatasetState.org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.Capacity}" - } - ], - "properties": { - "warning_threshold": "0.75", - "error_threshold": "0.9" - } - } - ] - }, - { - "layout_name": "default_hdfs_heatmap", - "section_name": "HDFS_HEATMAPS", - "display_name": "HDFS Heatmaps", - "widgetLayoutInfo": [ - { - "widget_name": "HDFS Bytes Read", - "default_section_name": "HDFS_HEATMAPS", - "description": "", - "widget_type": "HEATMAP", - "is_visible": true, - "metrics": [ - { - "name": "dfs.datanode.BytesRead._rate", - "metric_path": "metrics/dfs/datanode/bytes_read._rate", - "service_name": "HDFS", - "component_name": "DATANODE" - } - ], - "values": [ - { - "name": "HDFS Bytes Read", - "value": "${dfs.datanode.BytesRead._rate}" - } - ], - "properties": { - "display_unit": "MB", - "max_limit": "1024" - } - }, - { - "widget_name": "HDFS Bytes Written", - "description": "", - "widget_type": "HEATMAP", - "is_visible": false, - "metrics": [ - { - "name": "dfs.datanode.BytesWritten._rate", - "metric_path": "metrics/dfs/datanode/bytes_written._rate", - "service_name": "HDFS", - "component_name": "DATANODE" - } - ], - "values": [ - { - "name": "HDFS Bytes Written", - "value": "${dfs.datanode.BytesWritten._rate}" - } - ], - "properties": { - "display_unit": "MB", - "max_limit": "1024" - } - }, - { - "widget_name": "DataNode Garbage Collection Time", - "description": "", - "widget_type": "HEATMAP", - "is_visible": false, - "metrics": [ - { - "name": "Hadoop:service=DataNode,name=JvmMetrics.GcTimeMillis", - "metric_path": "metrics/jvm/gcTimeMillis", - "service_name": "HDFS", - "component_name": "DATANODE" - } - ], - "values": [ - { - "name": "DataNode Garbage Collection Time", - "value": "${Hadoop:service=DataNode,name=JvmMetrics.GcTimeMillis}" - } - ], - "properties": { - "display_unit": "ms", - "max_limit": "10000" - } - }, - { - "widget_name": "DataNode JVM Heap Memory Used", - "description": "", - "widget_type": "HEATMAP", - "is_visible": false, - "metrics": [ - { - "name": "Hadoop:service=DataNode,name=JvmMetrics.MemHeapUsedM", - "metric_path": "metrics/jvm/memHeapUsedM", - "service_name": "HDFS", - "component_name": "DATANODE" - } - ], - "values": [ - { - "name": "DataNode JVM Heap Memory Used", - "value": "${Hadoop:service=DataNode,name=JvmMetrics.MemHeapUsedM}" - } - ], - "properties": { - "display_unit": "MB", - "max_limit": "512" - } - }, - { - "widget_name": "DataNode JVM Heap Memory Committed", - "description": "", - "widget_type": "HEATMAP", - "is_visible": false, - "metrics": [ - { - "name": "Hadoop:service=DataNode,name=JvmMetrics.MemHeapCommittedM", - "metric_path": "metrics/jvm/memHeapCommittedM", - "service_name": "HDFS", - "component_name": "DATANODE" - } - ], - "values": [ - { - "name": "DataNode JVM Heap Memory Committed", - "value": "${Hadoop:service=DataNode,name=JvmMetrics.MemHeapCommittedM}" - } - ], - "properties": { - "display_unit": "MB", - "max_limit": "512" - } - }, - { - "widget_name": "DataNode Process Disk I/O Utilization", - "default_section_name": "HDFS_HEATMAPS", - "description": "", - "widget_type": "HEATMAP", - "is_visible": false, - "metrics": [ - { - "name": "dfs.datanode.BytesRead._rate", - "metric_path": "metrics/dfs/datanode/bytes_read._rate", - "service_name": "HDFS", - "component_name": "DATANODE" - }, - { - "name": "dfs.datanode.BytesWritten._rate", - "metric_path": "metrics/dfs/datanode/bytes_written._rate", - "service_name": "HDFS", - "component_name": "DATANODE" - }, - { - "name": "dfs.datanode.TotalReadTime._rate", - "metric_path": "metrics/dfs/datanode/TotalReadTime._rate", - "service_name": "HDFS", - "component_name": "DATANODE" - }, - { - "name": "dfs.datanode.TotalWriteTime._rate", - "metric_path": "metrics/dfs/datanode/TotalWriteTime._rate", - "service_name": "HDFS", - "component_name": "DATANODE" - } - ], - "values": [ - { - "name": "DataNode Process Disk I/O Utilization", - "value": "${((dfs.datanode.BytesRead._rate/dfs.datanode.TotalReadTime._rate)+(dfs.datanode.BytesWritten._rate/dfs.datanode.TotalWriteTime._rate))*50}" - } - ], - "properties": { - "display_unit": "%", - "max_limit": "100" - } - }, - { - "widget_name": "DataNode Process Network I/O Utilization", - "description": "", - "widget_type": "HEATMAP", - "is_visible": false, - "metrics": [ - { - "name": "dfs.datanode.RemoteBytesRead._rate", - "metric_path": "metrics/dfs/datanode/RemoteBytesRead._rate", - "service_name": "HDFS", - "component_name": "DATANODE" - }, - { - "name": "dfs.datanode.ReadsFromRemoteClient._rate", - "metric_path": "metrics/dfs/datanode/reads_from_remote_client._rate", - "service_name": "HDFS", - "component_name": "DATANODE" - }, - { - "name": "dfs.datanode.RemoteBytesWritten._rate", - "metric_path": "metrics/dfs/datanode/RemoteBytesWritten._rate", - "service_name": "HDFS", - "component_name": "DATANODE" - }, - { - "name": "dfs.datanode.WritesFromRemoteClient._rate", - "metric_path": "metrics/dfs/datanode/writes_from_remote_client._rate", - "service_name": "HDFS", - "component_name": "DATANODE" - } - ], - "values": [ - { - "name": "DataNode Process Network I/O Utilization", - "value": "${((dfs.datanode.RemoteBytesRead._rate/dfs.datanode.ReadsFromRemoteClient._rate)+(dfs.datanode.RemoteBytesWritten._rate/dfs.datanode.WritesFromRemoteClient._rate))*50}" - } - ], - "properties": { - "display_unit": "%", - "max_limit": "100" - } - }, - { - "widget_name": "HDFS Space Utilization", - "widget_type": "HEATMAP", - "is_visible": false, - "metrics": [ - { - "name": "FSDatasetState.org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.Remaining", - "metric_path": "metrics/FSDatasetState/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl/Remaining", - "service_name": "HDFS", - "component_name": "DATANODE" - }, - { - "name": "FSDatasetState.org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.Capacity", - "metric_path": "metrics/dfs/datanode/Capacity", - "service_name": "HDFS", - "component_name": "DATANODE" - } - ], - "values": [ - { - "name": "HDFS Space Utilization", - "value": "${((FSDatasetState.org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.Capacity - FSDatasetState.org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.Remaining)/FSDatasetState.org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.Capacity) * 100}" - } - ], - "properties": { - "display_unit": "%", - "max_limit": "100" - } - } - ] - } - ] -}
http://git-wip-us.apache.org/repos/asf/ambari/blob/d845449a/ambari-server/src/main/resources/stacks/HDP/3.0/services/HDFS/kerberos.json ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/3.0/services/HDFS/kerberos.json b/ambari-server/src/main/resources/stacks/HDP/3.0/services/HDFS/kerberos.json deleted file mode 100644 index 9000e95..0000000 --- a/ambari-server/src/main/resources/stacks/HDP/3.0/services/HDFS/kerberos.json +++ /dev/null @@ -1,246 +0,0 @@ -{ - "services": [ - { - "name": "HDFS", - "identities": [ - { - "name": "/spnego", - "principal": { - "configuration": "hdfs-site/dfs.web.authentication.kerberos.principal" - }, - "keytab": { - "configuration": "hdfs-site/dfs.web.authentication.kerberos.keytab" - } - }, - { - "name": "/smokeuser" - } - ], - "auth_to_local_properties" : [ - "core-site/hadoop.security.auth_to_local" - ], - "configurations": [ - { - "core-site": { - "hadoop.security.authentication": "kerberos", - "hadoop.security.authorization": "true", - "hadoop.proxyuser.HTTP.groups": "${hadoop-env/proxyuser_group}" - } - }, - { - "ranger-hdfs-audit": { - "xasecure.audit.jaas.Client.loginModuleName": "com.sun.security.auth.module.Krb5LoginModule", - "xasecure.audit.jaas.Client.loginModuleControlFlag": "required", - "xasecure.audit.jaas.Client.option.useKeyTab": "true", - "xasecure.audit.jaas.Client.option.storeKey": "false", - "xasecure.audit.jaas.Client.option.serviceName": "solr", - "xasecure.audit.destination.solr.force.use.inmemory.jaas.config": "true" - } - } - ], - "components": [ - { - "name": "HDFS_CLIENT", - "identities": [ - { - "name": "/HDFS/NAMENODE/hdfs" - } - ] - }, - { - "name": "NAMENODE", - "identities": [ - { - "name": "hdfs", - "principal": { - "value": "${hadoop-env/hdfs_user}-${cluster_name|toLower()}@${realm}", - "type" : "user" , - "configuration": "hadoop-env/hdfs_principal_name", - "local_username" : "${hadoop-env/hdfs_user}" - }, - "keytab": { - "file": "${keytab_dir}/hdfs.headless.keytab", - "owner": { - "name": "${hadoop-env/hdfs_user}", - "access": "r" - }, - "group": { - "name": "${cluster-env/user_group}", - "access": "" - }, - "configuration": "hadoop-env/hdfs_user_keytab" - } - }, - { - "name": "namenode_nn", - "principal": { - "value": "nn/_HOST@${realm}", - "type" : "service", - "configuration": "hdfs-site/dfs.namenode.kerberos.principal", - "local_username" : "${hadoop-env/hdfs_user}" - }, - "keytab": { - "file": "${keytab_dir}/nn.service.keytab", - "owner": { - "name": "${hadoop-env/hdfs_user}", - "access": "r" - }, - "group": { - "name": "${cluster-env/user_group}", - "access": "" - }, - "configuration": "hdfs-site/dfs.namenode.keytab.file" - } - }, - { - "name": "/spnego", - "principal": { - "configuration": "hdfs-site/dfs.namenode.kerberos.internal.spnego.principal" - } - }, - { - "name": "/HDFS/NAMENODE/namenode_nn", - "principal": { - "configuration": "ranger-hdfs-audit/xasecure.audit.jaas.Client.option.principal" - }, - "keytab": { - "configuration": "ranger-hdfs-audit/xasecure.audit.jaas.Client.option.keyTab" - } - } - ], - "configurations": [ - { - "hdfs-site": { - "dfs.block.access.token.enable": "true" - } - } - ] - }, - { - "name": "DATANODE", - "identities": [ - { - "name": "datanode_dn", - "principal": { - "value": "dn/_HOST@${realm}", - "type" : "service", - "configuration": "hdfs-site/dfs.datanode.kerberos.principal", - "local_username" : "${hadoop-env/hdfs_user}" - }, - "keytab": { - "file": "${keytab_dir}/dn.service.keytab", - "owner": { - "name": "${hadoop-env/hdfs_user}", - "access": "r" - }, - "group": { - "name": "${cluster-env/user_group}", - "access": "" - }, - "configuration": "hdfs-site/dfs.datanode.keytab.file" - } - } - ], - "configurations" : [ - { - "hdfs-site" : { - "dfs.datanode.address" : "0.0.0.0:1019", - "dfs.datanode.http.address": "0.0.0.0:1022" - } - } - ] - }, - { - "name": "SECONDARY_NAMENODE", - "identities": [ - { - "name": "secondary_namenode_nn", - "principal": { - "value": "nn/_HOST@${realm}", - "type" : "service", - "configuration": "hdfs-site/dfs.secondary.namenode.kerberos.principal", - "local_username" : "${hadoop-env/hdfs_user}" - }, - "keytab": { - "file": "${keytab_dir}/nn.service.keytab", - "owner": { - "name": "${hadoop-env/hdfs_user}", - "access": "r" - }, - "group": { - "name": "${cluster-env/user_group}", - "access": "" - }, - "configuration": "hdfs-site/dfs.secondary.namenode.keytab.file" - } - }, - { - "name": "/spnego", - "principal": { - "configuration": "hdfs-site/dfs.secondary.namenode.kerberos.internal.spnego.principal" - } - } - ] - }, - { - "name": "NFS_GATEWAY", - "identities": [ - { - "name": "nfsgateway", - "principal": { - "value": "nfs/_HOST@${realm}", - "type" : "service", - "configuration": "hdfs-site/nfs.kerberos.principal", - "local_username" : "${hadoop-env/hdfs_user}" - }, - "keytab": { - "file": "${keytab_dir}/nfs.service.keytab", - "owner": { - "name": "${hadoop-env/hdfs_user}", - "access": "r" - }, - "group": { - "name": "${cluster-env/user_group}", - "access": "" - }, - "configuration": "hdfs-site/nfs.keytab.file" - } - } - ] - }, - { - "name": "JOURNALNODE", - "identities": [ - { - "name": "journalnode_jn", - "principal": { - "value": "jn/_HOST@${realm}", - "type" : "service", - "configuration": "hdfs-site/dfs.journalnode.kerberos.principal", - "local_username" : "${hadoop-env/hdfs_user}" - }, - "keytab": { - "file": "${keytab_dir}/jn.service.keytab", - "owner": { - "name": "${hadoop-env/hdfs_user}", - "access": "r" - }, - "group": { - "name": "${cluster-env/user_group}", - "access": "" - }, - "configuration": "hdfs-site/dfs.journalnode.keytab.file" - } - }, - { - "name": "/spnego", - "principal": { - "configuration": "hdfs-site/dfs.journalnode.kerberos.internal.spnego.principal" - } - } - ] - } - ] - } - ] -} http://git-wip-us.apache.org/repos/asf/ambari/blob/d845449a/ambari-server/src/main/resources/stacks/HDP/3.0/services/HDFS/metainfo.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/3.0/services/HDFS/metainfo.xml b/ambari-server/src/main/resources/stacks/HDP/3.0/services/HDFS/metainfo.xml index 30c49c7..ef2027f 100644 --- a/ambari-server/src/main/resources/stacks/HDP/3.0/services/HDFS/metainfo.xml +++ b/ambari-server/src/main/resources/stacks/HDP/3.0/services/HDFS/metainfo.xml @@ -21,39 +21,8 @@ <service> <name>HDFS</name> <displayName>HDFS</displayName> - <comment>Apache Hadoop Distributed File System</comment> - <version>2.7.1.3.0</version> - <extends>common-services/HDFS/2.1.0.2.0</extends> - - <components> - <!-- NFS Gateway was added in HDP 2.3. --> - <component> - <name>NFS_GATEWAY</name> - <displayName>NFSGateway</displayName> - <cardinality>0+</cardinality> - <versionAdvertised>true</versionAdvertised> - <category>SLAVE</category> - <commandScript> - <script>scripts/nfsgateway.py</script> - <scriptType>PYTHON</scriptType> - <timeout>1200</timeout> - </commandScript> - <dependencies> - <dependency> - <name>HDFS/HDFS_CLIENT</name> - <scope>host</scope> - <auto-deploy> - <enabled>true</enabled> - </auto-deploy> - </dependency> - </dependencies> - </component> - - <component> - <name>ZKFC</name> - <versionAdvertised>true</versionAdvertised> - </component> - </components> + <version>3.0.0.3.0</version> + <extends>common-services/HDFS/3.0.0</extends> <osSpecifics> <osSpecific> @@ -171,20 +140,6 @@ </osSpecific> </osSpecifics> - <quickLinksConfigurations> - <quickLinksConfiguration> - <fileName>quicklinks.json</fileName> - <default>true</default> - </quickLinksConfiguration> - </quickLinksConfigurations> - - <themes> - <theme> - <fileName>theme.json</fileName> - <default>true</default> - </theme> - </themes> - </service> </services> </metainfo> http://git-wip-us.apache.org/repos/asf/ambari/blob/d845449a/ambari-server/src/main/resources/stacks/HDP/3.0/services/HDFS/quicklinks/quicklinks.json ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/3.0/services/HDFS/quicklinks/quicklinks.json b/ambari-server/src/main/resources/stacks/HDP/3.0/services/HDFS/quicklinks/quicklinks.json deleted file mode 100644 index 5318ba0..0000000 --- a/ambari-server/src/main/resources/stacks/HDP/3.0/services/HDFS/quicklinks/quicklinks.json +++ /dev/null @@ -1,80 +0,0 @@ -{ - "name": "default", - "description": "default quick links configuration", - "configuration": { - "protocol": - { - "type":"https", - "checks":[ - { - "property":"dfs.http.policy", - "desired":"HTTPS_ONLY", - "site":"hdfs-site" - } - ] - }, - - "links": [ - { - "name": "namenode_ui", - "label": "NameNode UI", - "component_name": "NAMENODE", - "url":"%@://%@:%@", - "requires_user_name": "false", - "port":{ - "http_property": "dfs.namenode.http-address", - "http_default_port": "50070", - "https_property": "dfs.namenode.https-address", - "https_default_port": "50470", - "regex": "\\w*:(\\d+)", - "site": "hdfs-site" - } - }, - { - "name": "namenode_logs", - "label": "NameNode Logs", - "component_name": "NAMENODE", - "url":"%@://%@:%@/logs", - "requires_user_name": "false", - "port":{ - "http_property": "dfs.namenode.http-address", - "http_default_port": "50070", - "https_property": "dfs.namenode.https-address", - "https_default_port": "50470", - "regex": "\\w*:(\\d+)", - "site": "hdfs-site" - } - }, - { - "name": "namenode_jmx", - "label": "NameNode JMX", - "component_name": "NAMENODE", - "url":"%@://%@:%@/jmx", - "requires_user_name": "false", - "port":{ - "http_property": "dfs.namenode.http-address", - "http_default_port": "50070", - "https_property": "dfs.namenode.https-address", - "https_default_port": "50470", - "regex": "\\w*:(\\d+)", - "site": "hdfs-site" - } - }, - { - "name": "Thread Stacks", - "label": "Thread Stacks", - "component_name": "NAMENODE", - "url":"%@://%@:%@/stacks", - "requires_user_name": "false", - "port":{ - "http_property": "dfs.namenode.http-address", - "http_default_port": "50070", - "https_property": "dfs.namenode.https-address", - "https_default_port": "50470", - "regex": "\\w*:(\\d+)", - "site": "hdfs-site" - } - } - ] - } -} \ No newline at end of file http://git-wip-us.apache.org/repos/asf/ambari/blob/d845449a/ambari-server/src/main/resources/stacks/HDP/3.0/services/HDFS/themes/theme.json ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/3.0/services/HDFS/themes/theme.json b/ambari-server/src/main/resources/stacks/HDP/3.0/services/HDFS/themes/theme.json deleted file mode 100644 index 6f2b797..0000000 --- a/ambari-server/src/main/resources/stacks/HDP/3.0/services/HDFS/themes/theme.json +++ /dev/null @@ -1,179 +0,0 @@ -{ - "name": "default", - "description": "Default theme for HDFS service", - "configuration": { - "layouts": [ - { - "name": "default", - "tabs": [ - { - "name": "settings", - "display-name": "Settings", - "layout": { - "tab-columns": "2", - "tab-rows": "1", - "sections": [ - { - "name": "section-namenode", - "display-name": "NameNode", - "row-index": "0", - "column-index": "0", - "row-span": "1", - "column-span": "1", - "section-columns": "1", - "section-rows": "1", - "subsections": [ - { - "name": "subsection-namenode-col1", - "row-index": "0", - "column-index": "0", - "row-span": "1", - "column-span": "1" - } - ] - }, - { - "name": "section-datanode", - "display-name": "DataNode", - "row-index": "0", - "column-index": "1", - "row-span": "1", - "column-span": "1", - "section-columns": "1", - "section-rows": "1", - "subsections": [ - { - "name": "subsection-datanode-col1", - "row-index": "0", - "column-index": "0", - "row-span": "1", - "column-span": "1" - } - ] - } - ] - } - } - ] - } - ], - "placement": { - "configuration-layout": "default", - "configs": [ - { - "config": "hdfs-site/dfs.namenode.name.dir", - "subsection-name": "subsection-namenode-col1" - }, - { - "config": "hadoop-env/namenode_heapsize", - "subsection-name": "subsection-namenode-col1" - }, - { - "config": "hdfs-site/dfs.namenode.handler.count", - "subsection-name": "subsection-namenode-col1" - }, - { - "config": "hdfs-site/dfs.namenode.safemode.threshold-pct", - "subsection-name": "subsection-namenode-col1" - }, - { - "config": "hdfs-site/dfs.datanode.data.dir", - "subsection-name": "subsection-datanode-col1" - }, - { - "config": "hdfs-site/dfs.datanode.failed.volumes.tolerated", - "subsection-name": "subsection-datanode-col1" - }, - { - "config": "hadoop-env/dtnode_heapsize", - "subsection-name": "subsection-datanode-col1" - }, - { - "config": "hdfs-site/dfs.datanode.max.transfer.threads", - "subsection-name": "subsection-datanode-col1" - } - ] - }, - "widgets": [ - { - "config": "hdfs-site/dfs.namenode.name.dir", - "widget": { - "type": "directories" - } - }, - { - "config": "hdfs-site/dfs.namenode.safemode.threshold-pct", - "widget": { - "type": "slider", - "units": [ - { - "unit-name": "percent" - } - ] - } - }, - { - "config": "hdfs-site/dfs.namenode.handler.count", - "widget": { - "type": "slider", - "units": [ - { - "unit-name": "int" - } - ] - } - }, - { - "config": "hadoop-env/namenode_heapsize", - "widget": { - "type": "slider", - "units": [ - { - "unit-name": "GB" - } - ] - } - }, - { - "config": "hdfs-site/dfs.datanode.failed.volumes.tolerated", - "widget": { - "type": "slider", - "units": [ - { - "unit-name": "int" - } - ] - } - }, - { - "config": "hdfs-site/dfs.datanode.data.dir", - "widget": { - "type": "directories" - } - }, - { - "config": "hadoop-env/dtnode_heapsize", - "widget": { - "type": "slider", - "units": [ - { - "unit-name": "GB" - } - ] - } - }, - { - "config": "hdfs-site/dfs.datanode.max.transfer.threads", - "widget": { - "type": "slider", - "units": [ - { - "unit-name": "int" - } - ] - } - } - ] - } -} - http://git-wip-us.apache.org/repos/asf/ambari/blob/d845449a/ambari-server/src/main/resources/stacks/HDP/3.0/services/YARN/YARN_widgets.json ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/3.0/services/YARN/YARN_widgets.json b/ambari-server/src/main/resources/stacks/HDP/3.0/services/YARN/YARN_widgets.json deleted file mode 100644 index 782f21d..0000000 --- a/ambari-server/src/main/resources/stacks/HDP/3.0/services/YARN/YARN_widgets.json +++ /dev/null @@ -1,670 +0,0 @@ -{ - "layouts": [ - { - "layout_name": "default_yarn_dashboard", - "display_name": "Standard YARN Dashboard", - "section_name": "YARN_SUMMARY", - "widgetLayoutInfo": [ - { - "widget_name": "Memory Utilization", - "description": "Percentage of total memory allocated to containers running in the cluster.", - "widget_type": "GRAPH", - "is_visible": true, - "metrics": [ - { - "name": "yarn.QueueMetrics.Queue=root.AllocatedMB", - "metric_path": "metrics/yarn/Queue/root/AllocatedMB", - "service_name": "YARN", - "component_name": "RESOURCEMANAGER", - "host_component_criteria": "host_components/HostRoles/ha_state=ACTIVE" - }, - { - "name": "yarn.QueueMetrics.Queue=root.AvailableMB", - "metric_path": "metrics/yarn/Queue/root/AvailableMB", - "service_name": "YARN", - "component_name": "RESOURCEMANAGER", - "host_component_criteria": "host_components/HostRoles/ha_state=ACTIVE" - } - ], - "values": [ - { - "name": "Memory Utilization", - "value": "${(yarn.QueueMetrics.Queue=root.AllocatedMB / (yarn.QueueMetrics.Queue=root.AllocatedMB + yarn.QueueMetrics.Queue=root.AvailableMB)) * 100}" - } - ], - "properties": { - "display_unit": "%", - "graph_type": "LINE", - "time_range": "1" - } - }, - { - "widget_name": "CPU Utilization", - "description": "Percentage of total virtual cores allocated to containers running in the cluster.", - "widget_type": "GRAPH", - "is_visible": true, - "metrics": [ - { - "name": "yarn.QueueMetrics.Queue=root.AllocatedVCores", - "metric_path": "metrics/yarn/Queue/root/AllocatedVCores", - "service_name": "YARN", - "component_name": "RESOURCEMANAGER", - "host_component_criteria": "host_components/HostRoles/ha_state=ACTIVE" - }, - { - "name": "yarn.QueueMetrics.Queue=root.AvailableVCores", - "metric_path": "metrics/yarn/Queue/root/AvailableVCores", - "service_name": "YARN", - "component_name": "RESOURCEMANAGER", - "host_component_criteria": "host_components/HostRoles/ha_state=ACTIVE" - } - ], - "values": [ - { - "name": "Total Allocatable CPU Utilized across NodeManager", - "value": "${(yarn.QueueMetrics.Queue=root.AllocatedVCores / (yarn.QueueMetrics.Queue=root.AllocatedVCores + yarn.QueueMetrics.Queue=root.AvailableVCores)) * 100}" - } - ], - "properties": { - "display_unit": "%", - "graph_type": "LINE", - "time_range": "1" - } - }, - { - "widget_name": "Bad Local Disks", - "description": "Number of unhealthy local disks across all NodeManagers.", - "widget_type": "NUMBER", - "is_visible": true, - "metrics": [ - { - "name": "yarn.NodeManagerMetrics.BadLocalDirs", - "metric_path": "metrics/yarn/BadLocalDirs", - "service_name": "YARN", - "component_name": "NODEMANAGER" - }, - { - "name": "yarn.NodeManagerMetrics.BadLogDirs", - "metric_path": "metrics/yarn/BadLogDirs", - "service_name": "YARN", - "component_name": "NODEMANAGER" - } - ], - "values": [ - { - "name": "Number of unhealthy local disks for NodeManager", - "value": "${yarn.NodeManagerMetrics.BadLocalDirs + yarn.NodeManagerMetrics.BadLogDirs}" - } - ], - "properties": { - "display_unit": "" - } - }, - { - "widget_name": "Container Failures", - "description": "Percentage of all containers failing in the cluster.", - "widget_type": "GRAPH", - "is_visible": true, - "metrics": [ - { - "name": "yarn.NodeManagerMetrics.ContainersFailed._rate", - "metric_path": "metrics/yarn/ContainersFailed._rate", - "service_name": "YARN", - "component_name": "NODEMANAGER" - }, - { - "name": "yarn.NodeManagerMetrics.ContainersCompleted._rate", - "metric_path": "metrics/yarn/ContainersCompleted._rate", - "service_name": "YARN", - "component_name": "NODEMANAGER" - }, - { - "name": "yarn.NodeManagerMetrics.ContainersLaunched._rate", - "metric_path": "metrics/yarn/ContainersLaunched._rate", - "service_name": "YARN", - "component_name": "NODEMANAGER" - }, - { - "name": "yarn.NodeManagerMetrics.ContainersIniting._sum", - "metric_path": "metrics/yarn/ContainersIniting._sum", - "service_name": "YARN", - "component_name": "NODEMANAGER" - }, - { - "name": "yarn.NodeManagerMetrics.ContainersKilled._rate", - "metric_path": "metrics/yarn/ContainersKilled._rate", - "service_name": "YARN", - "component_name": "NODEMANAGER" - }, - { - "name": "yarn.NodeManagerMetrics.ContainersRunning._sum", - "metric_path": "metrics/yarn/ContainersRunning._sum", - "service_name": "YARN", - "component_name": "NODEMANAGER" - } - ], - "values": [ - { - "name": "Container Failures", - "value": "${(yarn.NodeManagerMetrics.ContainersFailed._rate/(yarn.NodeManagerMetrics.ContainersFailed._rate + yarn.NodeManagerMetrics.ContainersCompleted._rate + yarn.NodeManagerMetrics.ContainersLaunched._rate + yarn.NodeManagerMetrics.ContainersIniting._sum + yarn.NodeManagerMetrics.ContainersKilled._rate + yarn.NodeManagerMetrics.ContainersRunning._sum)) * 100}" - } - ], - "properties": { - "display_unit": "%", - "graph_type": "LINE", - "time_range": "1" - } - }, - { - "widget_name": "App Failures", - "description": "Percentage of all launched applications failing in the cluster.", - "widget_type": "GRAPH", - "is_visible": true, - "metrics": [ - { - "name": "yarn.QueueMetrics.Queue=root.AppsFailed._rate", - "metric_path": "metrics/yarn/Queue/root/AppsFailed._rate", - "service_name": "YARN", - "component_name": "RESOURCEMANAGER", - "host_component_criteria": "host_components/HostRoles/ha_state=ACTIVE" - }, - { - "name": "yarn.QueueMetrics.Queue=root.AppsKilled._rate", - "metric_path": "metrics/yarn/Queue/root/AppsKilled._rate", - "service_name": "YARN", - "component_name": "RESOURCEMANAGER", - "host_component_criteria": "host_components/HostRoles/ha_state=ACTIVE" - }, - { - "name": "yarn.QueueMetrics.Queue=root.AppsPending", - "metric_path": "metrics/yarn/Queue/root/AppsPending", - "service_name": "YARN", - "component_name": "RESOURCEMANAGER", - "host_component_criteria": "host_components/HostRoles/ha_state=ACTIVE" - }, - { - "name": "yarn.QueueMetrics.Queue=root.AppsRunning", - "metric_path": "metrics/yarn/Queue/root/AppsRunning", - "service_name": "YARN", - "component_name": "RESOURCEMANAGER", - "host_component_criteria": "host_components/HostRoles/ha_state=ACTIVE" - }, - { - "name": "yarn.QueueMetrics.Queue=root.AppsSubmitted._rate", - "metric_path": "metrics/yarn/Queue/root/AppsSubmitted._rate", - "service_name": "YARN", - "component_name": "RESOURCEMANAGER", - "host_component_criteria": "host_components/HostRoles/ha_state=ACTIVE" - }, - { - "name": "yarn.QueueMetrics.Queue=root.AppsCompleted._rate", - "metric_path": "metrics/yarn/Queue/root/AppsCompleted._rate", - "service_name": "YARN", - "component_name": "RESOURCEMANAGER", - "host_component_criteria": "host_components/HostRoles/ha_state=ACTIVE" - } - ], - "values": [ - { - "name": "App Failures", - "value": "${(yarn.QueueMetrics.Queue=root.AppsFailed._rate/(yarn.QueueMetrics.Queue=root.AppsFailed._rate + yarn.QueueMetrics.Queue=root.AppsKilled._rate + yarn.QueueMetrics.Queue=root.AppsPending + yarn.QueueMetrics.Queue=root.AppsRunning + yarn.QueueMetrics.Queue=root.AppsSubmitted._rate + yarn.QueueMetrics.Queue=root.AppsCompleted._rate)) * 100}" - } - ], - "properties": { - "display_unit": "%", - "graph_type": "LINE", - "time_range": "1" - } - }, - { - "widget_name": "Pending Apps", - "description": "Count of applications waiting for cluster resources to become available.", - "widget_type": "GRAPH", - "is_visible": true, - "metrics": [ - { - "name": "yarn.QueueMetrics.Queue=root.AppsPending", - "metric_path": "metrics/yarn/Queue/root/AppsPending", - "service_name": "YARN", - "component_name": "RESOURCEMANAGER", - "host_component_criteria": "host_components/HostRoles/ha_state=ACTIVE" - } - ], - "values": [ - { - "name": "Pending Apps", - "value": "${yarn.QueueMetrics.Queue=root.AppsPending}" - } - ], - "properties": { - "display_unit": "Apps", - "graph_type": "LINE", - "time_range": "1" - } - }, - { - "widget_name": "Cluster Memory", - "description": "Percentage of memory used across all NodeManager hosts.", - "widget_type": "GRAPH", - "is_visible": true, - "metrics": [ - { - "name": "mem_total._sum", - "metric_path": "metrics/memory/mem_total._avg", - "service_name": "YARN", - "component_name": "NODEMANAGER" - }, - { - "name": "mem_free._sum", - "metric_path": "metrics/memory/mem_free._avg", - "service_name": "YARN", - "component_name": "NODEMANAGER" - } - ], - "values": [ - { - "name": "Memory utilization", - "value": "${((mem_total._sum - mem_free._sum)/mem_total._sum) * 100}" - } - ], - "properties": { - "display_unit": "%", - "graph_type": "LINE", - "time_range": "1" - } - }, - { - "widget_name": "Cluster Disk", - "description": "Sum of disk throughput for all NodeManager hosts.", - "widget_type": "GRAPH", - "is_visible": true, - "metrics": [ - { - "name": "read_bps._sum", - "metric_path": "metrics/disk/read_bps._sum", - "service_name": "YARN", - "component_name": "NODEMANAGER" - }, - { - "name": "write_bps._sum", - "metric_path": "metrics/disk/write_bps._sum", - "service_name": "YARN", - "component_name": "NODEMANAGER" - } - ], - "values": [ - { - "name": "Read throughput", - "value": "${read_bps._sum/1048576}" - }, - { - "name": "Write throughput", - "value": "${write_bps._sum/1048576}" - } - ], - "properties": { - "display_unit": "Mbps", - "graph_type": "LINE", - "time_range": "1" - } - }, - { - "widget_name": "Cluster Network", - "description": "Average of Network utilized across all NodeManager hosts.", - "default_section_name": "YARN_SUMMARY", - "widget_type": "GRAPH", - "is_visible": true, - "metrics": [ - { - "name": "pkts_in._avg", - "metric_path": "metrics/network/pkts_in._avg", - "service_name": "YARN", - "component_name": "NODEMANAGER" - }, - { - "name": "pkts_out._avg", - "metric_path": "metrics/network/pkts_out._avg", - "service_name": "YARN", - "component_name": "NODEMANAGER" - } - ], - "values": [ - { - "name": "Packets In", - "value": "${pkts_in._avg}" - }, - { - "name": "Packets Out", - "value": "${pkts_out._avg}" - } - ], - "properties": { - "graph_type": "LINE", - "time_range": "1" - } - }, - { - "widget_name": "Cluster CPU", - "description": "Percentage of CPU utilized across all NodeManager hosts.", - "default_section_name": "YARN_SUMMARY", - "widget_type": "GRAPH", - "is_visible": true, - "metrics": [ - { - "name": "cpu_system._sum", - "metric_path": "metrics/cpu/cpu_system._sum", - "service_name": "YARN", - "component_name": "NODEMANAGER" - }, - { - "name": "cpu_user._sum", - "metric_path": "metrics/cpu/cpu_user._sum", - "service_name": "YARN", - "component_name": "NODEMANAGER" - }, - { - "name": "cpu_nice._sum", - "metric_path": "metrics/cpu/cpu_nice._sum", - "service_name": "YARN", - "component_name": "NODEMANAGER" - }, - { - "name": "cpu_idle._sum", - "metric_path": "metrics/cpu/cpu_idle._sum", - "service_name": "YARN", - "component_name": "NODEMANAGER" - }, - { - "name": "cpu_wio._sum", - "metric_path": "metrics/cpu/cpu_wio._sum", - "service_name": "YARN", - "component_name": "NODEMANAGER" - } - ], - "values": [ - { - "name": "CPU utilization", - "value": "${((cpu_system._sum + cpu_user._sum + cpu_nice._sum)/(cpu_system._sum + cpu_user._sum + cpu_nice._sum + cpu_idle._sum + cpu_wio._sum)) * 100}" - } - ], - "properties": { - "graph_type": "LINE", - "time_range": "1", - "display_unit": "%" - } - } - ] - }, - { - "layout_name": "default_yarn_heatmap", - "display_name": "YARN Heatmaps", - "section_name": "YARN_HEATMAPS", - "widgetLayoutInfo": [ - { - "widget_name": "YARN local disk space utilization per NodeManager", - "description": "", - "widget_type": "HEATMAP", - "is_visible": true, - "metrics": [ - { - "name": "yarn.NodeManagerMetrics.GoodLocalDirsDiskUtilizationPerc", - "metric_path": "metrics/yarn/GoodLocalDirsDiskUtilizationPerc", - "service_name": "YARN", - "component_name": "NODEMANAGER" - }, - { - "name": "yarn.NodeManagerMetrics.GoodLogDirsDiskUtilizationPerc", - "metric_path": "metrics/yarn/GoodLogDirsDiskUtilizationPerc", - "service_name": "YARN", - "component_name": "NODEMANAGER" - } - ], - "values": [ - { - "name": "YARN local disk space utilization per NodeManager", - "value": "${(yarn.NodeManagerMetrics.GoodLocalDirsDiskUtilizationPerc + yarn.NodeManagerMetrics.GoodLogDirsDiskUtilizationPerc)/2}" - } - ], - "properties": { - "display_unit": "%", - "max_limit": "100" - } - }, - { - "widget_name": "Total Allocatable RAM Utilized per NodeManager", - "description": "", - "widget_type": "HEATMAP", - "is_visible": false, - "metrics": [ - { - "name": "yarn.NodeManagerMetrics.AllocatedGB", - "metric_path": "metrics/yarn/AllocatedGB", - "service_name": "YARN", - "component_name": "NODEMANAGER" - }, - { - "name": "yarn.NodeManagerMetrics.AvailableGB", - "metric_path": "metrics/yarn/AvailableGB", - "service_name": "YARN", - "component_name": "NODEMANAGER" - } - ], - "values": [ - { - "name": "Total Allocatable RAM Utilized per NodeManager", - "value": "${(yarn.NodeManagerMetrics.AllocatedGB/(yarn.NodeManagerMetrics.AvailableGB + yarn.NodeManagerMetrics.AllocatedGB)) * 100}" - } - ], - "properties": { - "display_unit": "%", - "max_limit": "100" - } - }, - { - "widget_name": "Total Allocatable CPU Utilized per NodeManager", - "description": "", - "widget_type": "HEATMAP", - "is_visible": false, - "metrics": [ - { - "name": "yarn.NodeManagerMetrics.AllocatedVCores", - "metric_path": "metrics/yarn/AllocatedVCores", - "service_name": "YARN", - "component_name": "NODEMANAGER" - }, - { - "name": "yarn.NodeManagerMetrics.AvailableVCores", - "metric_path": "metrics/yarn/AvailableVCores", - "service_name": "YARN", - "component_name": "NODEMANAGER" - } - ], - "values": [ - { - "name": "Total Allocatable CPU Utilized per NodeManager", - "value": "${(yarn.NodeManagerMetrics.AllocatedVCores/(yarn.NodeManagerMetrics.AllocatedVCores + yarn.NodeManagerMetrics.AvailableVCores)) * 100}" - } - ], - "properties": { - "display_unit": "%", - "max_limit": "100" - } - }, - { - "widget_name": "Container Failures", - "description": "", - "widget_type": "HEATMAP", - "is_visible": false, - "metrics": [ - { - "name": "yarn.NodeManagerMetrics.ContainersFailed._rate", - "metric_path": "metrics/yarn/ContainersFailed._rate", - "service_name": "YARN", - "component_name": "NODEMANAGER" - }, - { - "name": "yarn.NodeManagerMetrics.ContainersCompleted._rate", - "metric_path": "metrics/yarn/ContainersCompleted._rate", - "service_name": "YARN", - "component_name": "NODEMANAGER" - }, - { - "name": "yarn.NodeManagerMetrics.ContainersLaunched._rate", - "metric_path": "metrics/yarn/ContainersLaunched._rate", - "service_name": "YARN", - "component_name": "NODEMANAGER" - }, - { - "name": "yarn.NodeManagerMetrics.ContainersIniting", - "metric_path": "metrics/yarn/ContainersIniting", - "service_name": "YARN", - "component_name": "NODEMANAGER" - }, - { - "name": "yarn.NodeManagerMetrics.ContainersKilled._rate", - "metric_path": "metrics/yarn/ContainersKilled._rate", - "service_name": "YARN", - "component_name": "NODEMANAGER" - }, - { - "name": "yarn.NodeManagerMetrics.ContainersRunning", - "metric_path": "metrics/yarn/ContainersRunning", - "service_name": "YARN", - "component_name": "NODEMANAGER" - } - ], - "values": [ - { - "name": "Container Failures", - "value": "${(yarn.NodeManagerMetrics.ContainersFailed._rate/(yarn.NodeManagerMetrics.ContainersFailed._rate + yarn.NodeManagerMetrics.ContainersCompleted._rate + yarn.NodeManagerMetrics.ContainersLaunched._rate + yarn.NodeManagerMetrics.ContainersIniting + yarn.NodeManagerMetrics.ContainersKilled._rate + yarn.NodeManagerMetrics.ContainersRunning)) * 100}" - } - ], - "properties": { - "display_unit": "%", - "max_limit": "100" - } - }, - { - "widget_name": "NodeManager GC Time", - "description": "", - "widget_type": "HEATMAP", - "is_visible": false, - "metrics": [ - { - "name": "Hadoop:service=NodeManager,name=JvmMetrics.GcTimeMillis", - "metric_path": "metrics/jvm/gcTimeMillis", - "service_name": "YARN", - "component_name": "NODEMANAGER" - } - ], - "values": [ - { - "name": "NodeManager Garbage Collection Time", - "value": "${Hadoop:service=NodeManager,name=JvmMetrics.GcTimeMillis}" - } - ], - "properties": { - "display_unit": "ms", - "max_limit": "10000" - } - }, - { - "widget_name": "NodeManager JVM Heap Memory Used", - "description": "", - "widget_type": "HEATMAP", - "is_visible": false, - "metrics": [ - { - "name": "Hadoop:service=NodeManager,name=JvmMetrics.MemHeapUsedM", - "metric_path": "metrics/jvm/memHeapUsedM", - "service_name": "YARN", - "component_name": "NODEMANAGER" - } - ], - "values": [ - { - "name": "NodeManager JVM Heap Memory Used", - "value": "${Hadoop:service=NodeManager,name=JvmMetrics.MemHeapUsedM}" - } - ], - "properties": { - "display_unit": "MB", - "max_limit": "512" - } - }, - { - "widget_name": "Allocated Containers", - "description": "", - "widget_type": "HEATMAP", - "is_visible": false, - "metrics": [ - { - "name": "yarn.NodeManagerMetrics.AllocatedContainers", - "metric_path": "metrics/yarn/AllocatedContainers", - "service_name": "YARN", - "component_name": "NODEMANAGER" - } - ], - "values": [ - { - "name": "Allocated Containers", - "value": "${yarn.NodeManagerMetrics.AllocatedContainers}" - } - ], - "properties": { - "display_unit": "", - "max_limit": "100" - } - }, - { - "widget_name": "NodeManager RAM Utilized", - "description": "", - "widget_type": "HEATMAP", - "is_visible": false, - "metrics": [ - { - "name": "yarn.NodeManagerMetrics.AllocatedGB", - "metric_path": "metrics/yarn/AllocatedGB", - "service_name": "YARN", - "component_name": "NODEMANAGER" - } - ], - "values": [ - { - "name": "NodeManager RAM Utilized", - "value": "${yarn.NodeManagerMetrics.AllocatedGB}" - } - ], - "properties": { - "display_unit": "", - "max_limit": "100" - } - }, - { - "widget_name": "NodeManager CPU Utilized", - "description": "", - "widget_type": "HEATMAP", - "is_visible": false, - "metrics": [ - { - "name": "yarn.NodeManagerMetrics.AllocatedVCores", - "metric_path": "metrics/yarn/AllocatedVCores", - "service_name": "YARN", - "component_name": "NODEMANAGER" - } - ], - "values": [ - { - "name": "NodeManager CPU Utilized", - "value": "${yarn.NodeManagerMetrics.AllocatedVCores}" - } - ], - "properties": { - "display_unit": "", - "max_limit": "100" - } - } - ] - } - ] -} http://git-wip-us.apache.org/repos/asf/ambari/blob/d845449a/ambari-server/src/main/resources/stacks/HDP/3.0/services/YARN/configuration-mapred/mapred-env.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/3.0/services/YARN/configuration-mapred/mapred-env.xml b/ambari-server/src/main/resources/stacks/HDP/3.0/services/YARN/configuration-mapred/mapred-env.xml index a70fad3..deb4ef7 100644 --- a/ambari-server/src/main/resources/stacks/HDP/3.0/services/YARN/configuration-mapred/mapred-env.xml +++ b/ambari-server/src/main/resources/stacks/HDP/3.0/services/YARN/configuration-mapred/mapred-env.xml @@ -1,25 +1,23 @@ <?xml version="1.0"?> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> <!-- -/** - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ + Licensed to the Apache Software Foundation (ASF) under one or more + contributor license agreements. See the NOTICE file distributed with + this work for additional information regarding copyright ownership. + The ASF licenses this file to You under the Apache License, Version 2.0 + (the "License"); you may not use this file except in compliance with + the License. You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. --> -<configuration supports_adding_forbidden="true"> +<!-- Put site-specific property overrides in this file. --> +<configuration xmlns:xi="http://www.w3.org/2001/XInclude" supports_final="true"> <!-- These configs were inherited from HDP 2.2 --> <!-- mapred-env.sh --> <property> @@ -27,21 +25,21 @@ <display-name>mapred-env template</display-name> <description>This is the jinja template for mapred-env.sh file</description> <value> -# export JAVA_HOME=/home/y/libexec/jdk1.6.0/ + # export JAVA_HOME=/home/y/libexec/jdk1.6.0/ -export HADOOP_JOB_HISTORYSERVER_HEAPSIZE={{jobhistory_heapsize}} + export HADOOP_JOB_HISTORYSERVER_HEAPSIZE={{jobhistory_heapsize}} -export HADOOP_MAPRED_ROOT_LOGGER=INFO,RFA + export HADOOP_MAPRED_ROOT_LOGGER=INFO,RFA -#export HADOOP_JOB_HISTORYSERVER_OPTS= -#export HADOOP_MAPRED_LOG_DIR="" # Where log files are stored. $HADOOP_MAPRED_HOME/logs by default. -#export HADOOP_JHS_LOGGER=INFO,RFA # Hadoop JobSummary logger. -#export HADOOP_MAPRED_PID_DIR= # The pid files are stored. /tmp by default. -#export HADOOP_MAPRED_IDENT_STRING= #A string representing this instance of hadoop. $USER by default -#export HADOOP_MAPRED_NICENESS= #The scheduling priority for daemons. Defaults to 0. -export HADOOP_OPTS="-Dhdp.version=$HDP_VERSION $HADOOP_OPTS" -export HADOOP_OPTS="-Djava.io.tmpdir={{hadoop_java_io_tmpdir}} $HADOOP_OPTS" -export JAVA_LIBRARY_PATH="${JAVA_LIBRARY_PATH}:{{hadoop_java_io_tmpdir}}" + #export HADOOP_JOB_HISTORYSERVER_OPTS= + #export HADOOP_MAPRED_LOG_DIR="" # Where log files are stored. $HADOOP_MAPRED_HOME/logs by default. + #export HADOOP_JHS_LOGGER=INFO,RFA # Hadoop JobSummary logger. + #export HADOOP_MAPRED_PID_DIR= # The pid files are stored. /tmp by default. + #export HADOOP_MAPRED_IDENT_STRING= #A string representing this instance of hadoop. $USER by default + #export HADOOP_MAPRED_NICENESS= #The scheduling priority for daemons. Defaults to 0. + export HADOOP_OPTS="-Dhdp.version=$HDP_VERSION $HADOOP_OPTS" + export HADOOP_OPTS="-Djava.io.tmpdir={{hadoop_java_io_tmpdir}} $HADOOP_OPTS" + export JAVA_LIBRARY_PATH="${JAVA_LIBRARY_PATH}:{{hadoop_java_io_tmpdir}}" </value> <value-attributes> <type>content</type> http://git-wip-us.apache.org/repos/asf/ambari/blob/d845449a/ambari-server/src/main/resources/stacks/HDP/3.0/services/YARN/configuration-mapred/mapred-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/3.0/services/YARN/configuration-mapred/mapred-site.xml b/ambari-server/src/main/resources/stacks/HDP/3.0/services/YARN/configuration-mapred/mapred-site.xml index cef2b14..46f1c32 100644 --- a/ambari-server/src/main/resources/stacks/HDP/3.0/services/YARN/configuration-mapred/mapred-site.xml +++ b/ambari-server/src/main/resources/stacks/HDP/3.0/services/YARN/configuration-mapred/mapred-site.xml @@ -18,6 +18,16 @@ --> <!-- Put site-specific property overrides in this file. --> <configuration xmlns:xi="http://www.w3.org/2001/XInclude" supports_final="true"> + <property> + <name>mapreduce.application.classpath</name> + <value>$PWD/mr-framework/hadoop/share/hadoop/mapreduce/*:$PWD/mr-framework/hadoop/share/hadoop/mapreduce/lib/*:$PWD/mr-framework/hadoop/share/hadoop/common/*:$PWD/mr-framework/hadoop/share/hadoop/common/lib/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/lib/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/lib/*:$PWD/mr-framework/hadoop/share/hadoop/tools/lib/*:/usr/hdp/${hdp.version}/hadoop/lib/hadoop-lzo-0.6.0.${hdp.version}.jar:/etc/hadoop/conf/secure</value> + <description> + CLASSPATH for MR applications. A comma-separated list of CLASSPATH + entries. + </description> + <on-ambari-upgrade add="true"/> + </property> + <!-- These configs were inherited from HDP 2.2 --> <property> <name>mapreduce.admin.user.env</name> @@ -30,15 +40,6 @@ <on-ambari-upgrade add="true"/> </property> <property> - <name>mapreduce.application.classpath</name> - <value>$PWD/mr-framework/hadoop/share/hadoop/mapreduce/*:$PWD/mr-framework/hadoop/share/hadoop/mapreduce/lib/*:$PWD/mr-framework/hadoop/share/hadoop/common/*:$PWD/mr-framework/hadoop/share/hadoop/common/lib/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/lib/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/lib/*:$PWD/mr-framework/hadoop/share/hadoop/tools/lib/*:/usr/hdp/${hdp.version}/hadoop/lib/hadoop-lzo-0.6.0.${hdp.version}.jar:/etc/hadoop/conf/secure</value> - <description> - CLASSPATH for MR applications. A comma-separated list of CLASSPATH - entries. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> <name>mapreduce.application.framework.path</name> <value>/hdp/apps/${hdp.version}/mapreduce/mapreduce.tar.gz#mr-framework</value> <description/> @@ -74,61 +75,4 @@ <description/> <on-ambari-upgrade add="true"/> </property> - <property> - <name>mapreduce.reduce.shuffle.fetch.retry.enabled</name> - <value>1</value> - <description/> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>mapreduce.reduce.shuffle.fetch.retry.interval-ms</name> - <value>1000</value> - <description/> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>mapreduce.reduce.shuffle.fetch.retry.timeout-ms</name> - <value>30000</value> - <description/> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>mapreduce.job.emit-timeline-data</name> - <value>false</value> - <description/> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>mapreduce.jobhistory.bind-host</name> - <value>0.0.0.0</value> - <description/> - <on-ambari-upgrade add="true"/> - </property> - - <!-- These configs were inherited from HDP 2.3 --> - <property> - <name>mapreduce.jobhistory.recovery.enable</name> - <value>true</value> - <description>Enable the history server to store server state and recover - server state upon startup. If enabled then - mapreduce.jobhistory.recovery.store.class must be specified. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>mapreduce.jobhistory.recovery.store.class</name> - <value>org.apache.hadoop.mapreduce.v2.hs.HistoryServerLeveldbStateStoreService</value> - <description>The HistoryServerStateStoreService class to store history server - state for recovery. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>mapreduce.jobhistory.recovery.store.leveldb.path</name> - <value>/hadoop/mapreduce/jhs</value> - <description>The URI where history server state will be stored if HistoryServerLeveldbSystemStateStoreService - is configured as the recovery storage class. - </description> - <on-ambari-upgrade add="true"/> - </property> </configuration> http://git-wip-us.apache.org/repos/asf/ambari/blob/d845449a/ambari-server/src/main/resources/stacks/HDP/3.0/services/YARN/configuration/capacity-scheduler.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/3.0/services/YARN/configuration/capacity-scheduler.xml b/ambari-server/src/main/resources/stacks/HDP/3.0/services/YARN/configuration/capacity-scheduler.xml deleted file mode 100644 index 4768e46..0000000 --- a/ambari-server/src/main/resources/stacks/HDP/3.0/services/YARN/configuration/capacity-scheduler.xml +++ /dev/null @@ -1,71 +0,0 @@ -<?xml version="1.0"?> -<!-- - Licensed to the Apache Software Foundation (ASF) under one or more - contributor license agreements. See the NOTICE file distributed with - this work for additional information regarding copyright ownership. - The ASF licenses this file to You under the Apache License, Version 2.0 - (the "License"); you may not use this file except in compliance with - the License. You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - - Unless required by applicable law or agreed to in writing, software - distributed under the License is distributed on an "AS IS" BASIS, - WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - See the License for the specific language governing permissions and - limitations under the License. ---> -<configuration supports_final="false" supports_adding_forbidden="true"> - <!-- These configs were inherited from HDP 2.2 --> - <property> - <name>yarn.scheduler.capacity.resource-calculator</name> - <description> - The ResourceCalculator implementation to be used to compare Resources in the scheduler. - The default i.e. org.apache.hadoop.yarn.util.resource.DefaultResourseCalculator only uses - Memory while DominantResourceCalculator uses Dominant-resource to compare multi-dimensional - resources such as Memory, CPU etc. A Java ResourceCalculator class name is expected. - </description> - <value>org.apache.hadoop.yarn.util.resource.DefaultResourceCalculator</value> - <display-name>CPU Scheduling</display-name> - <value-attributes> - <type>value-list</type> - <entries> - <entry> - <value>org.apache.hadoop.yarn.util.resource.DominantResourceCalculator</value> - <label>Enabled</label> - </entry> - <entry> - <value>org.apache.hadoop.yarn.util.resource.DefaultResourceCalculator</value> - <label>Disabled</label> - </entry> - </entries> - <selection-cardinality>1</selection-cardinality> - </value-attributes> - <on-ambari-upgrade add="false"/> - </property> - <property> - <name>yarn.scheduler.capacity.root.accessible-node-labels</name> - <value>*</value> - <description/> - <on-ambari-upgrade add="false"/> - </property> - - <!-- In HDP 2.3, yarn.scheduler.capacity.root.default-node-label-expression was deleted --> - - <!-- These configs were inherited from HDP 2.5 --> - <property> - <name>capacity-scheduler</name> - <description>Enter key=value (one per line) for all properties of capacity-scheduler.xml</description> - <depends-on> - <property> - <type>hive-interactive-env</type> - <name>enable_hive_interactive</name> - </property> - <property> - <type>hive-interactive-env</type> - <name>llap_queue_capacity</name> - </property> - </depends-on> - <on-ambari-upgrade add="false"/> - </property> -</configuration>
