Please send vendor specific questions to that vendor's support mechanism. Since your issue appears to be with CDH, please use http://community.cloudera.com/<http://community.cloudera.com/>
Thanks & Regards Brahma Reddy Battula ________________________________ From: ZhuGe [t...@outlook.com] Sent: Wednesday, April 08, 2015 8:41 AM To: user@hadoop.apache.org Subject: hdfs ls command took more time to get response after update Hi all: Recently, I updated my hadoop cluster from hadoop-2.0.0-cdh4.3.0 to hadoop-2.5.0-cdh5.2.0. It works fine, however, a small problem is, when i use the hadoop fs -ls command in the terminal to get the list of files in the hdfs, it took much more time(10+ sec) to get the response compared to 2-3 secs before i update the version of hadoop.( get is slow too) Can any one exaplain a little bit of what might cause the problem or some configuration goes wrong? Below is the log: 15/04/08 10:51:18 DEBUG util.Shell: setsid exited with exit code 0 15/04/08 10:51:18 DEBUG conf.Configuration: parsing URL jar:file:/data/dbcenter/cdh5/hadoop-2.5.0-cdh5.2.0/share/hadoop/common/hadoop-common-2.5.0-cdh5.2.0.jar!/core-default.xml 15/04/08 10:51:18 DEBUG conf.Configuration: parsing input stream sun.net.www.protocol.jar.JarURLConnection$JarURLInputStream@57316e85 15/04/08 10:51:18 DEBUG conf.Configuration: parsing URL file:/data/dbcenter/cdh5/hadoop-2.5.0-cdh5.2.0/etc/hadoop/core-site.xml 15/04/08 10:51:18 DEBUG conf.Configuration: parsing input stream java.io.BufferedInputStream@31818dbc 15/04/08 10:51:19 DEBUG lib.MutableMetricsFactory: field org.apache.hadoop.metrics2.lib.MutableRate org.apache.hadoop.security.UserGroupInformation$UgiMetrics.loginSuccess with annotation @org.apache.hadoop.metrics2.annotation.Metric(valueName=Time, value=[Rate of successful kerberos logins and latency (milliseconds)], about=, type=DEFAULT, always=false, sampleName=Ops) 15/04/08 10:51:19 DEBUG lib.MutableMetricsFactory: field org.apache.hadoop.metrics2.lib.MutableRate org.apache.hadoop.security.UserGroupInformation$UgiMetrics.loginFailure with annotation @org.apache.hadoop.metrics2.annotation.Metric(valueName=Time, value=[Rate of failed kerberos logins and latency (milliseconds)], about=, type=DEFAULT, always=false, sampleName=Ops) 15/04/08 10:51:19 DEBUG lib.MutableMetricsFactory: field org.apache.hadoop.metrics2.lib.MutableRate org.apache.hadoop.security.UserGroupInformation$UgiMetrics.getGroups with annotation @org.apache.hadoop.metrics2.annotation.Metric(valueName=Time, value=[GetGroups], about=, type=DEFAULT, always=false, sampleName=Ops) 15/04/08 10:51:19 DEBUG impl.MetricsSystemImpl: UgiMetrics, User and group related metrics 15/04/08 10:51:19 DEBUG security.Groups: Creating new Groups object 15/04/08 10:51:19 DEBUG util.NativeCodeLoader: Trying to load the custom-built native-hadoop library... 15/04/08 10:51:19 DEBUG util.NativeCodeLoader: Failed to load native-hadoop with error: java.lang.UnsatisfiedLinkError: no hadoop in java.library.path 15/04/08 10:51:19 DEBUG util.NativeCodeLoader: java.library.path=/data/dbcenter/cdh5/hadoop-2.5.0-cdh5.2.0/lib/native 15/04/08 10:51:19 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 15/04/08 10:51:19 DEBUG util.PerformanceAdvisory: Falling back to shell based 15/04/08 10:51:19 DEBUG security.JniBasedUnixGroupsMappingWithFallback: Group mapping impl=org.apache.hadoop.security.ShellBasedUnixGroupsMapping 15/04/08 10:51:19 DEBUG security.Groups: Group mapping impl=org.apache.hadoop.security.JniBasedUnixGroupsMappingWithFallback; cacheTimeout=300000; warningDeltaMs=5000 15/04/08 10:51:19 DEBUG security.UserGroupInformation: hadoop login 15/04/08 10:51:19 DEBUG security.UserGroupInformation: hadoop login commit 15/04/08 10:51:19 DEBUG security.UserGroupInformation: using local user:UnixPrincipal: test 15/04/08 10:51:19 DEBUG security.UserGroupInformation: UGI loginUser:test (auth:SIMPLE) 15/04/08 10:51:19 DEBUG hdfs.BlockReaderLocal: dfs.client.use.legacy.blockreader.local = false 15/04/08 10:51:19 DEBUG hdfs.BlockReaderLocal: dfs.client.read.shortcircuit = false 15/04/08 10:51:19 DEBUG hdfs.BlockReaderLocal: dfs.client.domain.socket.data.traffic = false 15/04/08 10:51:19 DEBUG hdfs.BlockReaderLocal: dfs.domain.socket.path = 15/04/08 10:51:19 DEBUG hdfs.DFSClient: No KeyProvider found. 15/04/08 10:51:19 DEBUG hdfs.HAUtil: No HA service delegation token found for logical URI hdfs://tccluster:8020 15/04/08 10:51:19 DEBUG hdfs.BlockReaderLocal: dfs.client.use.legacy.blockreader.local = false 15/04/08 10:51:19 DEBUG hdfs.BlockReaderLocal: dfs.client.read.shortcircuit = false 15/04/08 10:51:19 DEBUG hdfs.BlockReaderLocal: dfs.client.domain.socket.data.traffic = false 15/04/08 10:51:19 DEBUG hdfs.BlockReaderLocal: dfs.domain.socket.path = 15/04/08 10:51:19 DEBUG retry.RetryUtils: multipleLinearRandomRetry = null 15/04/08 10:51:19 DEBUG ipc.Server: rpcKind=RPC_PROTOCOL_BUFFER, rpcRequestWrapperClass=class org.apache.hadoop.ipc.ProtobufRpcEngine$RpcRequestWrapper, rpcInvoker=org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker@5a2611a6 15/04/08 10:51:19 DEBUG ipc.Client: getting client out of cache: org.apache.hadoop.ipc.Client@285d4a6a 15/04/08 10:51:30 DEBUG util.PerformanceAdvisory: Both short-circuit local reads and UNIX domain socket are disabled. 15/04/08 10:51:30 DEBUG sasl.DataTransferSaslUtil: DataTransferProtocol not using SaslPropertiesResolver, no QOP found in configuration for dfs.data.transfer.protection 15/04/08 10:51:30 DEBUG ipc.Client: The ping interval is 60000 ms. 15/04/08 10:51:30 DEBUG ipc.Client: Connecting to master/192.168.1.13:8020 15/04/08 10:51:30 DEBUG ipc.Client: IPC Client (246890776) connection to master/192.168.1.13:8020 from test: starting, having connections 1 15/04/08 10:51:30 DEBUG ipc.Client: IPC Client (246890776) connection to master/192.168.1.13:8020 from test sending #0 15/04/08 10:51:30 DEBUG ipc.Client: IPC Client (246890776) connection to master/192.168.1.13:8020 from test got value #0 15/04/08 10:51:30 DEBUG ipc.ProtobufRpcEngine: Call: getFileInfo took 88ms 15/04/08 10:51:30 DEBUG ipc.Client: IPC Client (246890776) connection to master/192.168.1.13:8020 from test sending #1 15/04/08 10:51:30 DEBUG ipc.Client: IPC Client (246890776) connection to master/192.168.1.13:8020 from test got value #1 15/04/08 10:51:30 DEBUG ipc.ProtobufRpcEngine: Call: getListing took 2ms 15/04/08 10:51:30 DEBUG ipc.Client: stopping client from cache: org.apache.hadoop.ipc.Client@285d4a6a 15/04/08 10:51:30 DEBUG ipc.Client: removing client from cache: org.apache.hadoop.ipc.Client@285d4a6a 15/04/08 10:51:30 DEBUG ipc.Client: stopping actual client because no more references remain: org.apache.hadoop.ipc.Client@285d4a6a 15/04/08 10:51:30 DEBUG ipc.Client: Stopping client 15/04/08 10:51:30 DEBUG ipc.Client: IPC Client (246890776) connection to master/192.168.1.13:8020 from test: closed 15/04/08 10:51:30 DEBUG ipc.Client: IPC Client (246890776) connection to master/192.168.1.13:8020 from test: stopped, remaining connections 0