Are there more blocks on these hot DNs than there are on the cool
ones?   If you run a major compaction and then run your tests, does it
make a difference?
St.Ack

On Tue, May 17, 2011 at 8:03 PM, Weihua JIANG <[email protected]> wrote:
> -ROOT- and .META. table are not served by these hot region servers.
>
> I generate the key in random and verified at client by grepping .META.
> table and record the mapping from each query to its serving region
> server. It shows that each RS serves almost the same number of query
> requests.
>
> For GC hints, can you give me a link? I only found Todd's posts about
> GC tuning for write. But, in my case, I only perform query. So, the
> one I found seems no help to me.
>
> Thanks
> Weihua
>
> 2011/5/17 Michael Segel <[email protected]>:
>>
>> Sorry to jump in on the tail end.
>>
>> What do you mean to say that they key is generated randomly?
>>
>> I mean are you using a key and then applying a SHA-1 hash?
>>
>> Which node is serving your -ROOT- and META tables?
>>
>> Have you applied the GC hints recommended by Todd L in his blog?
>>
>>
>> Also you said:
>> '
>> And almost all these times are spent on DFSClient$BlockReader
>> to read data from DN.
>> '
>> What speed disks are you using and how many disks per node?
>> (you could be blocked on disk i/o.)
>>
>>
>> -Mike
>>
>>
>> ----------------------------------------
>>> Date: Tue, 17 May 2011 07:33:34 -0700
>>> Subject: Re: How to speedup Hbase query throughput
>>> From: [email protected]
>>> To: [email protected]
>>>
>>> Nice analysis.
>>>
>>> Can you figure the most popular blocks requested? You could figure
>>> which files they belong too by grepping the blocks in namenode log.
>>>
>>> It is odd that you have the sort of a request profile if your loading
>>> was even. I'd expect the DN distribution to be even.
>>>
>>> Sounds like hdfs-347 would help for sure.
>>>
>>> St.Ack
>>>
>>>
>>> On Tue, May 17, 2011 at 6:57 AM, Weihua JIANG  wrote:
>>> > No. The key is generated randomly. In theory, it shall distributed to
>>> > all the RSs equally.
>>> >
>>> > Thanks
>>> > Weihua
>>> >
>>> > 2011/5/17 Ted Dunning :
>>> >> Are your keys arranged so that you have a problem with a hot region?
>>> >>
>>> >> On Mon, May 16, 2011 at 11:18 PM, Weihua JIANG wrote:
>>> >>
>>> >>> I have not applied hdfs-347, but done some other experiments.
>>> >>>
>>> >>> I increased client thread to 2000 to put enough pressure on cluster. I
>>> >>> disabled RS block cache. The total TPS is still low (with Month+User
>>> >>> as row key, it is about 1300 for 10 RS+DN and with User+Month it is
>>> >>> 700).
>>> >>>
>>> >>> I used BTrace to log the time spent on each HTable.get on RS. It shows
>>> >>> that most of the GETs use 20~50ms and there are many GETs need
>>> >>> >1000ms. And almost all these times are spent on DFSClient$BlockReader
>>> >>> to read data from DN. But, the network usage is not high (<100Mb/s, we
>>> >>> have a giganet), so network is not a problem.
>>> >>>
>>> >>> Since for each DFS block read, there is a socket connection created. I
>>> >>> use netstat to caculate the TCP connections on 50010 port (DN listen
>>> >>> port) for each RS+DN server. It shows that there are always one or two
>>> >>> DNs have high such connection number (>200) while other DNs have low
>>> >>> number (<20). And the high connection DNs have high disk I/O usage
>>> >>> (about 100%) while other DNs have low disk I/O.  This phenoma lasts
>>> >>> for days and the hot machine is always the hot one.
>>> >>>
>>> >>> The high connection number mainly comes from local region server
>>> >>> request (~80%).
>>> >>>
>>> >>> According to the source code of DFSClient, it prefers to use local DN
>>> >>> to fetch block. But, why certain machine is so popular? All my servers
>>> >>> have almost the same configuration.
>>> >>>
>>> >>> 2011/4/29 Stack :
>>> >>> > Yes, you could try applying hdfs-347 to your hdfs as J-D suggests.  Do
>>> >>> > your numbers change if you run your client from more than one machine?
>>> >>> > St.Ack
>>> >>> >
>>> >>> > On Thu, Apr 28, 2011 at 2:56 PM, Jean-Daniel Cryans
>>> >>> wrote:
>>> >>> >> Seems to be a case of HDFS-347.
>>> >>> >>
>>> >>> >> J-D
>>> >>> >>
>>> >>> >> On Thu, Apr 28, 2011 at 12:55 AM, Weihua JIANG
>>> >>> wrote:
>>> >>> >>> After solving HBASE-3561, I successfully run hprof for RS and DN.
>>> >>> >>> Since block cache is useless in my case, I disabled it. I rerun my
>>> >>> >>> test with 14 RS+DNs and 1 client with 200 threads. But, the 
>>> >>> >>> throughput
>>> >>> >>> is still only about 700. No scalability shown in this case.
>>> >>> >>>
>>> >>> >>> Below is the hot spots in RS:
>>> >>> >>> CPU SAMPLES BEGIN (total = 1469756) Thu Apr 28 15:43:35 2011
>>> >>> >>> rank   self  accum   count trace method
>>> >>> >>>   1 44.33% 44.33%  651504 300612 
>>> >>> >>> sun.nio.ch.EPollArrayWrapper.epollWait
>>> >>> >>>   2 19.88% 64.21%  292221 301351 
>>> >>> >>> sun.nio.ch.EPollArrayWrapper.epollWait
>>> >>> >>>   3  8.88% 73.09%  130582 300554 
>>> >>> >>> sun.nio.ch.EPollArrayWrapper.epollWait
>>> >>> >>>   4  4.43% 77.52%   65106 301248 
>>> >>> >>> sun.nio.ch.EPollArrayWrapper.epollWait
>>> >>> >>>   5  4.43% 81.95%   65104 301249 
>>> >>> >>> sun.nio.ch.EPollArrayWrapper.epollWait
>>> >>> >>>   6  4.43% 86.38%   65100 301247 
>>> >>> >>> sun.nio.ch.EPollArrayWrapper.epollWait
>>> >>> >>>   7  4.43% 90.81%   65061 301266 
>>> >>> >>> sun.nio.ch.EPollArrayWrapper.epollWait
>>> >>> >>>   8  4.32% 95.13%   63465 301565 
>>> >>> >>> sun.nio.ch.EPollArrayWrapper.epollWait
>>> >>> >>>   9  2.31% 97.43%   33894 301555 
>>> >>> >>> sun.nio.ch.EPollArrayWrapper.epollWait
>>> >>> >>>  10  1.76% 99.19%   25841 301588 
>>> >>> >>> sun.nio.ch.EPollArrayWrapper.epollWait
>>> >>> >>>  11  0.48% 99.67%    7025 
>>> >>> >>> 301443sun.nio.ch.EPollArrayWrapper.epollWait
>>> >>> >>>  12  0.02% 99.69%     341 301568 sun.nio.ch.NativeThread.current
>>> >>> >>>  13  0.01% 99.71%     187 301535
>>> >>> >>> org.apache.hadoop.hbase.io.hfile.HFile$Reader.indexSize
>>> >>> >>>  14  0.01% 99.72%     186 301538
>>> >>> >>> org.apache.hadoop.hbase.io.hfile.HFile$Reader.indexSize
>>> >>> >>>  15  0.01% 99.73%     170 301625
>>> >>> org.apache.hadoop.util.DataChecksum.update
>>> >>> >>>  16  0.01% 99.74%     164 301579 
>>> >>> >>> sun.nio.ch.EPollArrayWrapper.epollWait
>>> >>> >>>  17  0.01% 99.75%     149 300938 
>>> >>> >>> sun.nio.ch.EPollArrayWrapper.epollWait
>>> >>> >>>
>>> >>> >>> TRACE 300612:
>>> >>> >>>
>>> >>>  sun.nio.ch.EPollArrayWrapper.epollWait(EPollArrayWrapper.java:Unknown 
>>> >>> line)
>>> >>> >>>        sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:215)
>>> >>> >>>        
>>> >>> >>> sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:65)
>>> >>> >>>        sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69)
>>> >>> >>>        sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80)
>>> >>> >>>        sun.nio.ch.SelectorImpl.select(SelectorImpl.java:84)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hbase.ipc.HBaseServer$Listener$Reader.run(HBaseServer.java:305)
>>> >>> >>>
>>> >>>  java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>>> >>> >>>
>>> >>>  java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>>> >>> >>>        java.lang.Thread.run(Thread.java:619)
>>> >>> >>> TRACE 301351:
>>> >>> >>>
>>> >>>  sun.nio.ch.EPollArrayWrapper.epollWait(EPollArrayWrapper.java:Unknown 
>>> >>> line)
>>> >>> >>>        sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:215)
>>> >>> >>>        
>>> >>> >>> sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:65)
>>> >>> >>>        sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69)
>>> >>> >>>        sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80)
>>> >>> >>>
>>> >>>  org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:332)
>>> >>> >>>
>>> >>>  org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:157)
>>> >>> >>>
>>> >>>  org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:155)
>>> >>> >>>
>>> >>>  org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:128)
>>> >>> >>>        
>>> >>> >>> java.io.BufferedInputStream.fill(BufferedInputStream.java:218)
>>> >>> >>>        
>>> >>> >>> java.io.BufferedInputStream.read1(BufferedInputStream.java:258)
>>> >>> >>>        
>>> >>> >>> java.io.BufferedInputStream.read(BufferedInputStream.java:317)
>>> >>> >>>        java.io.DataInputStream.read(DataInputStream.java:132)
>>> >>> >>>        org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:102)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hdfs.DFSClient$BlockReader.readChunk(DFSClient.java:1389)
>>> >>> >>>
>>> >>>  org.apache.hadoop.fs.FSInputChecker.readChecksumChunk(FSInputChecker.java:237)
>>> >>> >>>
>>> >>>  org.apache.hadoop.fs.FSInputChecker.fill(FSInputChecker.java:176)
>>> >>> >>>
>>> >>>  org.apache.hadoop.fs.FSInputChecker.read1(FSInputChecker.java:193)
>>> >>> >>>
>>> >>>  org.apache.hadoop.fs.FSInputChecker.read(FSInputChecker.java:158)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hdfs.DFSClient$BlockReader.read(DFSClient.java:1243)
>>> >>> >>> TRACE 300554:
>>> >>> >>>
>>> >>>  sun.nio.ch.EPollArrayWrapper.epollWait(EPollArrayWrapper.java:Unknown 
>>> >>> line)
>>> >>> >>>        sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:215)
>>> >>> >>>        
>>> >>> >>> sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:65)
>>> >>> >>>        sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69)
>>> >>> >>>        sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80)
>>> >>> >>>
>>> >>>  org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1107)
>>> >>> >>> TRACE 301248:
>>> >>> >>>
>>> >>>  sun.nio.ch.EPollArrayWrapper.epollWait(EPollArrayWrapper.java:Unknown 
>>> >>> line)
>>> >>> >>>        sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:215)
>>> >>> >>>        
>>> >>> >>> sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:65)
>>> >>> >>>        sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69)
>>> >>> >>>        sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hbase.ipc.HBaseServer$Responder.run(HBaseServer.java:588)
>>> >>> >>> TRACE 301249:
>>> >>> >>>
>>> >>>  sun.nio.ch.EPollArrayWrapper.epollWait(EPollArrayWrapper.java:Unknown 
>>> >>> line)
>>> >>> >>>        sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:215)
>>> >>> >>>        
>>> >>> >>> sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:65)
>>> >>> >>>        sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69)
>>> >>> >>>        sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80)
>>> >>> >>>        sun.nio.ch.SelectorImpl.select(SelectorImpl.java:84)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hbase.ipc.HBaseServer$Listener.run(HBaseServer.java:414)
>>> >>> >>> TRACE 301247:
>>> >>> >>>
>>> >>>  sun.nio.ch.EPollArrayWrapper.epollWait(EPollArrayWrapper.java:Unknown 
>>> >>> line)
>>> >>> >>>        sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:215)
>>> >>> >>>        
>>> >>> >>> sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:65)
>>> >>> >>>        sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69)
>>> >>> >>>        sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80)
>>> >>> >>>
>>> >>>  org.mortbay.io.nio.SelectorManager$SelectSet.doSelect(SelectorManager.java:498)
>>> >>> >>>
>>> >>>  org.mortbay.io.nio.SelectorManager.doSelect(SelectorManager.java:192)
>>> >>> >>>
>>> >>>  org.mortbay.jetty.nio.SelectChannelConnector.accept(SelectChannelConnector.java:124)
>>> >>> >>>
>>> >>>  org.mortbay.jetty.AbstractConnector$Acceptor.run(AbstractConnector.java:708)
>>> >>> >>>
>>> >>>  org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582)
>>> >>> >>> TRACE 301266:
>>> >>> >>>
>>> >>>  sun.nio.ch.EPollArrayWrapper.epollWait(EPollArrayWrapper.java:Unknown 
>>> >>> line)
>>> >>> >>>        sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:215)
>>> >>> >>>        
>>> >>> >>> sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:65)
>>> >>> >>>        sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69)
>>> >>> >>>        sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80)
>>> >>> >>>
>>> >>>  org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:332)
>>> >>> >>>
>>> >>>  org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:157)
>>> >>> >>>
>>> >>>  org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:155)
>>> >>> >>>
>>> >>>  org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:128)
>>> >>> >>>        java.io.DataInputStream.readFully(DataInputStream.java:178)
>>> >>> >>>        java.io.DataInputStream.readLong(DataInputStream.java:399)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hdfs.protocol.DataTransferProtocol$PipelineAck.readFields(DataTransferProtocol.java:120)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$ResponseProcessor.run(DFSClient.java:2638)
>>> >>> >>> TRACE 301565:
>>> >>> >>>
>>> >>>  sun.nio.ch.EPollArrayWrapper.epollWait(EPollArrayWrapper.java:Unknown 
>>> >>> line)
>>> >>> >>>        sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:215)
>>> >>> >>>        
>>> >>> >>> sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:65)
>>> >>> >>>        sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69)
>>> >>> >>>        sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80)
>>> >>> >>>
>>> >>>  org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:332)
>>> >>> >>>
>>> >>>  org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:157)
>>> >>> >>>
>>> >>>  org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:155)
>>> >>> >>>
>>> >>>  org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:128)
>>> >>> >>>        
>>> >>> >>> java.io.BufferedInputStream.fill(BufferedInputStream.java:218)
>>> >>> >>>        
>>> >>> >>> java.io.BufferedInputStream.read1(BufferedInputStream.java:258)
>>> >>> >>>        
>>> >>> >>> java.io.BufferedInputStream.read(BufferedInputStream.java:317)
>>> >>> >>>        java.io.DataInputStream.read(DataInputStream.java:132)
>>> >>> >>>        org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:102)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hdfs.DFSClient$BlockReader.readChunk(DFSClient.java:1389)
>>> >>> >>>
>>> >>>  org.apache.hadoop.fs.FSInputChecker.readChecksumChunk(FSInputChecker.java:237)
>>> >>> >>>
>>> >>>  org.apache.hadoop.fs.FSInputChecker.read1(FSInputChecker.java:189)
>>> >>> >>>
>>> >>>  org.apache.hadoop.fs.FSInputChecker.read(FSInputChecker.java:158)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hdfs.DFSClient$BlockReader.read(DFSClient.java:1249)
>>> >>> >>>
>>> >>>  org.apache.hadoop.fs.FSInputChecker.readFully(FSInputChecker.java:384)
>>> >>> >>>
>>> >>> >>>
>>> >>> >>> And the hot spot of DN is:
>>> >>> >>> CPU SAMPLES BEGIN (total = 1123925) Thu Apr 28 15:43:49 2011
>>> >>> >>> rank   self  accum   count trace method
>>> >>> >>>   1 12.38% 12.38%  139155 301547 
>>> >>> >>> sun.nio.ch.EPollArrayWrapper.epollWait
>>> >>> >>>   2 12.38% 24.76%  139154 301548 
>>> >>> >>> sun.nio.ch.EPollArrayWrapper.epollWait
>>> >>> >>>   3 10.76% 35.53%  120969 300368 
>>> >>> >>> java.net.PlainSocketImpl.socketAccept
>>> >>> >>>   4 10.74% 46.27%  120753 301213 
>>> >>> >>> sun.nio.ch.EPollArrayWrapper.epollWait
>>> >>> >>>   5 10.74% 57.01%  120744 301381 
>>> >>> >>> sun.nio.ch.EPollArrayWrapper.epollWait
>>> >>> >>>   6 10.74% 67.76%  120744 301384 
>>> >>> >>> sun.nio.ch.EPollArrayWrapper.epollWait
>>> >>> >>>   7 10.74% 78.50%  120743 301386 
>>> >>> >>> sun.nio.ch.EPollArrayWrapper.epollWait
>>> >>> >>>   8 10.68% 89.18%  120019 301383
>>> >>> sun.nio.ch.ServerSocketChannelImpl.accept0
>>> >>> >>>   9  8.11% 97.29%   91174 301475 
>>> >>> >>> sun.nio.ch.FileChannelImpl.transferTo0
>>> >>> >>>  10  1.44% 98.73%   16172 301479 java.io.FileInputStream.readBytes
>>> >>> >>>  11  0.40% 99.13%    4496 301477 java.io.FileInputStream.readBytes
>>> >>> >>>  12  0.23% 99.35%    2531 301497 
>>> >>> >>> sun.nio.ch.EPollArrayWrapper.epollWait
>>> >>> >>>  13  0.17% 99.53%    1958 301461 java.io.FileInputStream.readBytes
>>> >>> >>>  14  0.04% 99.57%     458 301468 java.io.UnixFileSystem.delete0
>>> >>> >>>  15  0.03% 99.59%     289 301520 java.lang.Thread.start0
>>> >>> >>>  16  0.02% 99.62%     278 301471
>>> >>> org.apache.hadoop.util.DataChecksum.update
>>> >>> >>>  17  0.02% 99.64%     204 301562
>>> >>> java.io.UnixFileSystem.getBooleanAttributes0
>>> >>> >>>  18  0.01% 99.65%     117 301587 
>>> >>> >>> sun.nio.ch.EPollArrayWrapper.epollCtl
>>> >>> >>>  19  0.01% 99.66%     113 301493 sun.nio.ch.FileDispatcher.write0
>>> >>> >>>
>>> >>> >>> TRACE 301547:
>>> >>> >>>
>>> >>>  sun.nio.ch.EPollArrayWrapper.epollWait(EPollArrayWrapper.java:Unknown 
>>> >>> line)
>>> >>> >>>        sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:215)
>>> >>> >>>        
>>> >>> >>> sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:65)
>>> >>> >>>        sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69)
>>> >>> >>>        sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80)
>>> >>> >>>
>>> >>>  org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:332)
>>> >>> >>>
>>> >>>  org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:157)
>>> >>> >>>
>>> >>>  org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:155)
>>> >>> >>>
>>> >>>  org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:128)
>>> >>> >>>        
>>> >>> >>> java.io.BufferedInputStream.fill(BufferedInputStream.java:218)
>>> >>> >>>        
>>> >>> >>> java.io.BufferedInputStream.read1(BufferedInputStream.java:258)
>>> >>> >>>        
>>> >>> >>> java.io.BufferedInputStream.read(BufferedInputStream.java:317)
>>> >>> >>>        java.io.DataInputStream.read(DataInputStream.java:132)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hdfs.server.datanode.BlockReceiver.readToBuf(BlockReceiver.java:267)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hdfs.server.datanode.BlockReceiver.readNextPacket(BlockReceiver.java:314)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:378)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:534)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:417)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:122)
>>> >>> >>> TRACE 301548:
>>> >>> >>>
>>> >>>  sun.nio.ch.EPollArrayWrapper.epollWait(EPollArrayWrapper.java:Unknown 
>>> >>> line)
>>> >>> >>>        sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:215)
>>> >>> >>>        
>>> >>> >>> sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:65)
>>> >>> >>>        sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69)
>>> >>> >>>        sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80)
>>> >>> >>>
>>> >>>  org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:332)
>>> >>> >>>
>>> >>>  org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:157)
>>> >>> >>>
>>> >>>  org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:155)
>>> >>> >>>
>>> >>>  org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:128)
>>> >>> >>>        java.io.DataInputStream.readFully(DataInputStream.java:178)
>>> >>> >>>        java.io.DataInputStream.readLong(DataInputStream.java:399)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hdfs.protocol.DataTransferProtocol$PipelineAck.readFields(DataTransferProtocol.java:120)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:894)
>>> >>> >>>        java.lang.Thread.run(Thread.java:619)
>>> >>> >>> TRACE 300368:
>>> >>> >>>
>>> >>>  java.net.PlainSocketImpl.socketAccept(PlainSocketImpl.java:Unknown 
>>> >>> line)
>>> >>> >>>        java.net.PlainSocketImpl.accept(PlainSocketImpl.java:390)
>>> >>> >>>        java.net.ServerSocket.implAccept(ServerSocket.java:453)
>>> >>> >>>        java.net.ServerSocket.accept(ServerSocket.java:421)
>>> >>> >>>
>>> >>>  sun.management.jmxremote.LocalRMIServerSocketFactory$1.accept(LocalRMIServerSocketFactory.java:34)
>>> >>> >>>
>>> >>>  sun.rmi.transport.tcp.TCPTransport$AcceptLoop.executeAcceptLoop(TCPTransport.java:369)
>>> >>> >>>
>>> >>>  sun.rmi.transport.tcp.TCPTransport$AcceptLoop.run(TCPTransport.java:341)
>>> >>> >>>        java.lang.Thread.run(Thread.java:619)
>>> >>> >>> TRACE 301213:
>>> >>> >>>
>>> >>>  sun.nio.ch.EPollArrayWrapper.epollWait(EPollArrayWrapper.java:Unknown 
>>> >>> line)
>>> >>> >>>        sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:215)
>>> >>> >>>        
>>> >>> >>> sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:65)
>>> >>> >>>        sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69)
>>> >>> >>>        sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80)
>>> >>> >>>
>>> >>>  org.mortbay.io.nio.SelectorManager$SelectSet.doSelect(SelectorManager.java:498)
>>> >>> >>>
>>> >>>  org.mortbay.io.nio.SelectorManager.doSelect(SelectorManager.java:192)
>>> >>> >>>
>>> >>>  org.mortbay.jetty.nio.SelectChannelConnector.accept(SelectChannelConnector.java:124)
>>> >>> >>>
>>> >>>  org.mortbay.jetty.AbstractConnector$Acceptor.run(AbstractConnector.java:708)
>>> >>> >>>
>>> >>>  org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582)
>>> >>> >>> TRACE 301381:
>>> >>> >>>
>>> >>>  sun.nio.ch.EPollArrayWrapper.epollWait(EPollArrayWrapper.java:Unknown 
>>> >>> line)
>>> >>> >>>        sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:215)
>>> >>> >>>        
>>> >>> >>> sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:65)
>>> >>> >>>        sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69)
>>> >>> >>>        sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80)
>>> >>> >>>        sun.nio.ch.SelectorImpl.select(SelectorImpl.java:84)
>>> >>> >>>
>>> >>>  org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:352)
>>> >>> >>> TRACE 301384:
>>> >>> >>>
>>> >>>  sun.nio.ch.EPollArrayWrapper.epollWait(EPollArrayWrapper.java:Unknown 
>>> >>> line)
>>> >>> >>>        sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:215)
>>> >>> >>>        
>>> >>> >>> sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:65)
>>> >>> >>>        sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69)
>>> >>> >>>        sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80)
>>> >>> >>>        org.apache.hadoop.ipc.Server$Responder.run(Server.java:636)
>>> >>> >>> TRACE 301386:
>>> >>> >>>
>>> >>>  sun.nio.ch.EPollArrayWrapper.epollWait(EPollArrayWrapper.java:Unknown 
>>> >>> line)
>>> >>> >>>        sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:215)
>>> >>> >>>        
>>> >>> >>> sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:65)
>>> >>> >>>        sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69)
>>> >>> >>>        sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80)
>>> >>> >>>        sun.nio.ch.SelectorImpl.select(SelectorImpl.java:84)
>>> >>> >>>        org.apache.hadoop.ipc.Server$Listener.run(Server.java:468)
>>> >>> >>> TRACE 301383:
>>> >>> >>>
>>> >>>  sun.nio.ch.ServerSocketChannelImpl.accept0(ServerSocketChannelImpl.java:Unknown
>>> >>> >>> line)
>>> >>> >>>
>>> >>>  sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:145)
>>> >>> >>>
>>> >>>  sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:84)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:130)
>>> >>> >>>        java.lang.Thread.run(Thread.java:619)
>>> >>> >>> TRACE 301475:
>>> >>> >>>
>>> >>>  sun.nio.ch.FileChannelImpl.transferTo0(FileChannelImpl.java:Unknown 
>>> >>> line)
>>> >>> >>>
>>> >>>  sun.nio.ch.FileChannelImpl.transferToDirectly(FileChannelImpl.java:415)
>>> >>> >>>        
>>> >>> >>> sun.nio.ch.FileChannelImpl.transferTo(FileChannelImpl.java:516)
>>> >>> >>>
>>> >>>  org.apache.hadoop.net.SocketOutputStream.transferToFully(SocketOutputStream.java:199)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hdfs.server.datanode.BlockSender.sendChunks(BlockSender.java:350)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hdfs.server.datanode.BlockSender.sendBlock(BlockSender.java:436)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hdfs.server.datanode.DataXceiver.readBlock(DataXceiver.java:214)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:114)
>>> >>> >>> TRACE 301479:
>>> >>> >>>        
>>> >>> >>> java.io.FileInputStream.readBytes(FileInputStream.java:Unknown
>>> >>> line)
>>> >>> >>>        java.io.FileInputStream.read(FileInputStream.java:199)
>>> >>> >>>        java.io.FilterInputStream.read(FilterInputStream.java:116)
>>> >>> >>>        
>>> >>> >>> java.io.BufferedInputStream.fill(BufferedInputStream.java:218)
>>> >>> >>>        
>>> >>> >>> java.io.BufferedInputStream.read1(BufferedInputStream.java:258)
>>> >>> >>>        
>>> >>> >>> java.io.BufferedInputStream.read(BufferedInputStream.java:317)
>>> >>> >>>        java.io.DataInputStream.readFully(DataInputStream.java:178)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hdfs.server.datanode.BlockSender.sendChunks(BlockSender.java:271)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hdfs.server.datanode.BlockSender.sendBlock(BlockSender.java:436)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hdfs.server.datanode.DataXceiver.readBlock(DataXceiver.java:214)
>>> >>> >>>
>>> >>>  org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:114)
>>> >>> >>>
>>> >>> >>> Thanks
>>> >>> >>> Weihua
>>> >>> >>>
>>> >>> >>> 2011/4/28 Weihua JIANG :
>>> >>> >>>> I am using CDH3U0. It is HBase 0.90.1, I think.
>>> >>> >>>>
>>> >>> >>>> Thanks
>>> >>> >>>> Weihua
>>> >>> >>>>
>>> >>> >>>> 2011/4/28 Stack :
>>> >>> >>>>> On Tue, Apr 26, 2011 at 6:02 PM, Weihua JIANG <
>>> >>> [email protected]> wrote:
>>> >>> >>>>>> I tried to enable HPROF on RS, but failed. If I added the HPROF
>>> >>> agent
>>> >>> >>>>>> in hbase-env.sh, RS startup reports an error said HPROF can't be
>>> >>> >>>>>> loaded twice. But, I am sure I only enabled it once. I don't know
>>> >>> >>>>>> where the problem is.
>>> >>> >>>>>>
>>> >>> >>>>>
>>> >>> >>>>> This sounds like 'HBASE-3561  OPTS arguments are duplicated'  Are 
>>> >>> >>>>> you
>>> >>> >>>>> running 0.90.2?
>>> >>> >>>>>
>>> >>> >>>>> St.Ack
>>> >>> >>>>>
>>> >>> >>>>
>>> >>> >>>
>>> >>> >>
>>> >>> >
>>> >>>
>>> >>
>>> >
>>
>

Reply via email to