[jira] [Resolved] (HBASE-16212) Many connections to datanode are created when doing a large scan

2016-11-07 Thread Zhihua Deng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhihua Deng resolved HBASE-16212.
-
Resolution: Not A Bug

> Many connections to datanode are created when doing a large scan 
> -
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, HBASE-16212.v2.patch
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.
> hadoop version: 2.7.1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HBASE-16212) Many connections to datanode are created when doing a large scan

2016-08-09 Thread Zhihua Deng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhihua Deng updated HBASE-16212:

Attachment: (was: regionserver-dfsinputstream.log)

> Many connections to datanode are created when doing a large scan 
> -
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, HBASE-16212.v2.patch
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.
> hadoop version: 2.7.1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HBASE-16212) Many connections to datanode are created when doing a large scan

2016-07-20 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15385485#comment-15385485
 ] 

Zhihua Deng commented on HBASE-16212:
-

Test on master branch, shows that the problem is rare so far. and Thanks 
[~stack]! 

> Many connections to datanode are created when doing a large scan 
> -
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, HBASE-16212.v2.patch, 
> regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.
> hadoop version: 2.7.1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Issue Comment Deleted] (HBASE-16212) Many connections to datanode are created when doing a large scan

2016-07-20 Thread Zhihua Deng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhihua Deng updated HBASE-16212:

Comment: was deleted

(was: Test on master branch, shows that the problem is rare.  Thanks [~stack]!)

> Many connections to datanode are created when doing a large scan 
> -
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, HBASE-16212.v2.patch, 
> regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.
> hadoop version: 2.7.1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HBASE-16212) Many connections to datanode are created when doing a large scan

2016-07-20 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15385463#comment-15385463
 ] 

Zhihua Deng commented on HBASE-16212:
-

Test on master branch, shows that the problem is rare.  Thanks [~stack]!

> Many connections to datanode are created when doing a large scan 
> -
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, HBASE-16212.v2.patch, 
> regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.
> hadoop version: 2.7.1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HBASE-16212) Many connections to datanode are created when doing a large scan

2016-07-19 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15383514#comment-15383514
 ] 

Zhihua Deng edited comment on HBASE-16212 at 7/19/16 6:39 AM:
--

Thanks [~stack]. From the logging, it implies that different threads share the 
same DFSInputStream instance, say 'defaultRpcServer.handler=7'(handler7) and 
'defaultRpcServer.handler=4'(handler4), for example. The original will prefect 
the next block header and cache the header into thread. When 
defaultRpcServer.handler=4 comes,  it first checks that the cached header 
offset is equal to the the block starting offset, unfortunately these two 
numbers are unequal(-1 != offset). The handler4 knows nothing about the block 
header,  though the header has been prefected by handler7.  The handler4 needs 
to seek the inputstream with the block starting offset for the header,  while 
the inputstream has been over read by 33 bytes(the header size). So a new 
connection to datanode should be recreated, the elder one will be closed. When 
the datanode writes to a closed channel, a socket exception will be raised. 
When the same case happens frequently, the datanode will be suffered from 
logging the message described as it is.


was (Author: dengzh):
Thanks [~stack]. From the logging, it implies that different threads share the 
same DFSInputStream instance, say 'defaultRpcServer.handler=7'(handler7) and 
'defaultRpcServer.handler=4'(handler4), for example. The original will prefect 
the next block header and cache the header into thread. When 
defaultRpcServer.handler=4 comes,  it first checks that the cached header 
offset is equal to the the block starting offset, unfortunately these two 
numbers are unequal(-1 != offset). The handler4 knows nothing about the block 
header,  though the header has been prefected by handler7.  The handler4 needs 
to seek the inputstream with the block starting offset for obtaining the 
header,  while the inputstream has been over read by 33 bytes(the header size). 
So a new connection to datanode should be recreated, the elder one will be 
closed. When the datanode writes to a closed channel, a socket exception will 
be raised. When the same case happens frequently, the datanode will be suffered 
from logging the message described as it is.

> Many connections to datanode are created when doing a large scan 
> -
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, HBASE-16212.v2.patch, 
> regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.
> hadoop version: 2.7.1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HBASE-16212) Many connections to datanode are created when doing a large scan

2016-07-19 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15383565#comment-15383565
 ] 

Zhihua Deng edited comment on HBASE-16212 at 7/19/16 6:32 AM:
--

Changing from threadlocal to synchronization, yes there will be a potential 
synchronization bottleneck, but it better than io operation.  So the question 
here is that how often the connection will be recreated for seeking + reading? 
The original threadlocal is declared as non static private field here,  it 
means that the created fsreaderimpl instance will be reused later on, also an 
inputstream is initiated when fsreaderimpl created. 
Taken the case described in the attached log, The synchronization way is much 
better than threadlocal when acts as a sequential read .
How about concurrent case?  the worst case: Thread1.readBlockInternal -> 
Thread2.readBlockInternal -> Thread3.readBlockInternal -> 
Thread1.readBlockInternal -> 
In this case, the synchronization way is equal to threadlocal when taken how 
many connections will be created into consideration.





was (Author: dengzh):
Changing from threadlocal to synchronization, yes there will be a potential 
synchronization bottleneck, but it better than io operation.  So the question 
here is that how often the connection will be recreated for seeking + reading? 
The original threadlocal is declared as non static private field here,  it 
means that the created fsreaderimpl instance will be reused later on, also an 
inputstream is initiated when fsreaderimpl created. 
Taken the case described in the attached log, The synchronization way is more 
better than threadlocal when acts as a sequential read .
How about concurrent case?  the worst case: Thread1.readBlockInternal -> 
Thread2.readBlockInternal -> Thread3.readBlockInternal -> 
Thread1.readBlockInternal -> 
In this case, the synchronization way is equal to threadlocal when taken how 
many connections will be created into consideration.




> Many connections to datanode are created when doing a large scan 
> -
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, HBASE-16212.v2.patch, 
> regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.
> hadoop version: 2.7.1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HBASE-16212) Many connections to datanode are created when doing a large scan

2016-07-19 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15383565#comment-15383565
 ] 

Zhihua Deng edited comment on HBASE-16212 at 7/19/16 6:24 AM:
--

Changing from threadlocal to synchronization, yes there will be a potential 
synchronization bottleneck, but it better than io operation.  So the question 
here is that how often the connection will be recreated for seeking + reading? 
The original threadlocal is declared as non static private field here,  it 
means that the created fsreaderimpl instance will be reused later on, also an 
inputstream is initiated when fsreaderimpl created. 
Taken the case described in the attached log, The synchronization way is more 
better than threadlocal when acts as a sequential read .
How about concurrent case?  the worst case: Thread1.readBlockInternal -> 
Thread2.readBlockInternal -> Thread3.readBlockInternal -> 
Thread1.readBlockInternal -> 
In this case, the synchronization way is equal to threadlocal when taken how 
many connections will be created into consideration.





was (Author: dengzh):
Changing from threadlocal to a common field, yes there will be a potential 
synchronization bottleneck, but it better than io operation.  So the question 
here is that how often the connection will be recreated for seeking + reading? 
The original threadlocal is declared as non static private field here,  it 
means that the created fsreaderimpl instance will be reused later on, also a 
inputstream is initiated when fsreaderimpl created. 
Taken the case described in the attached log, The synchronization way is more 
better than threadlocal when acts as a sequential read .
How about concurrent case?  the worst case: Thread1.readBlockInternal -> 
Thread2.readBlockInternal -> Thread3.readBlockInternal -> 
Thread1.readBlockInternal -> 
In this case, the synchronization way is equal to threadlocal when taken how 
many connections will be created into consideration.




> Many connections to datanode are created when doing a large scan 
> -
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, HBASE-16212.v2.patch, 
> regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.
> hadoop version: 2.7.1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HBASE-16212) Many connections to datanode are created when doing a large scan

2016-07-19 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15383514#comment-15383514
 ] 

Zhihua Deng edited comment on HBASE-16212 at 7/19/16 6:22 AM:
--

Thanks [~stack]. From the logging, it implies that different threads share the 
same DFSInputStream instance, say 'defaultRpcServer.handler=7'(handler7) and 
'defaultRpcServer.handler=4'(handler4), for example. The original will prefect 
the next block header and cache the header into thread. When 
defaultRpcServer.handler=4 comes,  it first checks that the cached header 
offset is equal to the the block starting offset, unfortunately these two 
numbers are unequal(-1 != offset). The handler4 knows nothing about the block 
header,  though the header has been prefected by handler7.  The handler4 needs 
to seek the inputstream with the block starting offset for obtaining the 
header,  while the inputstream has been over read by 33 bytes(the header size). 
So a new connection to datanode should be recreated, the elder one will be 
closed. When the datanode writes to a closed channel, a socket exception will 
be raised. When the same case happens frequently, the datanode will be suffered 
from logging the message described as it is.


was (Author: dengzh):
Thanks [~stack]. From the logging, it implies that different threads share the 
same DFSInputStream instance, say 'defaultRpcServer.handler=7'(handler7) and 
'defaultRpcServer.handler=4'(handler4), for example. The original will prefect 
the next block header and cache the header into thread. When 
defaultRpcServer.handler=4 comes,  it first checks that the cached header 
offset is equal to the the block starting offset, unfortunately these two 
numbers are unequal(-1 != offset). The handler4 knows nothing about the block 
header,  though the header has been prefected by handler7.  The handler4 needs 
to seek the inputstream with the block starting offset for obtaining the 
header,  while the inputstream has been over read by 33 bytes(the header size). 
So a new connection to datanode should be recreated, the elder one will be 
closed. When the datanode writes to a closed channel, an socket exception will 
be raised. When the same case happens frequently, the datanode will be suffered 
from logging the message described as it is.

> Many connections to datanode are created when doing a large scan 
> -
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, HBASE-16212.v2.patch, 
> regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.
> hadoop version: 2.7.1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HBASE-16212) Many connections to datanode are created when doing a large scan

2016-07-18 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15383565#comment-15383565
 ] 

Zhihua Deng edited comment on HBASE-16212 at 7/19/16 4:30 AM:
--

Changing from threadlocal to a common field, yes there will be a potential 
synchronization bottleneck, but it better than io operation.  So the question 
here is that how often the connection will be recreated for seeking + reading? 
The original threadlocal is declared as non static private field here,  it 
means that the created fsreaderimpl instance will be reused later on, also a 
inputstream is initiated when fsreaderimpl created. 
Taken the case described in the attached log, The synchronization way is more 
better than threadlocal when acts as a sequential read .
How about concurrent case?  the worst case: Thread1.readBlockInternal -> 
Thread2.readBlockInternal -> Thread3.readBlockInternal -> 
Thread1.readBlockInternal -> 
In this case, the synchronization way is equal to threadlocal when taken how 
many connections will be created into consideration.





was (Author: dengzh):
Changing from threadlocal to a common field, yes there will be a potential 
synchronization bottleneck, but it better than io operation.  So the question 
here is that how often the connection will be recreated for seeking + reading? 
The original threadlocal is declared as non static private field here,  it 
means that the created fsreaderimpl instance will be reused later on, also a 
inputstream is initiated when fsreaderimpl created. 
Taken the case described in the attached log, The synchronization way is more 
better than threadlocal.
How about concurrent case?  the worst case: Thread1.readBlockInternal -> 
Thread2.readBlockInternal -> Thread3.readBlockInternal -> 
Thread1.readBlockInternal -> 
In this case, the synchronization way is equal to threadlocal when taken how 
many connections will be created into consideration.




> Many connections to datanode are created when doing a large scan 
> -
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, HBASE-16212.v2.patch, 
> regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.
> hadoop version: 2.7.1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HBASE-16212) Many connections to datanode are created when doing a large scan

2016-07-18 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15383565#comment-15383565
 ] 

Zhihua Deng edited comment on HBASE-16212 at 7/19/16 4:22 AM:
--

Changing from threadlocal to a common field, yes there will be a potential 
synchronization bottleneck, but it better than io operation.  So the question 
here is that how often the connection will be recreated for seeking + reading? 
The original threadlocal is declared as non static private field here,  it 
means that the created fsreaderimpl instance will be reused later on, also a 
inputstream is initiated when fsreaderimpl created. 
Taken the case described in the attached log, The synchronization way is more 
better than threadlocal.
How about concurrent case?  the worst case: Thread1.readBlockInternal -> 
Thread2.readBlockInternal -> Thread3.readBlockInternal -> 
Thread1.readBlockInternal -> 
In this case, the synchronization way is equal to threadlocal when taken how 
many connections will be created into consideration.





was (Author: dengzh):
Changing from threadlocal to a common field, yes there will be a potential 
synchronization bottleneck, but it better than io operation.  So the question 
here is that how often the connection will be recreated for seeking + reading? 
The original threadlocal is declared as non static private field here,  it 
means that the created fsreaderimpl instance will be reused later on, also a 
inputstream is initiated when fsreaderimpl created. 
Taken the case described in the attached log, The synchronization way is more 
better than threadlocal.
How about concurrent case?  the worst case: Thread1.readBlockInternal -> 
Thread2.readBlockInternal -> Thread3.readBlockInternal -> 
Thread1.readBlockInternal -> 
In this case, the synchronization way is equal to threadlocal when taken how 
many connections will be created into consideration.




> Many connections to datanode are created when doing a large scan 
> -
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, HBASE-16212.v2.patch, 
> regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.
> hadoop version: 2.7.1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HBASE-16212) Many connections to datanode are created when doing a large scan

2016-07-18 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15383565#comment-15383565
 ] 

Zhihua Deng commented on HBASE-16212:
-

Changing from threadlocal to a common field, yes there will be a potential 
synchronization bottleneck, but it better than io operation.  So the question 
here is that how often the connection will be recreated for seeking + reading? 
The original threadlocal is declared as non static private field here,  it 
means that the created fsreaderimpl instance will be reused later on, also a 
inputstream is initiated when fsreaderimpl created. 
Taken the case described in the attached log, The synchronization way is more 
better than threadlocal.
How about concurrent case?  the worst case: Thread1.readBlockInternal -> 
Thread2.readBlockInternal -> Thread3.readBlockInternal -> 
Thread1.readBlockInternal -> 
In this case, the synchronization way is equal to threadlocal when taken how 
many connections will be created into consideration.




> Many connections to datanode are created when doing a large scan 
> -
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, HBASE-16212.v2.patch, 
> regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.
> hadoop version: 2.7.1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HBASE-16212) Many connections to datanode are created when doing a large scan

2016-07-18 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15383516#comment-15383516
 ] 

Zhihua Deng edited comment on HBASE-16212 at 7/19/16 3:11 AM:
--

Test it on my cluster, a hmaster with two regionservers based on hbase-1.1.2, 
it works. Further test should carry on the master brunch


was (Author: dengzh):
Test it on my cluster, a hmaster with two regionservers based on hbase-1.1.2, 
it works. Further test should carry on the master brunch

> Many connections to datanode are created when doing a large scan 
> -
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, HBASE-16212.v2.patch, 
> regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.
> hadoop version: 2.7.1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HBASE-16212) Many connections to datanode are created when doing a large scan

2016-07-18 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15383517#comment-15383517
 ] 

Zhihua Deng commented on HBASE-16212:
-

Test it on my cluster, a hmaster with two regionservers based on hbase-1.1.2, 
it works. Further test should carry on the master brunch

> Many connections to datanode are created when doing a large scan 
> -
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, HBASE-16212.v2.patch, 
> regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.
> hadoop version: 2.7.1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HBASE-16212) Many connections to datanode are created when doing a large scan

2016-07-18 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15383516#comment-15383516
 ] 

Zhihua Deng commented on HBASE-16212:
-

Test it on my cluster, a hmaster with two regionservers based on hbase-1.1.2, 
it works. Further test should carry on the master brunch

> Many connections to datanode are created when doing a large scan 
> -
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, HBASE-16212.v2.patch, 
> regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.
> hadoop version: 2.7.1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Issue Comment Deleted] (HBASE-16212) Many connections to datanode are created when doing a large scan

2016-07-18 Thread Zhihua Deng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhihua Deng updated HBASE-16212:

Comment: was deleted

(was: Test it on my cluster, a hmaster with two regionservers based on 
hbase-1.1.2, it works. Further test should carry on the master brunch)

> Many connections to datanode are created when doing a large scan 
> -
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, HBASE-16212.v2.patch, 
> regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.
> hadoop version: 2.7.1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HBASE-16212) Many connections to datanode are created when doing a large scan

2016-07-18 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15383514#comment-15383514
 ] 

Zhihua Deng commented on HBASE-16212:
-

Thanks [~stack]. From the logging, it implies that different threads share the 
same DFSInputStream instance, say 'defaultRpcServer.handler=7'(handler7) and 
'defaultRpcServer.handler=4'(handler4), for example. The original will prefect 
the next block header and cache the header into thread. When 
defaultRpcServer.handler=4 comes,  it first checks that the cached header 
offset is equal to the the block starting offset, unfortunately these two 
numbers are unequal(-1 != offset). The handler4 knows nothing about the block 
header,  though the header has been prefected by handler7.  The handler4 needs 
to seek the inputstream with the block starting offset for obtaining the 
header,  while the inputstream has been over read by 33 bytes(the header size). 
So a new connection to datanode should be recreated, the elder one will be 
closed. When the datanode writes to a closed channel, an socket exception will 
be raised. When the same case happens frequently, the datanode will be suffered 
from logging the message described as it is.

> Many connections to datanode are created when doing a large scan 
> -
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, HBASE-16212.v2.patch, 
> regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.
> hadoop version: 2.7.1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HBASE-16212) Many connections to datanode are created when doing a large scan

2016-07-17 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15381648#comment-15381648
 ] 

Zhihua Deng commented on HBASE-16212:
-

Similar to the case of HBASE-10676.  [~lhofhansl], [~zhaojianbo],  could you 
help review this patch, thanks.

> Many connections to datanode are created when doing a large scan 
> -
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, HBASE-16212.v2.patch, 
> regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.
> hadoop version: 2.7.1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HBASE-16212) Many connections to datanode are created when doing a large scan

2016-07-15 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15379100#comment-15379100
 ] 

Zhihua Deng commented on HBASE-16212:
-

Remove the wrong unit test, The TestHFileBlock#testConcurrentReading test the 
case i want to.

> Many connections to datanode are created when doing a large scan 
> -
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, HBASE-16212.v2.patch, 
> regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.
> hadoop version: 2.7.1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HBASE-16212) Many connections to datanode are created when doing a large scan

2016-07-15 Thread Zhihua Deng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhihua Deng updated HBASE-16212:

Attachment: HBASE-16212.v2.patch

> Many connections to datanode are created when doing a large scan 
> -
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, HBASE-16212.v2.patch, 
> regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.
> hadoop version: 2.7.1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HBASE-16212) Many connections to datanode are created when doing a large scan

2016-07-15 Thread Zhihua Deng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhihua Deng updated HBASE-16212:

Description: 
As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
is suffering from logging the same repeatedly. Adding log to DFSInputStream, it 
outputs as follows:

2016-07-10 21:31:42,147 INFO  
[B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
111506876, targetPos: 111506843
 ...
As the pos of this input stream is larger than targetPos(the pos trying to 
seek), A new connection to the datanode will be created, the older one will be 
closed as a consequence. When the wrong seeking ops are large, the datanode's 
block scanner info message is spamming logs, as well as many connections to the 
same datanode will be created.

hadoop version: 2.7.1



  was:
As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
is suffering from logging the same repeatedly. Adding log to DFSInputStream, it 
outputs as follows:

2016-07-10 21:31:42,147 INFO  
[B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
111506876, targetPos: 111506843
 ...
As the pos of this input stream is larger than targetPos(the pos trying to 
seek), A new connection to the datanode will be created, the older one will be 
closed as a consequence. When the wrong seeking ops are large, the datanode's 
block scanner info message is spamming logs, as well as many connections to the 
same datanode will be created.




> Many connections to datanode are created when doing a large scan 
> -
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.
> hadoop version: 2.7.1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HBASE-16212) Many connections to datanode are created when doing a large scan

2016-07-15 Thread Zhihua Deng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhihua Deng updated HBASE-16212:

Summary: Many connections to datanode are created when doing a large scan   
(was: Many Connections are created by wrong seeking pos when doing a large scan 
)

> Many connections to datanode are created when doing a large scan 
> -
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HBASE-16212) Many Connections are created by wrong seeking pos when doing a large scan

2016-07-15 Thread Zhihua Deng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhihua Deng updated HBASE-16212:

Summary: Many Connections are created by wrong seeking pos when doing a 
large scan   (was: Many Connections are created by wrong seeking pos on 
InputStream)

> Many Connections are created by wrong seeking pos when doing a large scan 
> --
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-15 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15378986#comment-15378986
 ] 

Zhihua Deng edited comment on HBASE-16212 at 7/15/16 8:08 AM:
--

I add log details to  DFSInputStream#seek(long targetPos):
{code:borderStyle=solid}
if(pos > targetPos) {
  DFSClient.LOG.info(dfsClient.getClientName() + " seek " + 
getCurrentDatanode() + " for " + getCurrentBlock() +
  ". pos: " + pos + ", targetPos: " + targetPos);
}
{code}
The attached file named 'regionserver-dfsinputstream.log' shows the process.

Also in one of datanodes throws an exceptions of client's sudden close:

java.net.SocketException: Original Exception : java.io.IOException: Connection 
reset by peer
at sun.nio.ch.FileChannelImpl.transferTo0(Native Method)
at 
sun.nio.ch.FileChannelImpl.transferToDirectlyInternal(FileChannelImpl.java:427)
at 
sun.nio.ch.FileChannelImpl.transferToDirectly(FileChannelImpl.java:492)
at sun.nio.ch.FileChannelImpl.transferTo(FileChannelImpl.java:607)
at 
org.apache.hadoop.net.SocketOutputStream.transferToFully(SocketOutputStream.java:223)
at 
org.apache.hadoop.hdfs.server.datanode.BlockSender.sendPacket(BlockSender.java:579)
at 
org.apache.hadoop.hdfs.server.datanode.BlockSender.doSendBlock(BlockSender.java:759)
at 
org.apache.hadoop.hdfs.server.datanode.BlockSender.sendBlock(BlockSender.java:706)
at 
org.apache.hadoop.hdfs.server.datanode.DataXceiver.readBlock(DataXceiver.java:551)
at 
org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opReadBlock(Receiver.java:116)
at 
org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:71)
at 
org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:251)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.io.IOException: Connection reset by peer
... 13 more





was (Author: dengzh):
I add log details to  DFSInputStream#seek(long targetPos):
{code:borderStyle=solid}
if(pos > targetPos) {
  DFSClient.LOG.info(dfsClient.getClientName() + " seek " + 
getCurrentDatanode() + " for " + getCurrentBlock() +
  ". pos: " + pos + ", targetPos: " + targetPos);
}
{code}
The attached file 'regionserver-dfsinputstream.log' shows the process.

> Many Connections are created by wrong seeking pos on InputStream
> 
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-15 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15378986#comment-15378986
 ] 

Zhihua Deng edited comment on HBASE-16212 at 7/15/16 8:09 AM:
--

I add log details to  DFSInputStream#seek(long targetPos):
{code:borderStyle=solid}
if(pos > targetPos) {
  DFSClient.LOG.info(dfsClient.getClientName() + " seek " + 
getCurrentDatanode() + " for " + getCurrentBlock() +
  ". pos: " + pos + ", targetPos: " + targetPos);
}
{code}
The attached file named 'regionserver-dfsinputstream.log' shows the process.

Also in one of datanodes throw an exception of such a sudden close by the 
client side:
java.net.SocketException: Original Exception : java.io.IOException: Connection 
reset by peer
at sun.nio.ch.FileChannelImpl.transferTo0(Native Method)
at 
sun.nio.ch.FileChannelImpl.transferToDirectlyInternal(FileChannelImpl.java:427)
at 
sun.nio.ch.FileChannelImpl.transferToDirectly(FileChannelImpl.java:492)
at sun.nio.ch.FileChannelImpl.transferTo(FileChannelImpl.java:607)
at 
org.apache.hadoop.net.SocketOutputStream.transferToFully(SocketOutputStream.java:223)
at 
org.apache.hadoop.hdfs.server.datanode.BlockSender.sendPacket(BlockSender.java:579)
at 
org.apache.hadoop.hdfs.server.datanode.BlockSender.doSendBlock(BlockSender.java:759)
at 
org.apache.hadoop.hdfs.server.datanode.BlockSender.sendBlock(BlockSender.java:706)
at 
org.apache.hadoop.hdfs.server.datanode.DataXceiver.readBlock(DataXceiver.java:551)
at 
org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opReadBlock(Receiver.java:116)
at 
org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:71)
at 
org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:251)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.io.IOException: Connection reset by peer
... 13 more





was (Author: dengzh):
I add log details to  DFSInputStream#seek(long targetPos):
{code:borderStyle=solid}
if(pos > targetPos) {
  DFSClient.LOG.info(dfsClient.getClientName() + " seek " + 
getCurrentDatanode() + " for " + getCurrentBlock() +
  ". pos: " + pos + ", targetPos: " + targetPos);
}
{code}
The attached file named 'regionserver-dfsinputstream.log' shows the process.

Also in one of datanodes throws an exceptions of client's sudden close:

java.net.SocketException: Original Exception : java.io.IOException: Connection 
reset by peer
at sun.nio.ch.FileChannelImpl.transferTo0(Native Method)
at 
sun.nio.ch.FileChannelImpl.transferToDirectlyInternal(FileChannelImpl.java:427)
at 
sun.nio.ch.FileChannelImpl.transferToDirectly(FileChannelImpl.java:492)
at sun.nio.ch.FileChannelImpl.transferTo(FileChannelImpl.java:607)
at 
org.apache.hadoop.net.SocketOutputStream.transferToFully(SocketOutputStream.java:223)
at 
org.apache.hadoop.hdfs.server.datanode.BlockSender.sendPacket(BlockSender.java:579)
at 
org.apache.hadoop.hdfs.server.datanode.BlockSender.doSendBlock(BlockSender.java:759)
at 
org.apache.hadoop.hdfs.server.datanode.BlockSender.sendBlock(BlockSender.java:706)
at 
org.apache.hadoop.hdfs.server.datanode.DataXceiver.readBlock(DataXceiver.java:551)
at 
org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opReadBlock(Receiver.java:116)
at 
org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:71)
at 
org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:251)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.io.IOException: Connection reset by peer
... 13 more




> Many Connections are created by wrong seeking pos on InputStream
> 
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a 

[jira] [Comment Edited] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-15 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15378986#comment-15378986
 ] 

Zhihua Deng edited comment on HBASE-16212 at 7/15/16 8:03 AM:
--

I add log details to  DFSInputStream#seek(long targetPos):
{code:borderStyle=solid}
if(pos > targetPos) {
  DFSClient.LOG.info(dfsClient.getClientName() + " seek " + 
getCurrentDatanode() + " for " + getCurrentBlock() +
  ". pos: " + pos + ", targetPos: " + targetPos);
}
{code}
The attached file 'regionserver-dfsinputstream.log' shows the process.


was (Author: dengzh):
I add log details to  DFSInputStream#seek(long targetPos):

if(pos > targetPos) {
  DFSClient.LOG.info(dfsClient.getClientName() + " seek " + 
getCurrentDatanode() + " for " + getCurrentBlock() +
  ". pos: " + pos + ", targetPos: " + targetPos);
}

The attached file 'regionserver-dfsinputstream.log' shows the process.

> Many Connections are created by wrong seeking pos on InputStream
> 
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-15 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15378986#comment-15378986
 ] 

Zhihua Deng commented on HBASE-16212:
-

I add log details to  DFSInputStream#seek(long targetPos):

if(pos > targetPos) {
  DFSClient.LOG.info(dfsClient.getClientName() + " seek " + 
getCurrentDatanode() + " for " + getCurrentBlock() +
  ". pos: " + pos + ", targetPos: " + targetPos);
}

The attached file 'regionserver-dfsinputstream.log' shows the process.

> Many Connections are created by wrong seeking pos on InputStream
> 
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-15 Thread Zhihua Deng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhihua Deng updated HBASE-16212:

Attachment: regionserver-dfsinputstream.log

> Many Connections are created by wrong seeking pos on InputStream
> 
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch, regionserver-dfsinputstream.log
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-15 Thread Zhihua Deng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhihua Deng updated HBASE-16212:

Attachment: HBASE-16212.patch

> Many Connections are created by wrong seeking pos on InputStream
> 
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
> Attachments: HBASE-16212.patch
>
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-15 Thread Zhihua Deng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhihua Deng updated HBASE-16212:

Status: Open  (was: Patch Available)

> Many Connections are created by wrong seeking pos on InputStream
> 
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-15 Thread Zhihua Deng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhihua Deng updated HBASE-16212:

Status: Patch Available  (was: Open)

> Many Connections are created by wrong seeking pos on InputStream
> 
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-15 Thread Zhihua Deng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhihua Deng updated HBASE-16212:

  Priority: Major  (was: Trivial)
Issue Type: Improvement  (was: Test)

> Many Connections are created by wrong seeking pos on InputStream
> 
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Issue Comment Deleted] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-15 Thread Zhihua Deng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhihua Deng updated HBASE-16212:

Comment: was deleted

(was: This jira should be closed as the problem this jira describes is ok.)

> Many Connections are created by wrong seeking pos on InputStream
> 
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Test
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
>Priority: Trivial
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-12 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15372772#comment-15372772
 ] 

Zhihua Deng commented on HBASE-16212:
-

This jira should be closed as the problem this jira describes is ok.

> Many Connections are created by wrong seeking pos on InputStream
> 
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Test
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
>Priority: Trivial
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-12 Thread Zhihua Deng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhihua Deng updated HBASE-16212:

Issue Type: Test  (was: Bug)

> Many Connections are created by wrong seeking pos on InputStream
> 
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Test
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
>Priority: Trivial
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-12 Thread Zhihua Deng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhihua Deng updated HBASE-16212:

Priority: Trivial  (was: Major)

> Many Connections are created by wrong seeking pos on InputStream
> 
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
>Priority: Trivial
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Issue Comment Deleted] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-12 Thread Zhihua Deng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhihua Deng updated HBASE-16212:

Comment: was deleted

(was: Some related logs from region server:
2016-07-10 21:31:42,147 INFO  
[B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
111506876, targetPos: 111506843
2016-07-10 21:31:42,715 INFO  
[B.defaultRpcServer.handler=25,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
113544644, targetPos: 113544611
2016-07-10 21:31:43,341 INFO  
[B.defaultRpcServer.handler=27,queue=0,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
115547269, targetPos: 115547236
2016-07-10 21:31:43,950 INFO  
[B.defaultRpcServer.handler=16,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
117532235, targetPos: 117532202
2016-07-10 21:31:43,988 INFO  [B.defaultRpcServer.handler=7,queue=1,port=16020] 
hdfs.DFSClient: DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
119567867, targetPos: 119567834
2016-07-10 21:31:45,228 INFO  
[B.defaultRpcServer.handler=19,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
121787264, targetPos: 121787231
2016-07-10 21:31:45,254 INFO  [B.defaultRpcServer.handler=0,queue=0,port=16020] 
hdfs.DFSClient: DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
123910032, targetPos: 12390
2016-07-10 21:31:46,402 INFO  
[B.defaultRpcServer.handler=26,queue=2,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
125905163, targetPos: 125905130
2016-07-10 21:31:47,027 INFO  
[B.defaultRpcServer.handler=24,queue=0,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
128028947, targetPos: 128028914
2016-07-10 21:31:47,649 INFO  
[B.defaultRpcServer.handler=10,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
130057763, targetPos: 130057730
)

> Many Connections are created by wrong seeking pos on InputStream
> 
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Issue Comment Deleted] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-12 Thread Zhihua Deng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhihua Deng updated HBASE-16212:

Comment: was deleted

(was: When the datanode writes to the closed connection, an exception will 
occur:
2016-06-30 11:21:34,320 TRACE org.apache.hadoop.hdfs.server.datanode.DataNode: 
DatanodeRegistration(10.130.1.29:50010, 
datanodeUuid=f3d795cc-2b3b-43b9-90c3-e4157c031d2c, infoPort=50075, 
infoSecurePort=0, ipcPort=50020, 
storageInfo=lv=-56;cid=CID-a99b693d-6f26-48fe-ad37-9f8162f70b22;nsid=920937379;c=0):Ignoring
 exception while serving 
BP-360285305-10.130.1.11-1444619256876:blk_1105510536_31776579 to 
/10.130.1.21:39933
java.net.SocketException: Original Exception : java.io.IOException: Connection 
reset by peer
at sun.nio.ch.FileChannelImpl.transferTo0(Native Method)
at 
sun.nio.ch.FileChannelImpl.transferToDirectlyInternal(FileChannelImpl.java:427)
at 
sun.nio.ch.FileChannelImpl.transferToDirectly(FileChannelImpl.java:492)
at sun.nio.ch.FileChannelImpl.transferTo(FileChannelImpl.java:607)
at 
org.apache.hadoop.net.SocketOutputStream.transferToFully(SocketOutputStream.java:223)
at 
org.apache.hadoop.hdfs.server.datanode.BlockSender.sendPacket(BlockSender.java:579)
at 
org.apache.hadoop.hdfs.server.datanode.BlockSender.doSendBlock(BlockSender.java:759)
at 
org.apache.hadoop.hdfs.server.datanode.BlockSender.sendBlock(BlockSender.java:706)
at 
org.apache.hadoop.hdfs.server.datanode.DataXceiver.readBlock(DataXceiver.java:551)
at 
org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opReadBlock(Receiver.java:116)
at 
org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:71)
at 
org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:251)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.io.IOException: Connection reset by peer
... 13 more
By analyzing the regionserver's log,  pos - targetPos = 33, which is equal to 
the length of block header, seen from HFileBlock.java.)

> Many Connections are created by wrong seeking pos on InputStream
> 
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Issue Comment Deleted] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-12 Thread Zhihua Deng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhihua Deng updated HBASE-16212:

Comment: was deleted

(was: The stack trace looks like this:

at org.apache.hadoop.hdfs.DFSInputStream.seek(DFSInputStream.java:1509)
at 
org.apache.hadoop.fs.FSDataInputStream.seek(FSDataInputStream.java:62)
at 
org.apache.hadoop.hbase.io.hfile.HFileBlock$AbstractFSReader.readAtOffset(HFileBlock.java:1357)
at 
org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderImpl.readBlockDataInternal(HFileBlock.java:1591)
at 
org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderImpl.readBlockData(HFileBlock.java:1470)
at 
org.apache.hadoop.hbase.io.hfile.HFileReaderV2.readBlock(HFileReaderV2.java:437)
at 
org.apache.hadoop.hbase.io.hfile.HFileReaderV2$AbstractScannerV2.readNextDataBlock(HFileReaderV2.java:708)
at 
org.apache.hadoop.hbase.io.hfile.HFileReaderV2$ScannerV2.isNextBlock(HFileReaderV2.java:833)
at 
org.apache.hadoop.hbase.io.hfile.HFileReaderV2$ScannerV2.positionForNextBlock(HFileReaderV2.java:828)
at 
org.apache.hadoop.hbase.io.hfile.HFileReaderV2$ScannerV2._next(HFileReaderV2.java:845)
at 
org.apache.hadoop.hbase.io.hfile.HFileReaderV2$ScannerV2.next(HFileReaderV2.java:865)
at 
org.apache.hadoop.hbase.regionserver.StoreFileScanner.next(StoreFileScanner.java:139)
at 
org.apache.hadoop.hbase.regionserver.KeyValueHeap.next(KeyValueHeap.java:108)
at 
org.apache.hadoop.hbase.regionserver.StoreScanner.next(StoreScanner.java:596)
at 
org.apache.hadoop.hbase.regionserver.KeyValueHeap.next(KeyValueHeap.java:147)
at 
org.apache.hadoop.hbase.regionserver.HRegion$RegionScannerImpl.populateResult(HRegion.java:5486)
at 
org.apache.hadoop.hbase.regionserver.HRegion$RegionScannerImpl.nextInternal(HRegion.java:5637)
at 
org.apache.hadoop.hbase.regionserver.HRegion$RegionScannerImpl.nextRaw(HRegion.java:5424)
at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.scan(RSRpcServices.java:2395))

> Many Connections are created by wrong seeking pos on InputStream
> 
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Issue Comment Deleted] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-12 Thread Zhihua Deng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhihua Deng updated HBASE-16212:

Comment: was deleted

(was: Yes, these are logs i added to the DFSInputStream#seek(long targetPos):
if(pos > targetPos) {
  DFSClient.LOG.info(dfsClient.getClientName() + " seek " + 
getCurrentDatanode() + " for " + getCurrentBlock() +
  ". pos: " + pos + ", targetPos: " + targetPos);
}

>From the log, it implies that the client(regionserver) reads behind the 
>inputstream with 33 bytes)

> Many Connections are created by wrong seeking pos on InputStream
> 
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-12 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15372280#comment-15372280
 ] 

Zhihua Deng commented on HBASE-16212:
-

The stack trace looks like this:

at org.apache.hadoop.hdfs.DFSInputStream.seek(DFSInputStream.java:1509)
at 
org.apache.hadoop.fs.FSDataInputStream.seek(FSDataInputStream.java:62)
at 
org.apache.hadoop.hbase.io.hfile.HFileBlock$AbstractFSReader.readAtOffset(HFileBlock.java:1357)
at 
org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderImpl.readBlockDataInternal(HFileBlock.java:1591)
at 
org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderImpl.readBlockData(HFileBlock.java:1470)
at 
org.apache.hadoop.hbase.io.hfile.HFileReaderV2.readBlock(HFileReaderV2.java:437)
at 
org.apache.hadoop.hbase.io.hfile.HFileReaderV2$AbstractScannerV2.readNextDataBlock(HFileReaderV2.java:708)
at 
org.apache.hadoop.hbase.io.hfile.HFileReaderV2$ScannerV2.isNextBlock(HFileReaderV2.java:833)
at 
org.apache.hadoop.hbase.io.hfile.HFileReaderV2$ScannerV2.positionForNextBlock(HFileReaderV2.java:828)
at 
org.apache.hadoop.hbase.io.hfile.HFileReaderV2$ScannerV2._next(HFileReaderV2.java:845)
at 
org.apache.hadoop.hbase.io.hfile.HFileReaderV2$ScannerV2.next(HFileReaderV2.java:865)
at 
org.apache.hadoop.hbase.regionserver.StoreFileScanner.next(StoreFileScanner.java:139)
at 
org.apache.hadoop.hbase.regionserver.KeyValueHeap.next(KeyValueHeap.java:108)
at 
org.apache.hadoop.hbase.regionserver.StoreScanner.next(StoreScanner.java:596)
at 
org.apache.hadoop.hbase.regionserver.KeyValueHeap.next(KeyValueHeap.java:147)
at 
org.apache.hadoop.hbase.regionserver.HRegion$RegionScannerImpl.populateResult(HRegion.java:5486)
at 
org.apache.hadoop.hbase.regionserver.HRegion$RegionScannerImpl.nextInternal(HRegion.java:5637)
at 
org.apache.hadoop.hbase.regionserver.HRegion$RegionScannerImpl.nextRaw(HRegion.java:5424)
at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.scan(RSRpcServices.java:2395)

> Many Connections are created by wrong seeking pos on InputStream
> 
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-12 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15372219#comment-15372219
 ] 

Zhihua Deng commented on HBASE-16212:
-

Yes, these are logs i added to the DFSInputStream#seek(long targetPos):
if(pos > targetPos) {
  DFSClient.LOG.info(dfsClient.getClientName() + " seek " + 
getCurrentDatanode() + " for " + getCurrentBlock() +
  ". pos: " + pos + ", targetPos: " + targetPos);
}

>From the log, it implies that the client(regionserver) reads behind the 
>inputstream with 33 bytes

> Many Connections are created by wrong seeking pos on InputStream
> 
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-11 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15372133#comment-15372133
 ] 

Zhihua Deng edited comment on HBASE-16212 at 7/12/16 3:44 AM:
--

Some related logs from region server:

2016-07-10 21:31:42,147 INFO  
[B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
111506876, targetPos: 111506843
2016-07-10 21:31:42,715 INFO  
[B.defaultRpcServer.handler=25,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
113544644, targetPos: 113544611
2016-07-10 21:31:43,341 INFO  
[B.defaultRpcServer.handler=27,queue=0,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
115547269, targetPos: 115547236
2016-07-10 21:31:43,950 INFO  
[B.defaultRpcServer.handler=16,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
117532235, targetPos: 117532202
2016-07-10 21:31:43,988 INFO  [B.defaultRpcServer.handler=7,queue=1,port=16020] 
hdfs.DFSClient: DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
119567867, targetPos: 119567834
2016-07-10 21:31:45,228 INFO  
[B.defaultRpcServer.handler=19,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
121787264, targetPos: 121787231
2016-07-10 21:31:45,254 INFO  [B.defaultRpcServer.handler=0,queue=0,port=16020] 
hdfs.DFSClient: DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
123910032, targetPos: 12390
2016-07-10 21:31:46,402 INFO  
[B.defaultRpcServer.handler=26,queue=2,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
125905163, targetPos: 125905130
2016-07-10 21:31:47,027 INFO  
[B.defaultRpcServer.handler=24,queue=0,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
128028947, targetPos: 128028914
2016-07-10 21:31:47,649 INFO  
[B.defaultRpcServer.handler=10,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
130057763, targetPos: 130057730



was (Author: dengzh):
Some related logs from region server:

2016-07-10 21:31:42,147 INFO  
[B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
111506876, targetPos: 111506843
2016-07-10 21:31:42,715 INFO  
[B.defaultRpcServer.handler=25,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
113544644, targetPos: 113544611
2016-07-10 21:31:43,341 INFO  
[B.defaultRpcServer.handler=27,queue=0,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
115547269, targetPos: 115547236
2016-07-10 21:31:43,950 INFO  
[B.defaultRpcServer.handler=16,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for 

[jira] [Comment Edited] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-11 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15372133#comment-15372133
 ] 

Zhihua Deng edited comment on HBASE-16212 at 7/12/16 3:44 AM:
--

Some related logs from region server:
2016-07-10 21:31:42,147 INFO  
[B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
111506876, targetPos: 111506843
2016-07-10 21:31:42,715 INFO  
[B.defaultRpcServer.handler=25,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
113544644, targetPos: 113544611
2016-07-10 21:31:43,341 INFO  
[B.defaultRpcServer.handler=27,queue=0,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
115547269, targetPos: 115547236
2016-07-10 21:31:43,950 INFO  
[B.defaultRpcServer.handler=16,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
117532235, targetPos: 117532202
2016-07-10 21:31:43,988 INFO  [B.defaultRpcServer.handler=7,queue=1,port=16020] 
hdfs.DFSClient: DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
119567867, targetPos: 119567834
2016-07-10 21:31:45,228 INFO  
[B.defaultRpcServer.handler=19,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
121787264, targetPos: 121787231
2016-07-10 21:31:45,254 INFO  [B.defaultRpcServer.handler=0,queue=0,port=16020] 
hdfs.DFSClient: DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
123910032, targetPos: 12390
2016-07-10 21:31:46,402 INFO  
[B.defaultRpcServer.handler=26,queue=2,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
125905163, targetPos: 125905130
2016-07-10 21:31:47,027 INFO  
[B.defaultRpcServer.handler=24,queue=0,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
128028947, targetPos: 128028914
2016-07-10 21:31:47,649 INFO  
[B.defaultRpcServer.handler=10,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
130057763, targetPos: 130057730



was (Author: dengzh):
Some related logs from region server:

2016-07-10 21:31:42,147 INFO  
[B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
111506876, targetPos: 111506843
2016-07-10 21:31:42,715 INFO  
[B.defaultRpcServer.handler=25,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
113544644, targetPos: 113544611
2016-07-10 21:31:43,341 INFO  
[B.defaultRpcServer.handler=27,queue=0,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
115547269, targetPos: 115547236
2016-07-10 21:31:43,950 INFO  
[B.defaultRpcServer.handler=16,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for 

[jira] [Commented] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-11 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15372133#comment-15372133
 ] 

Zhihua Deng commented on HBASE-16212:
-

Some related logs from region server:

2016-07-10 21:31:42,147 INFO  
[B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
111506876, targetPos: 111506843
2016-07-10 21:31:42,715 INFO  
[B.defaultRpcServer.handler=25,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
113544644, targetPos: 113544611
2016-07-10 21:31:43,341 INFO  
[B.defaultRpcServer.handler=27,queue=0,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
115547269, targetPos: 115547236
2016-07-10 21:31:43,950 INFO  
[B.defaultRpcServer.handler=16,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
117532235, targetPos: 117532202
2016-07-10 21:31:43,988 INFO  [B.defaultRpcServer.handler=7,queue=1,port=16020] 
hdfs.DFSClient: DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
119567867, targetPos: 119567834
2016-07-10 21:31:45,228 INFO  
[B.defaultRpcServer.handler=19,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
121787264, targetPos: 121787231
2016-07-10 21:31:45,254 INFO  [B.defaultRpcServer.handler=0,queue=0,port=16020] 
hdfs.DFSClient: DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
123910032, targetPos: 12390
2016-07-10 21:31:46,402 INFO  
[B.defaultRpcServer.handler=26,queue=2,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
125905163, targetPos: 125905130
2016-07-10 21:31:47,027 INFO  
[B.defaultRpcServer.handler=24,queue=0,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
128028947, targetPos: 128028914
2016-07-10 21:31:47,649 INFO  
[B.defaultRpcServer.handler=10,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
130057763, targetPos: 130057730


> Many Connections are created by wrong seeking pos on InputStream
> 
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-11 Thread Zhihua Deng (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15372130#comment-15372130
 ] 

Zhihua Deng commented on HBASE-16212:
-

When the datanode writes to the closed connection, an exception will occur:
2016-06-30 11:21:34,320 TRACE org.apache.hadoop.hdfs.server.datanode.DataNode: 
DatanodeRegistration(10.130.1.29:50010, 
datanodeUuid=f3d795cc-2b3b-43b9-90c3-e4157c031d2c, infoPort=50075, 
infoSecurePort=0, ipcPort=50020, 
storageInfo=lv=-56;cid=CID-a99b693d-6f26-48fe-ad37-9f8162f70b22;nsid=920937379;c=0):Ignoring
 exception while serving 
BP-360285305-10.130.1.11-1444619256876:blk_1105510536_31776579 to 
/10.130.1.21:39933
java.net.SocketException: Original Exception : java.io.IOException: Connection 
reset by peer
at sun.nio.ch.FileChannelImpl.transferTo0(Native Method)
at 
sun.nio.ch.FileChannelImpl.transferToDirectlyInternal(FileChannelImpl.java:427)
at 
sun.nio.ch.FileChannelImpl.transferToDirectly(FileChannelImpl.java:492)
at sun.nio.ch.FileChannelImpl.transferTo(FileChannelImpl.java:607)
at 
org.apache.hadoop.net.SocketOutputStream.transferToFully(SocketOutputStream.java:223)
at 
org.apache.hadoop.hdfs.server.datanode.BlockSender.sendPacket(BlockSender.java:579)
at 
org.apache.hadoop.hdfs.server.datanode.BlockSender.doSendBlock(BlockSender.java:759)
at 
org.apache.hadoop.hdfs.server.datanode.BlockSender.sendBlock(BlockSender.java:706)
at 
org.apache.hadoop.hdfs.server.datanode.DataXceiver.readBlock(DataXceiver.java:551)
at 
org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opReadBlock(Receiver.java:116)
at 
org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:71)
at 
org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:251)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.io.IOException: Connection reset by peer
... 13 more
By analyzing the regionserver's log,  pos - targetPos = 33, which is equal to 
the length of block header, seen from HFileBlock.java.

> Many Connections are created by wrong seeking pos on InputStream
> 
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
>
> As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
> is suffering from logging the same repeatedly. Adding log to DFSInputStream, 
> it outputs as follows:
> 2016-07-10 21:31:42,147 INFO  
> [B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
> DFSClient_NONMAPREDUCE_1984924661_1 seek 
> DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
>  for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
> 111506876, targetPos: 111506843
>  ...
> As the pos of this input stream is larger than targetPos(the pos trying to 
> seek), A new connection to the datanode will be created, the older one will 
> be closed as a consequence. When the wrong seeking ops are large, the 
> datanode's block scanner info message is spamming logs, as well as many 
> connections to the same datanode will be created.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-11 Thread Zhihua Deng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhihua Deng updated HBASE-16212:

Description: 
As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
is suffering from logging the same repeatedly. Adding log to DFSInputStream, it 
outputs as follows:

2016-07-10 21:31:42,147 INFO  
[B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
111506876, targetPos: 111506843
 ...
As the pos of this input stream is larger than targetPos(the pos trying to 
seek), A new connection to the datanode will be created, the older one will be 
closed as a consequence. When the wrong seeking ops are large, the datanode's 
block scanner info message is spamming logs, as well as many connections to the 
same datanode will be created.



  was:
As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
is suffering from logging the same repeatedly. Adding log to DFSInputStream, it 
outputs as follows:

 2016-07-10 21:31:42,147 INFO  
[B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
111506876, targetPos: 111506843
2016-07-10 21:31:42,715 INFO  
[B.defaultRpcServer.handler=25,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
113544644, targetPos: 113544611
2016-07-10 21:31:43,341 INFO  
[B.defaultRpcServer.handler=27,queue=0,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
115547269, targetPos: 115547236
2016-07-10 21:31:43,950 INFO  
[B.defaultRpcServer.handler=16,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
117532235, targetPos: 117532202
2016-07-10 21:31:43,988 INFO  [B.defaultRpcServer.handler=7,queue=1,port=16020] 
hdfs.DFSClient: DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
119567867, targetPos: 119567834
2016-07-10 21:31:45,228 INFO  
[B.defaultRpcServer.handler=19,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
121787264, targetPos: 121787231
2016-07-10 21:31:45,254 INFO  [B.defaultRpcServer.handler=0,queue=0,port=16020] 
hdfs.DFSClient: DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
123910032, targetPos: 12390
2016-07-10 21:31:46,402 INFO  
[B.defaultRpcServer.handler=26,queue=2,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
125905163, targetPos: 125905130
2016-07-10 21:31:47,027 INFO  
[B.defaultRpcServer.handler=24,queue=0,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
128028947, targetPos: 128028914
2016-07-10 21:31:47,649 INFO  
[B.defaultRpcServer.handler=10,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
130057763, targetPos: 130057730



> Many Connections are created by wrong seeking pos on InputStream
> 
>
> Key: HBASE-16212
> URL: https://issues.apache.org/jira/browse/HBASE-16212
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.1.2
>Reporter: Zhihua Deng
>

[jira] [Updated] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-11 Thread Zhihua Deng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhihua Deng updated HBASE-16212:

Description: 
As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
is suffering from logging the same repeatedly. Adding log to DFSInputStream, it 
outputs as follows:

 2016-07-10 21:31:42,147 INFO  
[B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
111506876, targetPos: 111506843
2016-07-10 21:31:42,715 INFO  
[B.defaultRpcServer.handler=25,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
113544644, targetPos: 113544611
2016-07-10 21:31:43,341 INFO  
[B.defaultRpcServer.handler=27,queue=0,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
115547269, targetPos: 115547236
2016-07-10 21:31:43,950 INFO  
[B.defaultRpcServer.handler=16,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
117532235, targetPos: 117532202
2016-07-10 21:31:43,988 INFO  [B.defaultRpcServer.handler=7,queue=1,port=16020] 
hdfs.DFSClient: DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
119567867, targetPos: 119567834
2016-07-10 21:31:45,228 INFO  
[B.defaultRpcServer.handler=19,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
121787264, targetPos: 121787231
2016-07-10 21:31:45,254 INFO  [B.defaultRpcServer.handler=0,queue=0,port=16020] 
hdfs.DFSClient: DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
123910032, targetPos: 12390
2016-07-10 21:31:46,402 INFO  
[B.defaultRpcServer.handler=26,queue=2,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
125905163, targetPos: 125905130
2016-07-10 21:31:47,027 INFO  
[B.defaultRpcServer.handler=24,queue=0,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
128028947, targetPos: 128028914
2016-07-10 21:31:47,649 INFO  
[B.defaultRpcServer.handler=10,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
130057763, targetPos: 130057730


  was:
As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
is suffering from logging the same repeatedly. Adding log to DFSInputStream, it 
outputs as follows:
 2016-07-10 21:31:42,147 INFO  
[B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
111506876, targetPos: 111506843
2016-07-10 21:31:42,715 INFO  
[B.defaultRpcServer.handler=25,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
113544644, targetPos: 113544611
2016-07-10 21:31:43,341 INFO  
[B.defaultRpcServer.handler=27,queue=0,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
115547269, targetPos: 115547236
2016-07-10 21:31:43,950 INFO  

[jira] [Created] (HBASE-16212) Many Connections are created by wrong seeking pos on InputStream

2016-07-11 Thread Zhihua Deng (JIRA)
Zhihua Deng created HBASE-16212:
---

 Summary: Many Connections are created by wrong seeking pos on 
InputStream
 Key: HBASE-16212
 URL: https://issues.apache.org/jira/browse/HBASE-16212
 Project: HBase
  Issue Type: Bug
Affects Versions: 1.1.2
Reporter: Zhihua Deng


As described in https://issues.apache.org/jira/browse/HDFS-8659, the datanode 
is suffering from logging the same repeatedly. Adding log to DFSInputStream, it 
outputs as follows:
 2016-07-10 21:31:42,147 INFO  
[B.defaultRpcServer.handler=22,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
111506876, targetPos: 111506843
2016-07-10 21:31:42,715 INFO  
[B.defaultRpcServer.handler=25,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
113544644, targetPos: 113544611
2016-07-10 21:31:43,341 INFO  
[B.defaultRpcServer.handler=27,queue=0,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
115547269, targetPos: 115547236
2016-07-10 21:31:43,950 INFO  
[B.defaultRpcServer.handler=16,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
117532235, targetPos: 117532202
2016-07-10 21:31:43,988 INFO  [B.defaultRpcServer.handler=7,queue=1,port=16020] 
hdfs.DFSClient: DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
119567867, targetPos: 119567834
2016-07-10 21:31:45,228 INFO  
[B.defaultRpcServer.handler=19,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
121787264, targetPos: 121787231
2016-07-10 21:31:45,254 INFO  [B.defaultRpcServer.handler=0,queue=0,port=16020] 
hdfs.DFSClient: DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
123910032, targetPos: 12390
2016-07-10 21:31:46,402 INFO  
[B.defaultRpcServer.handler=26,queue=2,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
125905163, targetPos: 125905130
2016-07-10 21:31:47,027 INFO  
[B.defaultRpcServer.handler=24,queue=0,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
128028947, targetPos: 128028914
2016-07-10 21:31:47,649 INFO  
[B.defaultRpcServer.handler=10,queue=1,port=16020] hdfs.DFSClient: 
DFSClient_NONMAPREDUCE_1984924661_1 seek 
DatanodeInfoWithStorage[10.130.1.29:50010,DS-086bc494-d862-470c-86e8-9cb7929985c6,DISK]
 for BP-360285305-10.130.1.11-1444619256876:blk_1109360829_35627143. pos: 
130057763, targetPos: 130057730



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)