[
https://issues.apache.org/jira/browse/HBASE-5322?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13440201#comment-13440201
]
Ian Varley commented on HBASE-5322:
-----------------------------------
A couple things:
- Have you tried it with caching turned off entirely? If you're caching 300
rows, it may be that it's constantly replacing those 300 rows as you scan
billions of rows
- It's also possible that your full table scan simply takes longer than the
timeout; if so, that's not a bug that needs fixing in HBase, that's just how it
is.
I know that it's possible to scan large numbers of rows, but I also know that
avoiding timeouts can be a bit of an art when you're working with large sets.
Generally speaking, in an interactive environment, HBase is really designed for
small scans & gets among a large data set, rather than scanning an entire data
set (which is more the province of Map/Reduce).
Lacking a specific stack trace, this issue is too general to follow up on. I
propose we close this, and then if you run into an identifiable stack trace in
the future, post it to the list and / or open a new ticket. Sound OK?
> RetriesExhaustedException: Trying to contact region server
> ----------------------------------------------------------
>
> Key: HBASE-5322
> URL: https://issues.apache.org/jira/browse/HBASE-5322
> Project: HBase
> Issue Type: Bug
> Affects Versions: 0.90.4
> Reporter: Karthik Pandian
>
> I have a hbase table which holds data for more than 10GB. Now I used the same
> client scanner to scan which fails and reports,
> "Could not seek StoreFileScanner[HFileScanner for reader reader=hdfs".
> This issue occurs only for the table which holds huge data and not for tables
> holding small data.
--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators:
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira