You can try increasing phoenix.query.timeoutMs (and hbase.client.scanner.timeout.period) on the client . https://phoenix.apache.org/tuning.html
On Fri, May 13, 2016 at 1:51 PM, 景涛 <[email protected]> wrote: > When I query from a very big table > It get errors as follow: > > java.lang.RuntimeException: > org.apache.phoenix.exception.PhoenixIOException: > org.apache.phoenix.exception.PhoenixIOException: Failed after attempts=36, > exceptions: > > Fri May 13 16:15:29 CST 2016, null, java.net.SocketTimeoutException: > callTimeout=60000, callDuration=60307: row '' on table 'IC_WHOLESALE_PRICE' > at > region=IC_WHOLESALE_PRICE,,1463121082822.e7e7cbd63c1831df75aee8842df5c7f6., > hostname=hadoop09,60020,1463119404385, seqNum=2 > > at sqlline.IncrementalRows.hasNext(IncrementalRows.java:73) > at sqlline.TableOutputFormat.print(TableOutputFormat.java:33) > at sqlline.SqlLine.print(SqlLine.java:1653) > at sqlline.Commands.execute(Commands.java:833) > at sqlline.Commands.sql(Commands.java:732) > at sqlline.SqlLine.dispatch(SqlLine.java:808) > at sqlline.SqlLine.begin(SqlLine.java:681) > at sqlline.SqlLine.start(SqlLine.java:398) > at sqlline.SqlLine.main(SqlLine.java:292) > > > Is this a client-side timeout, or do I need to change something > HBase-related on the server and restart the cluster? On master, or all > region servers? >
