Christophe S created PHOENIX-1655: ------------------------------------- Summary: SpoolTooBigToDiskException when getting a huge resultset Key: PHOENIX-1655 URL: https://issues.apache.org/jira/browse/PHOENIX-1655 Project: Phoenix Issue Type: Bug Affects Versions: 4.2 Environment: HDP2.2 Reporter: Christophe S
I am working with timeseries and I would like to extract all the values for a sensor (12M row for one year). I thought it would work with Phoenix as it would stream the result but it seems it can't. My table is as follows: |M| VARCHAR |13| |TS| TIMESTAMP |-| |V| FLOAT |-| |Q| CHAR |10| |N| VARCHAR |10| The code giving the exception: {code} final ResultSet rs = statement.executeQuery( "select m, ts, v from MY_TABLE where m = 'mySensor'"); while (rs.next()) { final String line = rs.getString(1) + "," + rs.getDate(2).getTime() + "," + rs.getFloat(3); ps.println(line); } {code} After a while I got the following stacktrace: {code} Exception in thread "main" org.apache.phoenix.exception.PhoenixIOException: org.apache.phoenix.iterate.SpoolTooBigToDiskException: result too big, max allowed(bytes): 1044971520 at org.apache.phoenix.util.ServerUtil.parseServerException(ServerUtil.java:101) at org.apache.phoenix.iterate.ParallelIterators.getIterators(ParallelIterators.java:289) at org.apache.phoenix.iterate.ConcatResultIterator.getIterators(ConcatResultIterator.java:44) at org.apache.phoenix.iterate.ConcatResultIterator.currentIterator(ConcatResultIterator.java:66) at org.apache.phoenix.iterate.ConcatResultIterator.next(ConcatResultIterator.java:86) at org.apache.phoenix.jdbc.PhoenixResultSet.next(PhoenixResultSet.java:732) {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)