The following tests are failing: /Functional/data-shapes/wide-columns/5000/1000rows/parquet/q163_DRILL-2046.q > > /Functional/data-shapes/wide-columns/5000/1000rows/parquet/q177_DRILL-2046.q > /Functional/data-shapes/wide-columns/5000/1000rows/parquet/q174.q > /Functional/data-shapes/wide-columns/5000/1000rows/parquet/ > /Functional/window_functions/multiple_partitions/q35.sql > > /Functional/data-shapes/wide-columns/5000/1000rows/parquet/q160_DRILL-1985.q > > /Functional/data-shapes/wide-columns/5000/1000rows/parquet/q162_DRILL-1985.q > /Functional/data-shapes/wide-columns/5000/1000rows/parquet/q165.q > /Functional/window_functions/multiple_partitions/q37.sql > /Functional/data-shapes/wide-columns/5000/1000rows/parquet/q171.q > > /Functional/data-shapes/wide-columns/5000/1000rows/parquet/q168_DRILL-2046.q > /Functional/window_functions/multiple_partitions/q36.sql > > /Functional/data-shapes/wide-columns/5000/1000rows/parquet/q159_DRILL-2046.q > /Functional/window_functions/multiple_partitions/q30.sql > > /Functional/data-shapes/wide-columns/5000/1000rows/parquet/large/q157_DRILL-1985.q > /Functional/window_functions/multiple_partitions/q22.sql
With one of the following errors: java.sql.SQLException: RESOURCE ERROR: One or more nodes ran out of memory > while executing the query. > Caused by: org.apache.drill.exec.exception.OutOfMemoryException: > org.apache.drill.exec.exception.OutOfMemoryException: Unable to allocate > sv2, and not enough batchGroups to spill > at > org.apache.drill.exec.physical.impl.xsort.ExternalSortBatch.innerNext(ExternalSortBatch.java:356) > ~[drill-java-exec-1.5.0-SNAPSHOT.jar:1.5.0-SNAPSHOT] or java.sql.SQLException: SYSTEM ERROR: DrillRuntimeException: Failed to > pre-allocate memory for SV. Existing recordCount*4 = 0, incoming batch > recordCount*4 = 3340 > Caused by: org.apache.drill.common.exceptions.DrillRuntimeException: > Failed to pre-allocate memory for SV. Existing recordCount*4 = 0, incoming > batch recordCount*4 = 3340 > at > org.apache.drill.exec.physical.impl.sort.SortRecordBatchBuilder.add(SortRecordBatchBuilder.java:116) > ~[drill-java-exec-1.5.0-SNAPSHOT.jar:1.5.0-SNAPSHOT] > at > org.apache.drill.exec.physical.impl.xsort.ExternalSortBatch.innerNext(ExternalSortBatch.java:451) > ~[drill-java-exec-1.5.0-SNAPSHOT.jar:1.5.0-SNAPSHOT] On Wed, Dec 30, 2015 at 12:42 PM, Jacques Nadeau <[email protected]> wrote: > I'll let Steven answer your question directly. > > FYI, we are running a regression suite that was forked from the MapR repo a > month or so ago because we had to fix a bunch of things to make it work > with Apache Hadoop. (There was a thread about this back then and we haven't > yet figured out how to merge both suites.) It is possible that he had a > successful run but the failures are happening on items that you've recently > added to your suite. > > It is also possible (likely?) that the configuration settings for our > regression clusters are not the same. > > -- > Jacques Nadeau > CTO and Co-Founder, Dremio > > On Wed, Dec 30, 2015 at 12:37 PM, Abdel Hakim Deneche < > [email protected] > > wrote: > > > Steven, > > > > were you able to successfully run the regression tests on the transfer > > patch ? I just tried and saw several queries running out of memory ! > > > > On Wed, Dec 30, 2015 at 11:46 AM, Abdel Hakim Deneche < > > [email protected] > > > wrote: > > > > > Created DRILL-4236 <https://issues.apache.org/jira/browse/DRILL-4236> > to > > > keep track of this improvement. > > > > > > On Wed, Dec 30, 2015 at 11:01 AM, Jacques Nadeau <[email protected]> > > > wrote: > > > > > >> Since the accounting changed (more accurate), the termination > condition > > >> for > > >> the sort operator will be different than before. In fact, this likely > > will > > >> be sooner since our accounting is much larger than previously (since > we > > >> correctly consider the entire allocation rather than simply the used > > >> allocation). > > >> > > >> Hakim, > > >> Steven and I were discussing the need to update the ExternalSort > > operator > > >> to use the new allocator functionality to better manage its memory > > >> envelope. Would you be interested in working on this since you seem to > > be > > >> working with that code the most? Basically, it used to be that there > was > > >> no > > >> way the sort operator would be able to correctly detect a memory > > condition > > >> and so it jumped through a bunch of hoops to try to figure out the > > >> termination condition.With the transfer accounting in place, this code > > can > > >> be greatly simplified to just use the current operator memory > > allocation. > > >> > > >> -- > > >> Jacques Nadeau > > >> CTO and Co-Founder, Dremio > > >> > > >> On Wed, Dec 30, 2015 at 10:48 AM, rahul challapalli < > > >> [email protected]> wrote: > > >> > > >> > I installed the latest master and ran this query. So > > >> > planner.memory.max_query_memory_per_node should have been the > default > > >> > value. I switched back to 1.4.0 branch and this query completed > > >> > successfully. > > >> > > > >> > On Wed, Dec 30, 2015 at 10:37 AM, Abdel Hakim Deneche < > > >> > [email protected] > > >> > > wrote: > > >> > > > >> > > Rahul, > > >> > > > > >> > > How much memory was assigned to the sort operator ( > > >> > > planner.memory.max_query_memory_per_node) ? > > >> > > > > >> > > On Wed, Dec 30, 2015 at 9:54 AM, rahul challapalli < > > >> > > [email protected]> wrote: > > >> > > > > >> > > > I am seeing an OOM error while executing a simple CTAS query. I > > >> raised > > >> > > > DRILL-4324 for this. The query mentioned in the JIRA used to > > >> complete > > >> > > > successfully without any issue prior to 1.5. Any idea what could > > >> have > > >> > > > caused the regression? > > >> > > > > > >> > > > - Rahul > > >> > > > > > >> > > > > >> > > > > >> > > > > >> > > -- > > >> > > > > >> > > Abdelhakim Deneche > > >> > > > > >> > > Software Engineer > > >> > > > > >> > > <http://www.mapr.com/> > > >> > > > > >> > > > > >> > > Now Available - Free Hadoop On-Demand Training > > >> > > < > > >> > > > > >> > > > >> > > > http://www.mapr.com/training?utm_source=Email&utm_medium=Signature&utm_campaign=Free%20available > > >> > > > > > >> > > > > >> > > > >> > > > > > > > > > > > > -- > > > > > > Abdelhakim Deneche > > > > > > Software Engineer > > > > > > <http://www.mapr.com/> > > > > > > > > > Now Available - Free Hadoop On-Demand Training > > > < > > > http://www.mapr.com/training?utm_source=Email&utm_medium=Signature&utm_campaign=Free%20available > > > > > > > > > > > > > > -- > > > > Abdelhakim Deneche > > > > Software Engineer > > > > <http://www.mapr.com/> > > > > > > Now Available - Free Hadoop On-Demand Training > > < > > > http://www.mapr.com/training?utm_source=Email&utm_medium=Signature&utm_campaign=Free%20available > > > > > > -- Abdelhakim Deneche Software Engineer <http://www.mapr.com/> Now Available - Free Hadoop On-Demand Training <http://www.mapr.com/training?utm_source=Email&utm_medium=Signature&utm_campaign=Free%20available>
