Misha Dmitriev created HIVE-19668: ------------------------------------- Summary: 11.8% of the heap wasted due to duplicate org.antlr.runtime.CommonToken's Key: HIVE-19668 URL: https://issues.apache.org/jira/browse/HIVE-19668 Project: Hive Issue Type: Improvement Components: HiveServer2 Affects Versions: 3.0.0 Reporter: Misha Dmitriev Assignee: Misha Dmitriev Attachments: image-2018-05-22-17-41-39-572.png
I've recently analyzed a HS2 heap dump, obtained when there was a huge memory spike during compilation of some big query. The analysis was done with jxray ([www.jxray.com).|http://www.jxray.com)./] It turns out that more than 90% of the 20G heap was used by data structures associated with query parsing ({{org.apache.hadoop.hive.ql.parse.QBExpr}}). There are probably multiple opportunities for optimizations here. One of them is to stop the code from creating duplicate instances of {{org.antlr.runtime.CommonToken}} class. See a sample of these objects in the attached image: !image-2018-05-22-17-41-39-572.png|width=879,height=399! Looks like these particular {{CommonToken}} objects are constants, that don't change once created. I see some code, e.g. in {{org.apache.hadoop.hive.ql.parse.CalcitePlanner}}, where such objects are apparently repeatedly created with e.g. {{new CommonToken(HiveParser.TOK_INSERT, "TOK_INSERT")}} If these 33 token kinds are instead created once and reused, we will save more than 1/10th of the heap in this scenario. Plus, since these objects are small but very numerous, getting rid of them will remove a gread deal of pressure from the GC. -- This message was sent by Atlassian JIRA (v7.6.3#76005)