[ 
https://issues.apache.org/jira/browse/SPARK-16713?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Qifan Pu updated SPARK-16713:
-----------------------------
    Description: Ideally, we would wish codegen methods to be less than 8KB for 
bytecode size. Beyond 8K JIT won't compile and can cause performance 
degradation. We have seen this for queries with wide schema (30+ fields), where 
agg_doAggregateWithKeys() can be more than 8K. This is also a major reason for 
performance regression when we enable fash aggregate hashmap (such as using 
VectorizedHashMapGenerator.scala).  (was: Ideally, we would wish codegen 
methods to be less than 8KB for bytecode size. Beyond 8K JIT won't compile and 
can cause performance degradation. We have seen this for queries with wide 
schema (30+ fields). This is also a major reason for performance regression 
when we enable fash aggregate hashmap (such as using 
VectorizedHashMapGenerator.scala).)

> Limit codegen method size to 8KB
> --------------------------------
>
>                 Key: SPARK-16713
>                 URL: https://issues.apache.org/jira/browse/SPARK-16713
>             Project: Spark
>          Issue Type: Improvement
>          Components: Spark Core, SQL
>    Affects Versions: 2.0.0
>            Reporter: Qifan Pu
>
> Ideally, we would wish codegen methods to be less than 8KB for bytecode size. 
> Beyond 8K JIT won't compile and can cause performance degradation. We have 
> seen this for queries with wide schema (30+ fields), where 
> agg_doAggregateWithKeys() can be more than 8K. This is also a major reason 
> for performance regression when we enable fash aggregate hashmap (such as 
> using VectorizedHashMapGenerator.scala).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to