GitHub user sameeragarwal opened a pull request:

    https://github.com/apache/spark/pull/12710

    [SPARK-14929][SQL] Disable vectorized map for wide schemas & high-precision 
decimals

    ## What changes were proposed in this pull request?
    
    While the vectorized hash map in `TungstenAggregate` is currently supported 
for all primitive data types during partial aggregation, this patch only 
enables the hash map for a subset of cases that've been verified to show 
performance improvements on our benchmarks subject to an internal conf that 
sets an upper limit on the maximum length of the aggregate key/value schema. 
This list of supported use-cases should be expanded over time.
    
    ## How was this patch tested?
    
    This is no new change in functionality so existing tests should suffice. 
Performance tests were done on TPCDS benchmarks.

You can merge this pull request into a Git repository by running:

    $ git pull https://github.com/sameeragarwal/spark vectorized-enable

Alternatively you can review and apply these changes as the patch at:

    https://github.com/apache/spark/pull/12710.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

    This closes #12710
    
----
commit f48eba15ab1804c7848dd14f0ee3bb051500934f
Author: Sameer Agarwal <[email protected]>
Date:   2016-04-26T20:02:31Z

    Not enable vectorized hashmap for wide schemas and high-precision decimals

----


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to