GitHub user nburoojy opened a pull request:

    https://github.com/apache/spark/pull/9526

    [SPARK-9301] [SQL] Add collect_set and collect_list aggregate functions

    For now they are thin wrappers around the corresponding Hive UDAFs.
    
    One limitation with these in Hive 0.13.0 is they only support aggregating 
primitive types.
    
    I chose snake_case here instead of camelCase because it seems to be used in 
the majority of the multi-word fns.
    
    Do we also want to add these to `functions.py`?
    
    This approach was recommended here: 
https://github.com/apache/spark/pull/8592#issuecomment-154247089
    
    @marmbrus @rxin 

You can merge this pull request into a Git repository by running:

    $ git pull https://github.com/nburoojy/spark nick/udaf-alias

Alternatively you can review and apply these changes as the patch at:

    https://github.com/apache/spark/pull/9526.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

    This closes #9526
    
----
commit 289ace5b2577cb53f84b98407597f9dde98755d7
Author: Nick Buroojy <[email protected]>
Date:   2015-11-06T19:12:20Z

    [SPARK-9301] [SQL] Add collect_set and collect_list aggregate functions
    
    For now they are thin wrappers around the corresponding Hive UDAFs.

----


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to