Github user davies commented on the pull request:

    https://github.com/apache/spark/pull/13107#issuecomment-221993686
  
    @rxin After fixing those two, we still have some other limits (the number 
of elements should be less than 512 mm), especially for on-heap mode. There are:
    1) the largest memory block is 8G in on-heap mode, so the number of 
elements should be less than 512M.
    2) for both Radix sort and time sort, the underlying array could not 2G (or 
overflow), the number of records should be less than 1G
    3) For sorted iterator, the underlying array could not be larger than 2G.
    
    So we still need to check the number of elements, then do spilling, or the 
job could fail in unexpected way.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to