Github user zecevicp commented on the issue:

    https://github.com/apache/spark/pull/21109
  
    Sorry, I don't quite understand your question. This already applies (only) 
to equijoins if there are additional range conditions on secondary columns. So 
if Spark rewrites those range conditions (?), and you end up with two equijoins 
(doesn't sound like a realistic scenario), then this doesn't apply at all.
    But if you have an equi-join which cannot be performed because you need to 
match a huge number of rows, and you can narrow down the search window using 
range conditions, then the advantage is that this makes it feasible and/or 
much, much faster.
    
    Regarding the second point, I don't want to tell you what to do, but on my 
part I can say that this has been tested with unit tests and on real, large 
datasets and I believe it should be safe to merge. But it can also wait for 
2.5/3.0...



---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to