Github user nkronenfeld commented on the pull request:
https://github.com/apache/spark/pull/5565#issuecomment-94554558
@srowen I'm sure, for what it's worth, that part works out - it's whether
the surrounding application structure can fit in that paradigm that is the
issue.
Just to be clear, I didn't really expect this to ever get merged in as is
:-) I am hoping for a few things from this PR:
1. I wanted a list of ununified points the various APIs needed to be unified
2. I wanted to know what sort of compatibility was needed (code-compatible,
which with one exception, this is, vs. binary compatible, which it definitely
isn't)
3. I was hoping to foster in the community some sense that this was a goal
we could work towards
* So that when we got to the next point where we could make compatibility
changes, we would be ready to do so
* And to try and prevent further changes from moving the APIs even
farther appart.
4. And, lastly, I was hoping that someone whos scala was better than mine
might have some ideas on how to do this without as many API changes (for
instance, modifying this so that we could have all 4 versions of reduceByKey,
with the bad one deprecated, and still have the compiler intuit types correctly)
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]