Github user jmahonin commented on the pull request: https://github.com/apache/phoenix/pull/59#issuecomment-88176830 Thanks for the feedback @mravi , point comments below: 1: Right, I'll try get that sorted out. The original phoenix-spark library would not work with 1.7 for some reason, but that may no longer be the case. 2: Good catch. I think IntelliJ did something a little funny here on me, that file was supposed to be in the main hierarchy. 3 / 4: It's my first kick at extending Spark (and Phoenix for that matter), but the naming scheme and file separation was modelled off of DataStax' Spark-Cassandra connector, which I figured is as good a model as any: https://github.com/datastax/spark-cassandra-connector/tree/master/spark-cassandra-connector/src/main/scala/com/datastax/spark/connector In theory, doing it that way means a user can have just one import to get all the nice implicit definitions: `import org.apache.spark.phoenix._` 5: I've never had much luck with getting the Scala integration working well on any IDE, I just run 'mvn test' from the CLI. Re: Good to haves 1. I totally agree, but I don't think I can afford the cycles at the moment. My hope was that by modelling after the spark-cassandra-connector, it would be relatively painless to add for either a third party, or myself in the hopefully not-too-distant-future. 2. Great idea, I hadn't actually seen that usage with Spark SQL yet. We're still using the RDD API internally. On a quick glance it looks fairly straight-forward to implement.
--- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---