Github user jmahonin commented on the pull request:

    https://github.com/apache/phoenix/pull/59#issuecomment-88176830
  
    Thanks for the feedback @mravi , point comments below:
    
    1: Right, I'll try get that sorted out. The original phoenix-spark library 
would not work with 1.7 for some reason, but that may no longer be the case.
    
    2: Good catch. I think IntelliJ did something a little funny here on me, 
that file was supposed to be in the main hierarchy.
    
    3 / 4: It's my first kick at extending Spark (and Phoenix for that matter), 
but the naming scheme and file separation was modelled off of DataStax' 
Spark-Cassandra connector, which I figured is as good a model as any:
    
https://github.com/datastax/spark-cassandra-connector/tree/master/spark-cassandra-connector/src/main/scala/com/datastax/spark/connector
    
    In theory, doing it that way means a user can have just one import to get 
all the nice implicit definitions:
    `import org.apache.spark.phoenix._`
    
    5: I've never had much luck with getting the Scala integration working well 
on any IDE, I just run 'mvn test' from the CLI.
    
    Re: Good to haves
    1. I totally agree, but I don't think I can afford the cycles at the 
moment. My hope was that by modelling after the spark-cassandra-connector, it 
would be relatively painless to add for either a third party, or myself in the 
hopefully not-too-distant-future.
    2. Great idea, I hadn't actually seen that usage with Spark SQL yet. We're 
still using the RDD API internally. On a quick glance it looks fairly 
straight-forward to implement.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---

Reply via email to