[ 
https://issues.apache.org/jira/browse/PHOENIX-2040?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14586539#comment-14586539
 ] 

Nick Dimiduk commented on PHOENIX-2040:
---------------------------------------

I think this looks alright to me, so long as we're in agreement we want to be 
able to launch spark jobs from the client jar.

[~jmahonin] have you had a chance to test that sqlline.py and psql.py both work 
after this patch is applied? As we've seen, these uberjar assemblies are pretty 
fragile. I've done my testing against a local mode HBase instance, basically 
it's enough to make sure everything is running in different JVMs. If that 
testing passes, +1 from me as well.

> Mark spark/scala dependencies as 'provided'
> -------------------------------------------
>
>                 Key: PHOENIX-2040
>                 URL: https://issues.apache.org/jira/browse/PHOENIX-2040
>             Project: Phoenix
>          Issue Type: Bug
>            Reporter: Josh Mahonin
>            Assignee: Josh Mahonin
>             Fix For: 5.0.0, 4.5.0
>
>         Attachments: PHOENIX-2040.patch
>
>
> The Spark runtime provides both the scala library, as well as the Spark 
> dependencies, so these should be marked as 'provided' in the phoenix-spark 
> module. This greatly reduces the size of the resulting client JAR.
> This patch also adds back phoenix-spark to the list of modules in the 
> assembly JAR, to be included in the client JAR.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to