[ 
https://issues.apache.org/jira/browse/LIVY-749?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gyorgy Gal updated LIVY-749:
----------------------------
    Fix Version/s: 0.10.0
                       (was: 0.9.0)

This issue has been moved to the 0.10.0 release as part of a bulk update. If 
you feel this is moved out inappropriately, feel free to provide justification 
and reset the Fix Version to 0.9.0.

> Datanucleus jars are uploaded to hdfs unnecessarily when starting a livy 
> session
> --------------------------------------------------------------------------------
>
>                 Key: LIVY-749
>                 URL: https://issues.apache.org/jira/browse/LIVY-749
>             Project: Livy
>          Issue Type: Bug
>          Components: Server
>    Affects Versions: 0.6.0, 0.7.0
>            Reporter: shanyu zhao
>            Priority: Major
>             Fix For: 0.10.0
>
>          Time Spent: 20m
>  Remaining Estimate: 0h
>
> If we start any Livy session with hive support 
> (livy.repl.enable-hive-context=true), we see that 3 datanucleus jars are 
> uploaded to HDFS and downloaded to drivers/executors:
> Uploading resource file:/opt/spark/jars/datanucleus-api-jdo-3.2.6.jar -> 
> hdfs://namenode/user/test1/.sparkStaging/application_1581024490249_0002/datanucleus-api-jdo-3.2.6.jar
> ...
> These 3 datanucleus jars are not needed bacause they are already included in 
> Spark 2.x jars folder.
> The reason is because in InteractiveSession.scala, method 
> mergeHiveSiteAndHiveDeps(), we merged datanucleus jars to spark.jars list 
> with method datanucleusJars(). We should remove datanucleusJars() function.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to