[
https://issues.apache.org/jira/browse/LIVY-749?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Gyorgy Gal updated LIVY-749:
----------------------------
Fix Version/s: 0.10.0
(was: 0.9.0)
This issue has been moved to the 0.10.0 release as part of a bulk update. If
you feel this is moved out inappropriately, feel free to provide justification
and reset the Fix Version to 0.9.0.
> Datanucleus jars are uploaded to hdfs unnecessarily when starting a livy
> session
> --------------------------------------------------------------------------------
>
> Key: LIVY-749
> URL: https://issues.apache.org/jira/browse/LIVY-749
> Project: Livy
> Issue Type: Bug
> Components: Server
> Affects Versions: 0.6.0, 0.7.0
> Reporter: shanyu zhao
> Priority: Major
> Fix For: 0.10.0
>
> Time Spent: 20m
> Remaining Estimate: 0h
>
> If we start any Livy session with hive support
> (livy.repl.enable-hive-context=true), we see that 3 datanucleus jars are
> uploaded to HDFS and downloaded to drivers/executors:
> Uploading resource file:/opt/spark/jars/datanucleus-api-jdo-3.2.6.jar ->
> hdfs://namenode/user/test1/.sparkStaging/application_1581024490249_0002/datanucleus-api-jdo-3.2.6.jar
> ...
> These 3 datanucleus jars are not needed bacause they are already included in
> Spark 2.x jars folder.
> The reason is because in InteractiveSession.scala, method
> mergeHiveSiteAndHiveDeps(), we merged datanucleus jars to spark.jars list
> with method datanucleusJars(). We should remove datanucleusJars() function.
--
This message was sent by Atlassian Jira
(v8.20.10#820010)