[ https://issues.apache.org/jira/browse/LIVY-749?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
shanyu zhao updated LIVY-749: ----------------------------- Issue Type: Bug (was: New Feature) > Datanucleus jars are uploaded to hdfs unnecessarily when starting a livy > session > -------------------------------------------------------------------------------- > > Key: LIVY-749 > URL: https://issues.apache.org/jira/browse/LIVY-749 > Project: Livy > Issue Type: Bug > Components: Server > Affects Versions: 0.6.0, 0.7.0 > Reporter: shanyu zhao > Priority: Major > > If we start any Livy session with hive support > (livy.repl.enable-hive-context=true), we see that 3 datanucleus jars are > uploaded to HDFS and downloaded to drivers/executors: > Uploading resource file:/opt/spark/jars/datanucleus-api-jdo-3.2.6.jar -> > hdfs://namenode/user/test1/.sparkStaging/application_1581024490249_0002/datanucleus-api-jdo-3.2.6.jar > ... > These 3 datanucleus jars are not needed bacause they are already included in > Spark 2.x jars folder. > The reason is because in InteractiveSession.scala, method > mergeHiveSiteAndHiveDeps(), we merged datanucleus jars to spark.jars list > with method datanucleusJars(). We should remove datanucleusJars() function. -- This message was sent by Atlassian Jira (v8.3.4#803005)