Sumit Chauhan created LIVY-600: ---------------------------------- Summary: Default UDF not found when spark job is deployed through livy Key: LIVY-600 URL: https://issues.apache.org/jira/browse/LIVY-600 Project: Livy Issue Type: Bug Components: API, Core Affects Versions: 0.6.0 Environment: Centos 7 , Spark 2.3.1 Reporter: Sumit Chauhan
First time invocation after deployment through livy works fine. However on subsequent invocation it is not able to find the default udf written inside spark sql implicit val client = new LivyClientBuilder() .setURI(new URI("http://ABC:8998")) .setConf("spark.sql.crossJoin.enabled", "true") etConf("spark.cassandra.connection.host",ConfigUtils.getCassConnectionHost) .setConf("spark.cassandra.connection.port", "9042") .setConf("spark.cassandra.auth.username", authDetails._1) .setConf("spark.cassandra.auth.password", authDetails._2) .setConf("spark.cassandra.output.consistency.level" ,"LOCAL_ONE") .setConf("spark.cassandra.output.batch.size.rows", "auto") .setConf("spark.cassandra.output.concurrent.writes", "500") .setConf("spark.cassandra.output.batch.size.bytes", "100000") .setConf("spark.cassandra.output.throughput_mb_per_sec", "1") .setConf("spark.executor.memory" ,"4G") .setConf("spark.sql.crossJoin.enabled", "true") .setConf("spark.app.name","livy_poc") .build(); client.addJar(new URI("""hdfs://ABC:8020/user/livy/myJar.jar""")).get() Stack Trace or a permanent function registered in the database 'default'.; line 9 pos 33 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.analysis.Analyzer$LookupFunctions$$anonfun$apply$15$$anonfun$applyOrElse$50.apply(Analyzer.scala:1200) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.analysis.Analyzer$LookupFunctions$$anonfun$apply$15$$anonfun$applyOrElse$50.apply(Analyzer.scala:1200) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.analysis.package$.withPosition(package.scala:53) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.analysis.Analyzer$LookupFunctions$$anonfun$apply$15.applyOrElse(Analyzer.scala:1199) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.analysis.Analyzer$LookupFunctions$$anonfun$apply$15.applyOrElse(Analyzer.scala:1197) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$2.apply(TreeNode.scala:267) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$2.apply(TreeNode.scala:267) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:266) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$transformDown$1.apply(TreeNode.scala:272) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$transformDown$1.apply(TreeNode.scala:272) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:306) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:304) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:272) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$transformDown$1.apply(TreeNode.scala:272) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$transformDown$1.apply(TreeNode.scala:272) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:306) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:304) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:272) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.plans.QueryPlan$$anonfun$transformExpressionsDown$1.apply(QueryPlan.scala:85) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.plans.QueryPlan$$anonfun$transformExpressionsDown$1.apply(QueryPlan.scala:85) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.plans.QueryPlan$$anonfun$1.apply(QueryPlan.scala:107) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.plans.QueryPlan$$anonfun$1.apply(QueryPlan.scala:107) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.plans.QueryPlan.transformExpression$1(QueryPlan.scala:106) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.plans.QueryPlan.org$apache$spark$sql$catalyst$plans$QueryPlan$$recursiveTransform$1(QueryPlan.scala:118) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.plans.QueryPlan$$anonfun$org$apache$spark$sql$catalyst$plans$QueryPlan$$recursiveTransform$1$1.apply(QueryPlan.scala:122) 19/06/12 15:26:26 INFO LineBufferedStream: at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234) 19/06/12 15:26:26 INFO LineBufferedStream: at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234) 19/06/12 15:26:26 INFO LineBufferedStream: at scala.collection.immutable.List.foreach(List.scala:381) 19/06/12 15:26:26 INFO LineBufferedStream: at scala.collection.TraversableLike$class.map(TraversableLike.scala:234) 19/06/12 15:26:26 INFO LineBufferedStream: at scala.collection.immutable.List.map(List.scala:285) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.plans.QueryPlan.org$apache$spark$sql$catalyst$plans$QueryPlan$$recursiveTransform$1(QueryPlan.scala:122) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.plans.QueryPlan$$anonfun$2.apply(QueryPlan.scala:127) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.plans.QueryPlan.mapExpressions(QueryPlan.scala:127) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.plans.QueryPlan.transformExpressionsDown(QueryPlan.scala:85) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.plans.QueryPlan.transformExpressions(QueryPlan.scala:76) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.plans.QueryPlan$$anonfun$transformAllExpressions$1.applyOrElse(QueryPlan.scala:138) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.plans.QueryPlan$$anonfun$transformAllExpressions$1.applyOrElse(QueryPlan.scala:137) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$2.apply(TreeNode.scala:267) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$2.apply(TreeNode.scala:267) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:266) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.trees.TreeNode.transform(TreeNode.scala:256) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.plans.QueryPlan.transformAllExpressions(QueryPlan.scala:137) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.analysis.Analyzer$LookupFunctions$.apply(Analyzer.scala:1197) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.analysis.Analyzer$LookupFunctions$.apply(Analyzer.scala:1196) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:87) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:84) 19/06/12 15:26:26 INFO LineBufferedStream: at scala.collection.IndexedSeqOptimized$class.foldl(IndexedSeqOptimized.scala:57) 19/06/12 15:26:26 INFO LineBufferedStream: at scala.collection.IndexedSeqOptimized$class.foldLeft(IndexedSeqOptimized.scala:66) 19/06/12 15:26:26 INFO LineBufferedStream: at scala.collection.mutable.WrappedArray.foldLeft(WrappedArray.scala:35) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:84) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:76) 19/06/12 15:26:26 INFO LineBufferedStream: at scala.collection.immutable.List.foreach(List.scala:381) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:76) 19/06/12 15:26:26 INFO LineBufferedStream: at org.apache.spark.sql.catalyst.analysis.Analyzer.org$apache$spark$sql$catalyst$analysis$Analyzer$$executeSameContext(Analyzer.scala:124) -- This message was sent by Atlassian JIRA (v7.6.3#76005)