Sumit Chauhan created LIVY-600:
----------------------------------

             Summary: Default UDF not found when spark job is deployed through 
livy
                 Key: LIVY-600
                 URL: https://issues.apache.org/jira/browse/LIVY-600
             Project: Livy
          Issue Type: Bug
          Components: API, Core
    Affects Versions: 0.6.0
         Environment: Centos 7 , Spark 2.3.1
            Reporter: Sumit Chauhan


First time invocation after deployment through livy works fine.

However on subsequent invocation it is not able to find the default udf written 
inside spark sql

 


implicit val client = new LivyClientBuilder()
 .setURI(new URI("http://ABC:8998";))
 .setConf("spark.sql.crossJoin.enabled", "true")
etConf("spark.cassandra.connection.host",ConfigUtils.getCassConnectionHost)
 .setConf("spark.cassandra.connection.port", "9042")
 .setConf("spark.cassandra.auth.username", authDetails._1)
 .setConf("spark.cassandra.auth.password", authDetails._2)
 .setConf("spark.cassandra.output.consistency.level" ,"LOCAL_ONE")
.setConf("spark.cassandra.output.batch.size.rows", "auto")
 .setConf("spark.cassandra.output.concurrent.writes", "500")
 .setConf("spark.cassandra.output.batch.size.bytes", "100000")
 .setConf("spark.cassandra.output.throughput_mb_per_sec", "1")
 .setConf("spark.executor.memory" ,"4G")
 .setConf("spark.sql.crossJoin.enabled", "true")
 .setConf("spark.app.name","livy_poc")
 .build();

 client.addJar(new URI("""hdfs://ABC:8020/user/livy/myJar.jar""")).get()

 

Stack Trace

or a permanent function registered in the database 'default'.; line 9 pos 33
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.analysis.Analyzer$LookupFunctions$$anonfun$apply$15$$anonfun$applyOrElse$50.apply(Analyzer.scala:1200)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.analysis.Analyzer$LookupFunctions$$anonfun$apply$15$$anonfun$applyOrElse$50.apply(Analyzer.scala:1200)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.analysis.package$.withPosition(package.scala:53)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.analysis.Analyzer$LookupFunctions$$anonfun$apply$15.applyOrElse(Analyzer.scala:1199)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.analysis.Analyzer$LookupFunctions$$anonfun$apply$15.applyOrElse(Analyzer.scala:1197)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$2.apply(TreeNode.scala:267)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$2.apply(TreeNode.scala:267)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:266)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$transformDown$1.apply(TreeNode.scala:272)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$transformDown$1.apply(TreeNode.scala:272)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:306)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:304)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:272)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$transformDown$1.apply(TreeNode.scala:272)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$transformDown$1.apply(TreeNode.scala:272)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:306)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:304)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:272)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.plans.QueryPlan$$anonfun$transformExpressionsDown$1.apply(QueryPlan.scala:85)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.plans.QueryPlan$$anonfun$transformExpressionsDown$1.apply(QueryPlan.scala:85)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.plans.QueryPlan$$anonfun$1.apply(QueryPlan.scala:107)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.plans.QueryPlan$$anonfun$1.apply(QueryPlan.scala:107)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.plans.QueryPlan.transformExpression$1(QueryPlan.scala:106)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.plans.QueryPlan.org$apache$spark$sql$catalyst$plans$QueryPlan$$recursiveTransform$1(QueryPlan.scala:118)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.plans.QueryPlan$$anonfun$org$apache$spark$sql$catalyst$plans$QueryPlan$$recursiveTransform$1$1.apply(QueryPlan.scala:122)
19/06/12 15:26:26 INFO LineBufferedStream: at 
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
19/06/12 15:26:26 INFO LineBufferedStream: at 
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
19/06/12 15:26:26 INFO LineBufferedStream: at 
scala.collection.immutable.List.foreach(List.scala:381)
19/06/12 15:26:26 INFO LineBufferedStream: at 
scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
19/06/12 15:26:26 INFO LineBufferedStream: at 
scala.collection.immutable.List.map(List.scala:285)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.plans.QueryPlan.org$apache$spark$sql$catalyst$plans$QueryPlan$$recursiveTransform$1(QueryPlan.scala:122)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.plans.QueryPlan$$anonfun$2.apply(QueryPlan.scala:127)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.plans.QueryPlan.mapExpressions(QueryPlan.scala:127)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.plans.QueryPlan.transformExpressionsDown(QueryPlan.scala:85)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.plans.QueryPlan.transformExpressions(QueryPlan.scala:76)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.plans.QueryPlan$$anonfun$transformAllExpressions$1.applyOrElse(QueryPlan.scala:138)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.plans.QueryPlan$$anonfun$transformAllExpressions$1.applyOrElse(QueryPlan.scala:137)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$2.apply(TreeNode.scala:267)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$2.apply(TreeNode.scala:267)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:266)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.trees.TreeNode.transform(TreeNode.scala:256)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.plans.QueryPlan.transformAllExpressions(QueryPlan.scala:137)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.analysis.Analyzer$LookupFunctions$.apply(Analyzer.scala:1197)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.analysis.Analyzer$LookupFunctions$.apply(Analyzer.scala:1196)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:87)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:84)
19/06/12 15:26:26 INFO LineBufferedStream: at 
scala.collection.IndexedSeqOptimized$class.foldl(IndexedSeqOptimized.scala:57)
19/06/12 15:26:26 INFO LineBufferedStream: at 
scala.collection.IndexedSeqOptimized$class.foldLeft(IndexedSeqOptimized.scala:66)
19/06/12 15:26:26 INFO LineBufferedStream: at 
scala.collection.mutable.WrappedArray.foldLeft(WrappedArray.scala:35)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:84)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:76)
19/06/12 15:26:26 INFO LineBufferedStream: at 
scala.collection.immutable.List.foreach(List.scala:381)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:76)
19/06/12 15:26:26 INFO LineBufferedStream: at 
org.apache.spark.sql.catalyst.analysis.Analyzer.org$apache$spark$sql$catalyst$analysis$Analyzer$$executeSameContext(Analyzer.scala:124)

 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to