[ 
https://issues.apache.org/jira/browse/SPARK-44600?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yang Jie updated SPARK-44600:
-----------------------------
    Description: 
[https://github.com/apache/spark/actions/runs/5727123477/job/15518895421]

 
{code:java}
- SPARK-15236: use Hive catalog *** FAILED ***
18137  isContain was true Interpreter output contained 'Exception':
18138  Welcome to
18139        ____              __
18140       / __/__  ___ _____/ /__
18141      _\ \/ _ \/ _ `/ __/  '_/
18142     /___/ .__/\_,_/_/ /_/\_\   version 4.0.0-SNAPSHOT
18143        /_/
18144           
18145  Using Scala version 2.12.18 (OpenJDK 64-Bit Server VM, Java 1.8.0_372)
18146  Type in expressions to have them evaluated.
18147  Type :help for more information.
18148  
18149  scala> 
18150  scala> java.lang.NoClassDefFoundError: 
org/sparkproject/guava/cache/CacheBuilder
18151    at 
org.apache.spark.sql.catalyst.catalog.SessionCatalog.<init>(SessionCatalog.scala:197)
18152    at 
org.apache.spark.sql.internal.BaseSessionStateBuilder.catalog$lzycompute(BaseSessionStateBuilder.scala:153)
18153    at 
org.apache.spark.sql.internal.BaseSessionStateBuilder.catalog(BaseSessionStateBuilder.scala:152)
18154    at 
org.apache.spark.sql.internal.BaseSessionStateBuilder.v2SessionCatalog$lzycompute(BaseSessionStateBuilder.scala:166)
18155    at 
org.apache.spark.sql.internal.BaseSessionStateBuilder.v2SessionCatalog(BaseSessionStateBuilder.scala:166)
18156    at 
org.apache.spark.sql.internal.BaseSessionStateBuilder.catalogManager$lzycompute(BaseSessionStateBuilder.scala:168)
18157    at 
org.apache.spark.sql.internal.BaseSessionStateBuilder.catalogManager(BaseSessionStateBuilder.scala:168)
18158    at 
org.apache.spark.sql.internal.BaseSessionStateBuilder$$anon$1.<init>(BaseSessionStateBuilder.scala:185)
18159    at 
org.apache.spark.sql.internal.BaseSessionStateBuilder.analyzer(BaseSessionStateBuilder.scala:185)
18160    at 
org.apache.spark.sql.internal.BaseSessionStateBuilder.$anonfun$build$2(BaseSessionStateBuilder.scala:374)
18161    at 
org.apache.spark.sql.internal.SessionState.analyzer$lzycompute(SessionState.scala:92)
18162    at 
org.apache.spark.sql.internal.SessionState.analyzer(SessionState.scala:92)
18163    at 
org.apache.spark.sql.execution.QueryExecution.$anonfun$analyzed$1(QueryExecution.scala:77)
18164    at 
org.apache.spark.sql.catalyst.QueryPlanningTracker.measurePhase(QueryPlanningTracker.scala:138)
18165    at 
org.apache.spark.sql.execution.QueryExecution.$anonfun$executePhase$2(QueryExecution.scala:219)
18166    at 
org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546)
18167    at 
org.apache.spark.sql.execution.QueryExecution.$anonfun$executePhase$1(QueryExecution.scala:219)
18168    at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
18169    at 
org.apache.spark.sql.execution.QueryExecution.executePhase(QueryExecution.scala:218)
18170    at 
org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:77)
18171    at 
org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:74)
18172    at 
org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:66)
18173    at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100)
18174    at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
18175    at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:98)
18176    at 
org.apache.spark.sql.SparkSession.$anonfun$sql$4(SparkSession.scala:691)
18177    at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
18178    at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:682)
18179    at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:713)
18180    at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:744)
18181    ... 100 elided
18182  Caused by: java.lang.ClassNotFoundException: 
org.sparkproject.guava.cache.CacheBuilder
18183    at java.net.URLClassLoader.findClass(URLClassLoader.java:387)
18184    at java.lang.ClassLoader.loadClass(ClassLoader.java:418)
18185    at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352)
18186    at java.lang.ClassLoader.loadClass(ClassLoader.java:351)
18187    ... 130 more
18188  
18189  scala>      | 
18190  scala> :quit (ReplSuite.scala:83) {code}

> Make `repl` module daily test pass
> ----------------------------------
>
>                 Key: SPARK-44600
>                 URL: https://issues.apache.org/jira/browse/SPARK-44600
>             Project: Spark
>          Issue Type: Sub-task
>          Components: Tests
>    Affects Versions: 4.0.0
>            Reporter: Yang Jie
>            Priority: Major
>
> [https://github.com/apache/spark/actions/runs/5727123477/job/15518895421]
>  
> {code:java}
> - SPARK-15236: use Hive catalog *** FAILED ***
> 18137  isContain was true Interpreter output contained 'Exception':
> 18138  Welcome to
> 18139        ____              __
> 18140       / __/__  ___ _____/ /__
> 18141      _\ \/ _ \/ _ `/ __/  '_/
> 18142     /___/ .__/\_,_/_/ /_/\_\   version 4.0.0-SNAPSHOT
> 18143        /_/
> 18144           
> 18145  Using Scala version 2.12.18 (OpenJDK 64-Bit Server VM, Java 1.8.0_372)
> 18146  Type in expressions to have them evaluated.
> 18147  Type :help for more information.
> 18148  
> 18149  scala> 
> 18150  scala> java.lang.NoClassDefFoundError: 
> org/sparkproject/guava/cache/CacheBuilder
> 18151    at 
> org.apache.spark.sql.catalyst.catalog.SessionCatalog.<init>(SessionCatalog.scala:197)
> 18152    at 
> org.apache.spark.sql.internal.BaseSessionStateBuilder.catalog$lzycompute(BaseSessionStateBuilder.scala:153)
> 18153    at 
> org.apache.spark.sql.internal.BaseSessionStateBuilder.catalog(BaseSessionStateBuilder.scala:152)
> 18154    at 
> org.apache.spark.sql.internal.BaseSessionStateBuilder.v2SessionCatalog$lzycompute(BaseSessionStateBuilder.scala:166)
> 18155    at 
> org.apache.spark.sql.internal.BaseSessionStateBuilder.v2SessionCatalog(BaseSessionStateBuilder.scala:166)
> 18156    at 
> org.apache.spark.sql.internal.BaseSessionStateBuilder.catalogManager$lzycompute(BaseSessionStateBuilder.scala:168)
> 18157    at 
> org.apache.spark.sql.internal.BaseSessionStateBuilder.catalogManager(BaseSessionStateBuilder.scala:168)
> 18158    at 
> org.apache.spark.sql.internal.BaseSessionStateBuilder$$anon$1.<init>(BaseSessionStateBuilder.scala:185)
> 18159    at 
> org.apache.spark.sql.internal.BaseSessionStateBuilder.analyzer(BaseSessionStateBuilder.scala:185)
> 18160    at 
> org.apache.spark.sql.internal.BaseSessionStateBuilder.$anonfun$build$2(BaseSessionStateBuilder.scala:374)
> 18161    at 
> org.apache.spark.sql.internal.SessionState.analyzer$lzycompute(SessionState.scala:92)
> 18162    at 
> org.apache.spark.sql.internal.SessionState.analyzer(SessionState.scala:92)
> 18163    at 
> org.apache.spark.sql.execution.QueryExecution.$anonfun$analyzed$1(QueryExecution.scala:77)
> 18164    at 
> org.apache.spark.sql.catalyst.QueryPlanningTracker.measurePhase(QueryPlanningTracker.scala:138)
> 18165    at 
> org.apache.spark.sql.execution.QueryExecution.$anonfun$executePhase$2(QueryExecution.scala:219)
> 18166    at 
> org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546)
> 18167    at 
> org.apache.spark.sql.execution.QueryExecution.$anonfun$executePhase$1(QueryExecution.scala:219)
> 18168    at 
> org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
> 18169    at 
> org.apache.spark.sql.execution.QueryExecution.executePhase(QueryExecution.scala:218)
> 18170    at 
> org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:77)
> 18171    at 
> org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:74)
> 18172    at 
> org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:66)
> 18173    at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100)
> 18174    at 
> org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
> 18175    at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:98)
> 18176    at 
> org.apache.spark.sql.SparkSession.$anonfun$sql$4(SparkSession.scala:691)
> 18177    at 
> org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
> 18178    at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:682)
> 18179    at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:713)
> 18180    at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:744)
> 18181    ... 100 elided
> 18182  Caused by: java.lang.ClassNotFoundException: 
> org.sparkproject.guava.cache.CacheBuilder
> 18183    at java.net.URLClassLoader.findClass(URLClassLoader.java:387)
> 18184    at java.lang.ClassLoader.loadClass(ClassLoader.java:418)
> 18185    at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352)
> 18186    at java.lang.ClassLoader.loadClass(ClassLoader.java:351)
> 18187    ... 130 more
> 18188  
> 18189  scala>      | 
> 18190  scala> :quit (ReplSuite.scala:83) {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to