[ 
https://issues.apache.org/jira/browse/SPARK-43592?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17724291#comment-17724291
 ] 

Yuming Wang commented on SPARK-43592:
-------------------------------------

Maybe same issue: SPARK-32475?

> NoSuchMethodError in Spark 3.4 with JDK8u362 & JDK8u372
> -------------------------------------------------------
>
>                 Key: SPARK-43592
>                 URL: https://issues.apache.org/jira/browse/SPARK-43592
>             Project: Spark
>          Issue Type: Bug
>          Components: Spark Core
>    Affects Versions: 3.4.0
>         Environment: JDK: JDK8u362, JDK8u372
> Kubernetes
> Spark 3.4
>            Reporter: Shivam Kasat
>            Priority: Critical
>              Labels: JDK1.8, java, jdk11
>
> My project was on spark 3.3 with JDK8u362 and I tried updating it to spark 
> 3.4, Official documentation of spark 3.4 says it works with JDK8u362 and 
> above but when I tried upgrading docker base image of spark to JDK8u362 and 
> JDK8u372 it is failing at runtime with below error, For JDK8u362 it throws 
> error for Java.nio.CharBuffer.position method and for JDK8u372 it throws 
> error for java.nio.ByteBuffer.flip method. But when I run with JDK11 image in 
> spark Docker file it works fine. Am I missing anything or how to fix this 
> issue as I want to run it with JDK8.
> {code:java}
> ava.lang.NoSuchMethodError: 
> java.nio.CharBuffer.position(I)Ljava/nio/CharBuffer;
>         at 
> org.apache.spark.sql.catalyst.parser.ParserUtils$.unescapeSQLString(ParserUtils.scala:220)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.ParserUtils$.string(ParserUtils.scala:95)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.$anonfun$createString$2(AstBuilder.scala:2632)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> scala.collection.TraversableLike.$anonfun$map$1(TraversableLike.scala:286) 
> ~[scala-library-2.12.17.jar:?]
>         at scala.collection.Iterator.foreach(Iterator.scala:943) 
> ~[scala-library-2.12.17.jar:?]
>         at scala.collection.Iterator.foreach$(Iterator.scala:943) 
> ~[scala-library-2.12.17.jar:?]
>         at scala.collection.AbstractIterator.foreach(Iterator.scala:1431) 
> ~[scala-library-2.12.17.jar:?]
>         at scala.collection.IterableLike.foreach(IterableLike.scala:74) 
> ~[scala-library-2.12.17.jar:?]
>         at scala.collection.IterableLike.foreach$(IterableLike.scala:73) 
> ~[scala-library-2.12.17.jar:?]
>         at scala.collection.AbstractIterable.foreach(Iterable.scala:56) 
> ~[scala-library-2.12.17.jar:?]
>         at scala.collection.TraversableLike.map(TraversableLike.scala:286) 
> ~[scala-library-2.12.17.jar:?]
>         at scala.collection.TraversableLike.map$(TraversableLike.scala:279) 
> ~[scala-library-2.12.17.jar:?]
>         at scala.collection.AbstractTraversable.map(Traversable.scala:108) 
> ~[scala-library-2.12.17.jar:?]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.createString(AstBuilder.scala:2632)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.$anonfun$visitStringLiteral$1(AstBuilder.scala:2618)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.ParserUtils$.withOrigin(ParserUtils.scala:160)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitStringLiteral(AstBuilder.scala:2618)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitStringLiteral(AstBuilder.scala:58)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.SqlBaseParser$StringLiteralContext.accept(SqlBaseParser.java:19511)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitChildren(AstBuilder.scala:73)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.SqlBaseParserBaseVisitor.visitConstantDefault(SqlBaseParserBaseVisitor.java:1735)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.SqlBaseParser$ConstantDefaultContext.accept(SqlBaseParser.java:18373)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitChildren(AstBuilder.scala:73)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.SqlBaseParserBaseVisitor.visitValueExpressionDefault(SqlBaseParserBaseVisitor.java:1567)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.SqlBaseParser$ValueExpressionDefaultContext.accept(SqlBaseParser.java:17491)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.typedVisit(AstBuilder.scala:63)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.expression(AstBuilder.scala:1630)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.$anonfun$withPredicate$1(AstBuilder.scala:1870)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.ParserUtils$.withOrigin(ParserUtils.scala:160)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.withPredicate(AstBuilder.scala:1784)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.$anonfun$visitPredicated$1(AstBuilder.scala:1768)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.ParserUtils$.withOrigin(ParserUtils.scala:160)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitPredicated(AstBuilder.scala:1765)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitPredicated(AstBuilder.scala:58)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.SqlBaseParser$PredicatedContext.accept(SqlBaseParser.java:16909)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitChildren(AstBuilder.scala:73)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.SqlBaseParserBaseVisitor.visitExpression(SqlBaseParserBaseVisitor.java:1518)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.SqlBaseParser$ExpressionContext.accept(SqlBaseParser.java:16766)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.typedVisit(AstBuilder.scala:63)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.expression(AstBuilder.scala:1630)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.$anonfun$visitParenthesizedExpression$1(AstBuilder.scala:2361)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.ParserUtils$.withOrigin(ParserUtils.scala:160)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitParenthesizedExpression(AstBuilder.scala:2361)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitParenthesizedExpression(AstBuilder.scala:58)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.SqlBaseParser$ParenthesizedExpressionContext.accept(SqlBaseParser.java:18036)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitChildren(AstBuilder.scala:73)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.SqlBaseParserBaseVisitor.visitValueExpressionDefault(SqlBaseParserBaseVisitor.java:1567)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.SqlBaseParser$ValueExpressionDefaultContext.accept(SqlBaseParser.java:17491)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.typedVisit(AstBuilder.scala:63)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.expression(AstBuilder.scala:1630)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.$anonfun$visitPredicated$1(AstBuilder.scala:1766)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.ParserUtils$.withOrigin(ParserUtils.scala:160)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitPredicated(AstBuilder.scala:1765)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitPredicated(AstBuilder.scala:58)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.SqlBaseParser$PredicatedContext.accept(SqlBaseParser.java:16909)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.typedVisit(AstBuilder.scala:63)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.expression(AstBuilder.scala:1630)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.$anonfun$visitLogicalBinary$4(AstBuilder.scala:1694)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at scala.collection.SeqLike.$anonfun$reverseMap$2(SeqLike.scala:295) 
> ~[scala-library-2.12.17.jar:?]
>         at scala.collection.SeqLike.reverseMap(SeqLike.scala:294) 
> ~[scala-library-2.12.17.jar:?]
>         at scala.collection.SeqLike.reverseMap$(SeqLike.scala:289) 
> ~[scala-library-2.12.17.jar:?]
>         at scala.collection.AbstractSeq.reverseMap(Seq.scala:45) 
> ~[scala-library-2.12.17.jar:?]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.$anonfun$visitLogicalBinary$1(AstBuilder.scala:1694)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.ParserUtils$.withOrigin(ParserUtils.scala:160)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitLogicalBinary(AstBuilder.scala:1669)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitLogicalBinary(AstBuilder.scala:58)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.SqlBaseParser$LogicalBinaryContext.accept(SqlBaseParser.java:16958)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.typedVisit(AstBuilder.scala:63)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.expression(AstBuilder.scala:1630)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.withWhereClause(AstBuilder.scala:703)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.$anonfun$visitCommonSelectQueryClausePlan$2(AstBuilder.scala:807)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.ParserUtils$EnhancedLogicalPlan$.optionalMap$extension(ParserUtils.scala:256)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitCommonSelectQueryClausePlan(AstBuilder.scala:807)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.$anonfun$withSelectQuerySpecification$1(AstBuilder.scala:788)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.ParserUtils$.withOrigin(ParserUtils.scala:160)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.withSelectQuerySpecification(AstBuilder.scala:776)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.$anonfun$visitRegularQuerySpecification$1(AstBuilder.scala:668)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.ParserUtils$.withOrigin(ParserUtils.scala:160)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitRegularQuerySpecification(AstBuilder.scala:656)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitRegularQuerySpecification(AstBuilder.scala:58)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.SqlBaseParser$RegularQuerySpecificationContext.accept(SqlBaseParser.java:10386)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitChildren(AstBuilder.scala:73)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.SqlBaseParserBaseVisitor.visitQueryPrimaryDefault(SqlBaseParserBaseVisitor.java:902)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.SqlBaseParser$QueryPrimaryDefaultContext.accept(SqlBaseParser.java:9891)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitChildren(AstBuilder.scala:73)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.SqlBaseParserBaseVisitor.visitQueryTermDefault(SqlBaseParserBaseVisitor.java:888)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.SqlBaseParser$QueryTermDefaultContext.accept(SqlBaseParser.java:9658)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.typedVisit(AstBuilder.scala:63)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.plan(AstBuilder.scala:114) 
> ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.$anonfun$visitQuery$1(AstBuilder.scala:120)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.ParserUtils$.withOrigin(ParserUtils.scala:160)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitQuery(AstBuilder.scala:119)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitQuery(AstBuilder.scala:58)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.SqlBaseParser$QueryContext.accept(SqlBaseParser.java:6891)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitChildren(AstBuilder.scala:73)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.SqlBaseParserBaseVisitor.visitStatementDefault(SqlBaseParserBaseVisitor.java:69)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.SqlBaseParser$StatementDefaultContext.accept(SqlBaseParser.java:1988)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.antlr.v4.runtime.tree.AbstractParseTreeVisitor.visit(AbstractParseTreeVisitor.java:18)
>  ~[antlr4-runtime-4.9.3.jar:4.9.3]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.$anonfun$visitSingleStatement$1(AstBuilder.scala:80)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.ParserUtils$.withOrigin(ParserUtils.scala:160)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitSingleStatement(AstBuilder.scala:80)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AbstractSqlParser.$anonfun$parsePlan$2(ParseDriver.scala:92)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.ParserUtils$.withOrigin(ParserUtils.scala:160)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AbstractSqlParser.$anonfun$parsePlan$1(ParseDriver.scala:92)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AbstractSqlParser.parse(ParseDriver.scala:127)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.execution.SparkSqlParser.parse(SparkSqlParser.scala:52) 
> ~[spark-sql_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.parser.AbstractSqlParser.parsePlan(ParseDriver.scala:89)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.SparkSession.$anonfun$sql$2(SparkSession.scala:633) 
> ~[spark-sql_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.catalyst.QueryPlanningTracker.measurePhase(QueryPlanningTracker.scala:111)
>  ~[spark-catalyst_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:632) 
> ~[spark-sql_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:827) 
> ~[spark-sql_2.12-3.4.0.jar:3.4.0]
>         at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:630) 
> ~[spark-sql_2.12-3.4.0.jar:3.4.0]
>         at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:671) 
> ~[spark-sql_2.12-3.4.0.jar:3.4.0]
>         at 
> com.cisco.intersight.cep.AdvisoryProcessorTask.processActions(AdvisoryProcessorTask.java:376)
>  ~[orion.jar:?]
>         at 
> com.cisco.intersight.cep.AdvisoryProcessorTask.processAdvisory(AdvisoryProcessorTask.java:329)
>  ~[orion.jar:?]
>         at 
> com.cisco.intersight.cep.AdvisoryProcessorTask.processAdvisories(AdvisoryProcessorTask.java:306)
>  ~[orion.jar:?]
>         at 
> com.cisco.intersight.cep.AdvisoryProcessorTask.run(AdvisoryProcessorTask.java:286)
>  ~[orion.jar:?]
>         at 
> com.cisco.intersight.cep.AdvisoryDriver.main(AdvisoryDriver.java:115) 
> ~[orion.jar:?]
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 
> ~[?:1.8.0_352]
>         at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
> ~[?:1.8.0_352]
>         at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>  ~[?:1.8.0_352]
>         at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_352]
>         at 
> org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52) 
> ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:1020)
>  ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:192) 
> ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:215) 
> ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:91) 
> ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1111) 
> ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1120) 
> ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) 
> ~[spark-core_2.12-3.4.0.jar:3.4.0]% {code}
> Error with JDK8u362
> {code:java}
> org.apache.spark.SparkException: Job aborted due to stage failure: Task 
> serialization failed: java.lang.NoSuchMethodError: 
> java.nio.ByteBuffer.flip()Ljava/nio/ByteBuffer;
> java.lang.NoSuchMethodError: java.nio.ByteBuffer.flip()Ljava/nio/ByteBuffer;
>         at 
> org.apache.spark.util.io.ChunkedByteBufferOutputStream.toChunkedByteBuffer(ChunkedByteBufferOutputStream.scala:115)
>         at 
> org.apache.spark.broadcast.TorrentBroadcast$.blockifyObject(TorrentBroadcast.scala:362)
>         at 
> org.apache.spark.broadcast.TorrentBroadcast.writeBlocks(TorrentBroadcast.scala:160)
>         at 
> org.apache.spark.broadcast.TorrentBroadcast.<init>(TorrentBroadcast.scala:99)
>         at 
> org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:38)
>         at 
> org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:78)
>         at 
> org.apache.spark.SparkContext.broadcastInternal(SparkContext.scala:1548)
>         at org.apache.spark.SparkContext.broadcast(SparkContext.scala:1530)
>         at 
> org.apache.spark.scheduler.DAGScheduler.submitMissingTasks(DAGScheduler.scala:1535)
>         at 
> org.apache.spark.scheduler.DAGScheduler.submitStage(DAGScheduler.scala:1353)
>         at 
> org.apache.spark.scheduler.DAGScheduler.handleMapStageSubmitted(DAGScheduler.scala:1334)
>         at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:2934)
>         at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2923)
>         at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2912)
>         at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49)
>         at 
> org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2785)
>  ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:2721)
>  ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:2720)
>  ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at 
> scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62) 
> ~[scala-library-2.12.17.jar:?]
>         at 
> scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55) 
> ~[scala-library-2.12.17.jar:?]
>         at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49) 
> ~[scala-library-2.12.17.jar:?]
>         at 
> org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:2720) 
> ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.scheduler.DAGScheduler.submitMissingTasks(DAGScheduler.scala:1545)
>  ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.scheduler.DAGScheduler.submitStage(DAGScheduler.scala:1353) 
> ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.scheduler.DAGScheduler.handleMapStageSubmitted(DAGScheduler.scala:1334)
>  ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:2934)
>  ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2923)
>  ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2912)
>  ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49) 
> ~[spark-core_2.12-3.4.0.jar:3.4.0]
> Caused by: java.lang.NoSuchMethodError: 
> java.nio.ByteBuffer.flip()Ljava/nio/ByteBuffer;
>         at 
> org.apache.spark.util.io.ChunkedByteBufferOutputStream.toChunkedByteBuffer(ChunkedByteBufferOutputStream.scala:115)
>  ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.broadcast.TorrentBroadcast$.blockifyObject(TorrentBroadcast.scala:362)
>  ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.broadcast.TorrentBroadcast.writeBlocks(TorrentBroadcast.scala:160)
>  ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.broadcast.TorrentBroadcast.<init>(TorrentBroadcast.scala:99) 
> ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:38)
>  ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:78)
>  ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.SparkContext.broadcastInternal(SparkContext.scala:1548) 
> ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at org.apache.spark.SparkContext.broadcast(SparkContext.scala:1530) 
> ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         at 
> org.apache.spark.scheduler.DAGScheduler.submitMissingTasks(DAGScheduler.scala:1535)
>  ~[spark-core_2.12-3.4.0.jar:3.4.0]
>         ... 6 more 
>  {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to