[ https://issues.apache.org/jira/browse/SPARK-43278?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
jiangjiguang0719 updated SPARK-43278: ------------------------------------- Description: Java version: 1.8.0_331, Apache Maven 3.8.4 I run next steps: # git clone [https://github.com/apache/spark.git] # git checkout -b v3.3.0 3.3.0 # mvn clean install -DskipTests # copy hive-site.xml to examples/src/main/resources/ # execute TPC-H Q6 {code:java} public static void main(String[] args) throws InterruptedException { SparkConf sparkConf = new SparkConf() .setAppName("demo") .setMaster("local[1]") ; SparkSession sparkSession = SparkSession.builder() .config(sparkConf) .enableHiveSupport() .getOrCreate(); sparkSession.sql("use local_tpch_sf10_uncompressed_etl"); sparkSession.sql(TPCH.SQL6).show(); } {code} get the error info: Exception in thread "main" java.lang.NoSuchMethodError: java.nio.ByteBuffer.flip()Ljava/nio/ByteBuffer; at org.apache.spark.util.io.ChunkedByteBufferOutputStream.toChunkedByteBuffer(ChunkedByteBufferOutputStream.scala:115) at org.apache.spark.broadcast.TorrentBroadcast$.blockifyObject(TorrentBroadcast.scala:325) at org.apache.spark.broadcast.TorrentBroadcast.writeBlocks(TorrentBroadcast.scala:140) at org.apache.spark.broadcast.TorrentBroadcast.<init>(TorrentBroadcast.scala:95) at org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:34) at org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:75) at org.apache.spark.SparkContext.broadcast(SparkContext.scala:1529) at org.apache.spark.sql.execution.datasources.parquet.ParquetFileFormat.buildReaderWithPartitionValues(ParquetFileFormat.scala:235) at org.apache.spark.sql.execution.FileSourceScanExec.inputRDD$lzycompute(DataSourceScanExec.scala:457) at org.apache.spark.sql.execution.FileSourceScanExec.inputRDD(DataSourceScanExec.scala:448) at org.apache.spark.sql.execution.FileSourceScanExec.doExecuteColumnar(DataSourceScanExec.scala:547) at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeColumnar$1(SparkPlan.scala:221) at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(SparkPlan.scala:232) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:229) at org.apache.spark.sql.execution.SparkPlan.executeColumnar(SparkPlan.scala:217) was: Java version: 1.8.0_331, Apache Maven 3.8.4 I run next steps: # git clone [https://github.com/apache/spark.git] # git checkout -b v3.3.0 3.3.0 # mvn clean install -DskipTests # copy hive-site.xml to examples/src/main/resources/ # execute TPC-H Q6 !image-2023-04-25-17-14-50-392.png|width=437,height=246! get the error info !image-2023-04-25-17-15-57-874.png|width=466,height=161! > Exception in thread "main" java.lang.NoSuchMethodError: > java.nio.ByteBuffer.flip()Ljava/nio/ByteBuffer; > ------------------------------------------------------------------------------------------------------- > > Key: SPARK-43278 > URL: https://issues.apache.org/jira/browse/SPARK-43278 > Project: Spark > Issue Type: Bug > Components: Java API > Affects Versions: 3.3.0 > Reporter: jiangjiguang0719 > Priority: Major > > Java version: 1.8.0_331, Apache Maven 3.8.4 > I run next steps: > # git clone [https://github.com/apache/spark.git] > # git checkout -b v3.3.0 3.3.0 > # mvn clean install -DskipTests > # copy hive-site.xml to examples/src/main/resources/ > # execute TPC-H Q6 > > {code:java} > public static void main(String[] args) throws InterruptedException { > SparkConf sparkConf = new SparkConf() > .setAppName("demo") > .setMaster("local[1]") > ; > SparkSession sparkSession = SparkSession.builder() > .config(sparkConf) > .enableHiveSupport() > .getOrCreate(); > sparkSession.sql("use local_tpch_sf10_uncompressed_etl"); > sparkSession.sql(TPCH.SQL6).show(); > } {code} > > > get the error info: > Exception in thread "main" java.lang.NoSuchMethodError: > java.nio.ByteBuffer.flip()Ljava/nio/ByteBuffer; > at > org.apache.spark.util.io.ChunkedByteBufferOutputStream.toChunkedByteBuffer(ChunkedByteBufferOutputStream.scala:115) > at > org.apache.spark.broadcast.TorrentBroadcast$.blockifyObject(TorrentBroadcast.scala:325) > at > org.apache.spark.broadcast.TorrentBroadcast.writeBlocks(TorrentBroadcast.scala:140) > at > org.apache.spark.broadcast.TorrentBroadcast.<init>(TorrentBroadcast.scala:95) > at > org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:34) > at > org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:75) > at org.apache.spark.SparkContext.broadcast(SparkContext.scala:1529) > at > org.apache.spark.sql.execution.datasources.parquet.ParquetFileFormat.buildReaderWithPartitionValues(ParquetFileFormat.scala:235) > at > org.apache.spark.sql.execution.FileSourceScanExec.inputRDD$lzycompute(DataSourceScanExec.scala:457) > at > org.apache.spark.sql.execution.FileSourceScanExec.inputRDD(DataSourceScanExec.scala:448) > at > org.apache.spark.sql.execution.FileSourceScanExec.doExecuteColumnar(DataSourceScanExec.scala:547) > at > org.apache.spark.sql.execution.SparkPlan.$anonfun$executeColumnar$1(SparkPlan.scala:221) > at > org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(SparkPlan.scala:232) > at > org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) > at > org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:229) > at > org.apache.spark.sql.execution.SparkPlan.executeColumnar(SparkPlan.scala:217) -- This message was sent by Atlassian Jira (v8.20.10#820010) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org