[ https://issues.apache.org/jira/browse/SPARK-31068?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Sean R. Owen reassigned SPARK-31068: ------------------------------------ Assignee: Lantao Jin > IllegalArgumentException in BroadcastExchangeExec > ------------------------------------------------- > > Key: SPARK-31068 > URL: https://issues.apache.org/jira/browse/SPARK-31068 > Project: Spark > Issue Type: Bug > Components: SQL > Affects Versions: 3.0.0 > Reporter: Lantao Jin > Assignee: Lantao Jin > Priority: Major > > {code} > Caused by: org.apache.spark.SparkException: Failed to materialize query > stage: BroadcastQueryStage 0 > +- BroadcastExchange HashedRelationBroadcastMode(List(input[0, string, true], > input[1, bigint, true], input[2, int, true])) > +- *(1) Project [guid#138126, session_skey#138127L, seqnum#138132] > +- *(1) Filter ((((isnotnull(session_start_dt#138129) && > (session_start_dt#138129 = 2020-01-01)) && isnotnull(seqnum#138132)) && > isnotnull(session_skey#138127L)) && isnotnull(guid#138126)) > +- *(1) FileScan parquet p_soj_cl_t.clav_events[guid#138126, > session_skey#138127L, session_start_dt#138129, seqnum#138132] DataFilters: > [isnotnull(session_start_dt#138129), (session_start_dt#138129 = 2020-01-01), > isnotnull(seqnum#138..., Format: Parquet, Location: > TahoeLogFileIndex[hdfs://hermes-rno/workspaces/P_SOJ_CL_T/clav_events], > PartitionFilters: [], PushedFilters: [IsNotNull(session_start_dt), > EqualTo(session_start_dt,2020-01-01), IsNotNull(seqnum), IsNotNull(..., > ReadSchema: > struct<guid:string,session_skey:bigint,session_start_dt:string,seqnum:int>, > SelectedBucketsCount: 1000 out of 1000, UsedIndexes: [] > at > org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec$$anonfun$generateFinalPlan$3.apply(AdaptiveSparkPlanExec.scala:230) > at > org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec$$anonfun$generateFinalPlan$3.apply(AdaptiveSparkPlanExec.scala:225) > at scala.collection.immutable.List.foreach(List.scala:381) > at > org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.generateFinalPlan(AdaptiveSparkPlanExec.scala:225) > ... 48 more > Caused by: java.lang.IllegalArgumentException: Initial capacity 670166426 > exceeds maximum capacity of 536870912 > at > org.apache.spark.unsafe.map.BytesToBytesMap.<init>(BytesToBytesMap.java:196) > at > org.apache.spark.unsafe.map.BytesToBytesMap.<init>(BytesToBytesMap.java:219) > at > org.apache.spark.sql.execution.joins.UnsafeHashedRelation$.apply(HashedRelation.scala:340) > at > org.apache.spark.sql.execution.joins.HashedRelation$.apply(HashedRelation.scala:123) > at > org.apache.spark.sql.execution.joins.HashedRelationBroadcastMode.transform(HashedRelation.scala:964) > at > org.apache.spark.sql.execution.joins.HashedRelationBroadcastMode.transform(HashedRelation.scala:952) > at > org.apache.spark.sql.execution.exchange.BroadcastExchangeExec$$anonfun$relationFuture$1$$anonfun$apply$9.apply(BroadcastExchangeExec.scala:220) > at > org.apache.spark.sql.execution.exchange.BroadcastExchangeExec$$anonfun$relationFuture$1$$anonfun$apply$9.apply(BroadcastExchangeExec.scala:207) > at > org.apache.spark.sql.execution.SQLExecution$.withExecutionId(SQLExecution.scala:128) > at > org.apache.spark.sql.execution.exchange.BroadcastExchangeExec$$anonfun$relationFuture$1.apply(BroadcastExchangeExec.scala:206) > at > org.apache.spark.sql.execution.exchange.BroadcastExchangeExec$$anonfun$relationFuture$1.apply(BroadcastExchangeExec.scala:172) > at > scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24) > at > scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24) > ... 3 more > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org