[jira] [Updated] (HUDI-1711) Avro Schema Exception with Spark 3.0 in 0.7
[ https://issues.apache.org/jira/browse/HUDI-1711?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nishith Agarwal updated HUDI-1711: -- Labels: sev:critical user-support-issues (was: sev:triage user-support-issues) > Avro Schema Exception with Spark 3.0 in 0.7 > --- > > Key: HUDI-1711 > URL: https://issues.apache.org/jira/browse/HUDI-1711 > Project: Apache Hudi > Issue Type: Bug > Components: DeltaStreamer >Reporter: Balaji Varadarajan >Assignee: sivabalan narayanan >Priority: Major > Labels: sev:critical, user-support-issues > > GH: [https://github.com/apache/hudi/issues/2705] > > > {{21/03/22 10:10:35 WARN util.package: Truncated the string representation of > a plan since it was too large. This behavior can be adjusted by setting > 'spark.sql.debug.maxToStringFields'. > 21/03/22 10:10:35 ERROR executor.Executor: Exception in task 0.0 in stage 1.0 > (TID 1) > java.lang.RuntimeException: Error while decoding: > java.lang.NegativeArraySizeException: -1255727808 > createexternalrow(if (isnull(input[0, > struct, > true])) null else createexternalrow(if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].id, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].name.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].type.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].url.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].user.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].password.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].create_time.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].create_user.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].update_time.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].update_user.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].del_flag, StructField(id,IntegerType,false), > StructField(name,StringType,true), StructField(type,StringType,true), > StructField(url,StringType,true), StructField(user,StringType,true), > StructField(password,StringType,true), > StructField(create_time,StringType,true), > StructField(create_user,StringType,true), > StructField(update_time,StringType,true), > StructField(update_user,StringType,true), > StructField(del_flag,IntegerType,true)), if (isnull(input[1, > struct, > true])) null else createexternalrow(if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].id, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].name.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].type.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].url.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].user.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].password.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].create_time.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].create_user.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].update_time.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].update_user.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].del_flag, StructField(id,IntegerType,false), > StructField(name,StringType,true), StructField(type,StringType,true), > StructField(url,StringType,true), StructField(user,StringType,true), > StructField(password,StringType,true), > StructField(create_time,StringType,true), > StructField(create_user,StringType,true), > StructField(update_time,StringType,true), > StructField(update_user,StringType,true), > StructField(del_flag,IntegerType,true)), if (isnull(input[2, > struct, > false])) null else createexternalrow(if (input[2, > struct, > false].isNullAt) null else input[2, > struct, > false].version.toString, if (input[2, > struct, > false].isNullAt) null else input[2, > struct, > false].connector.toString, if (input[2, > struct, > false].isNullAt) null else input[2, > struct, > false].name.toString, if (input[2, > struct, > false].isNullAt) null else input[2, > struct, > false].ts_ms, if (input[2, > struct, > false].isNullAt) null else input[2, > struct, > false].snapshot.toString, if (input
[jira] [Updated] (HUDI-1711) Avro Schema Exception with Spark 3.0 in 0.7
[ https://issues.apache.org/jira/browse/HUDI-1711?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] sivabalan narayanan updated HUDI-1711: -- Labels: sev:triage user-support-issues (was: sev:critical user-support-issues) > Avro Schema Exception with Spark 3.0 in 0.7 > --- > > Key: HUDI-1711 > URL: https://issues.apache.org/jira/browse/HUDI-1711 > Project: Apache Hudi > Issue Type: Bug > Components: DeltaStreamer >Reporter: Balaji Varadarajan >Priority: Major > Labels: sev:triage, user-support-issues > > GH: [https://github.com/apache/hudi/issues/2705] > > > {{21/03/22 10:10:35 WARN util.package: Truncated the string representation of > a plan since it was too large. This behavior can be adjusted by setting > 'spark.sql.debug.maxToStringFields'. > 21/03/22 10:10:35 ERROR executor.Executor: Exception in task 0.0 in stage 1.0 > (TID 1) > java.lang.RuntimeException: Error while decoding: > java.lang.NegativeArraySizeException: -1255727808 > createexternalrow(if (isnull(input[0, > struct, > true])) null else createexternalrow(if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].id, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].name.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].type.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].url.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].user.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].password.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].create_time.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].create_user.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].update_time.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].update_user.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].del_flag, StructField(id,IntegerType,false), > StructField(name,StringType,true), StructField(type,StringType,true), > StructField(url,StringType,true), StructField(user,StringType,true), > StructField(password,StringType,true), > StructField(create_time,StringType,true), > StructField(create_user,StringType,true), > StructField(update_time,StringType,true), > StructField(update_user,StringType,true), > StructField(del_flag,IntegerType,true)), if (isnull(input[1, > struct, > true])) null else createexternalrow(if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].id, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].name.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].type.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].url.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].user.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].password.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].create_time.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].create_user.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].update_time.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].update_user.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].del_flag, StructField(id,IntegerType,false), > StructField(name,StringType,true), StructField(type,StringType,true), > StructField(url,StringType,true), StructField(user,StringType,true), > StructField(password,StringType,true), > StructField(create_time,StringType,true), > StructField(create_user,StringType,true), > StructField(update_time,StringType,true), > StructField(update_user,StringType,true), > StructField(del_flag,IntegerType,true)), if (isnull(input[2, > struct, > false])) null else createexternalrow(if (input[2, > struct, > false].isNullAt) null else input[2, > struct, > false].version.toString, if (input[2, > struct, > false].isNullAt) null else input[2, > struct, > false].connector.toString, if (input[2, > struct, > false].isNullAt) null else input[2, > struct, > false].name.toString, if (input[2, > struct, > false].isNullAt) null else input[2, > struct, > false].ts_ms, if (input[2, > struct, > false].isNullAt) null else input[2, > struct, > false].snapshot.toString, if (input[2, > struct, > false].isNullAt) nu
[jira] [Updated] (HUDI-1711) Avro Schema Exception with Spark 3.0 in 0.7
[ https://issues.apache.org/jira/browse/HUDI-1711?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] sivabalan narayanan updated HUDI-1711: -- Labels: sev:critical user-support-issues (was: ) > Avro Schema Exception with Spark 3.0 in 0.7 > --- > > Key: HUDI-1711 > URL: https://issues.apache.org/jira/browse/HUDI-1711 > Project: Apache Hudi > Issue Type: Bug > Components: DeltaStreamer >Reporter: Balaji Varadarajan >Priority: Major > Labels: sev:critical, user-support-issues > > GH: [https://github.com/apache/hudi/issues/2705] > > > {{21/03/22 10:10:35 WARN util.package: Truncated the string representation of > a plan since it was too large. This behavior can be adjusted by setting > 'spark.sql.debug.maxToStringFields'. > 21/03/22 10:10:35 ERROR executor.Executor: Exception in task 0.0 in stage 1.0 > (TID 1) > java.lang.RuntimeException: Error while decoding: > java.lang.NegativeArraySizeException: -1255727808 > createexternalrow(if (isnull(input[0, > struct, > true])) null else createexternalrow(if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].id, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].name.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].type.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].url.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].user.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].password.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].create_time.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].create_user.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].update_time.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].update_user.toString, if (input[0, > struct, > true].isNullAt) null else input[0, > struct, > true].del_flag, StructField(id,IntegerType,false), > StructField(name,StringType,true), StructField(type,StringType,true), > StructField(url,StringType,true), StructField(user,StringType,true), > StructField(password,StringType,true), > StructField(create_time,StringType,true), > StructField(create_user,StringType,true), > StructField(update_time,StringType,true), > StructField(update_user,StringType,true), > StructField(del_flag,IntegerType,true)), if (isnull(input[1, > struct, > true])) null else createexternalrow(if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].id, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].name.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].type.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].url.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].user.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].password.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].create_time.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].create_user.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].update_time.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].update_user.toString, if (input[1, > struct, > true].isNullAt) null else input[1, > struct, > true].del_flag, StructField(id,IntegerType,false), > StructField(name,StringType,true), StructField(type,StringType,true), > StructField(url,StringType,true), StructField(user,StringType,true), > StructField(password,StringType,true), > StructField(create_time,StringType,true), > StructField(create_user,StringType,true), > StructField(update_time,StringType,true), > StructField(update_user,StringType,true), > StructField(del_flag,IntegerType,true)), if (isnull(input[2, > struct, > false])) null else createexternalrow(if (input[2, > struct, > false].isNullAt) null else input[2, > struct, > false].version.toString, if (input[2, > struct, > false].isNullAt) null else input[2, > struct, > false].connector.toString, if (input[2, > struct, > false].isNullAt) null else input[2, > struct, > false].name.toString, if (input[2, > struct, > false].isNullAt) null else input[2, > struct, > false].ts_ms, if (input[2, > struct, > false].isNullAt) null else input[2, > struct, > false].snapshot.toString, if (input[2, > struct, > false].isNullAt) null else input[2, > struct, >