[jira] [Updated] (SPARK-9340) CatalystSchemaConverter and CatalystRowConverter don't handle unannotated repeated fields correctly
[ https://issues.apache.org/jira/browse/SPARK-9340?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Cheng Lian updated SPARK-9340: -- Sprint: Spark 1.5 doc/QA sprint Target Version/s: 1.5.0 CatalystSchemaConverter and CatalystRowConverter don't handle unannotated repeated fields correctly --- Key: SPARK-9340 URL: https://issues.apache.org/jira/browse/SPARK-9340 Project: Spark Issue Type: Bug Components: SQL Affects Versions: 1.2.0, 1.3.0, 1.4.0, 1.5.0 Reporter: Damian Guy Assignee: Cheng Lian Attachments: ParquetTypesConverterTest.scala SPARK-6776 and SPARK-6777 followed {{parquet-avro}} to implement backwards-compatibility rules defined in {{parquet-format}} spec. However, both Spark SQL and {{parquet-avro}} neglected the following statement in {{parquet-format}}: {quote} This does not affect repeated fields that are not annotated: A repeated field that is neither contained by a {{LIST}}- or {{MAP}}-annotated group nor annotated by {{LIST}} or {{MAP}} should be interpreted as a required list of required elements where the element type is the type of the field. {quote} One of the consequences is that, Parquet files generated by {{parquet-protobuf}} containing unannotated repeated fields are not correctly converted to Catalyst arrays. For example, the following Parquet schema {noformat} message root { repeated int32 f1 } {noformat} should be converted to {noformat} StructType(StructField(f1, ArrayType(IntegerType, containsNull = false), nullable = false) :: Nil) {noformat} But now it triggers an {{AnalysisException}}. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Updated] (SPARK-9340) CatalystSchemaConverter and CatalystRowConverter don't handle unannotated repeated fields correctly
[ https://issues.apache.org/jira/browse/SPARK-9340?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Cheng Lian updated SPARK-9340: -- Summary: CatalystSchemaConverter and CatalystRowConverter don't handle unannotated repeated fields correctly (was: ParquetTypeConverter incorrectly handling of repeated types results in schema mismatch) CatalystSchemaConverter and CatalystRowConverter don't handle unannotated repeated fields correctly --- Key: SPARK-9340 URL: https://issues.apache.org/jira/browse/SPARK-9340 Project: Spark Issue Type: Bug Components: SQL Affects Versions: 1.2.0, 1.3.0, 1.4.0, 1.5.0 Reporter: Damian Guy Attachments: ParquetTypesConverterTest.scala The way ParquetTypesConverter handles primitive repeated types results in an incompatible schema being used for querying data. For example, given a schema like so: message root { repeated int32 repeated_field; } Spark produces a read schema like: message root { optional int32 repeated_field; } These are incompatible and all attempts to read fail. In ParquetTypesConverter.toDataType: if (parquetType.isPrimitive) { toPrimitiveDataType(parquetType.asPrimitiveType, isBinaryAsString, isInt96AsTimestamp) } else {...} The if condition should also have !parquetType.isRepetition(Repetition.REPEATED) And then this case will need to be handled in the else -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Updated] (SPARK-9340) CatalystSchemaConverter and CatalystRowConverter don't handle unannotated repeated fields correctly
[ https://issues.apache.org/jira/browse/SPARK-9340?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Cheng Lian updated SPARK-9340: -- Description: SPARK-6776 and SPARK-6777 followed {{parquet-avro}} to implement backwards-compatibility rules defined in {{parquet-format}} spec. However, both Spark SQL and {{parquet-avro}} neglected the following statement in {{parquet-format}}: {quote} This does not affect repeated fields that are not annotated: A repeated field that is neither contained by a {{LIST}}- or {{MAP}}-annotated group nor annotated by {{LIST}} or {{MAP}} should be interpreted as a required list of required elements where the element type is the type of the field. {quote} One of the consequences is that, Parquet files generated by {{parquet-protobuf}} containing unannotated repeated fields are not correctly converted to Catalyst arrays. For example, the following Parquet schema {noformat} message root { repeated int32 f1 } {noformat} should be converted to {noformat} StructType(StructField(f1, ArrayType(IntegerType, containsNull = false), nullable = false) :: Nil) {noformat} But now it triggers an {{AnalysisException}}. was: The way ParquetTypesConverter handles primitive repeated types results in an incompatible schema being used for querying data. For example, given a schema like so: message root { repeated int32 repeated_field; } Spark produces a read schema like: message root { optional int32 repeated_field; } These are incompatible and all attempts to read fail. In ParquetTypesConverter.toDataType: if (parquetType.isPrimitive) { toPrimitiveDataType(parquetType.asPrimitiveType, isBinaryAsString, isInt96AsTimestamp) } else {...} The if condition should also have !parquetType.isRepetition(Repetition.REPEATED) And then this case will need to be handled in the else CatalystSchemaConverter and CatalystRowConverter don't handle unannotated repeated fields correctly --- Key: SPARK-9340 URL: https://issues.apache.org/jira/browse/SPARK-9340 Project: Spark Issue Type: Bug Components: SQL Affects Versions: 1.2.0, 1.3.0, 1.4.0, 1.5.0 Reporter: Damian Guy Attachments: ParquetTypesConverterTest.scala SPARK-6776 and SPARK-6777 followed {{parquet-avro}} to implement backwards-compatibility rules defined in {{parquet-format}} spec. However, both Spark SQL and {{parquet-avro}} neglected the following statement in {{parquet-format}}: {quote} This does not affect repeated fields that are not annotated: A repeated field that is neither contained by a {{LIST}}- or {{MAP}}-annotated group nor annotated by {{LIST}} or {{MAP}} should be interpreted as a required list of required elements where the element type is the type of the field. {quote} One of the consequences is that, Parquet files generated by {{parquet-protobuf}} containing unannotated repeated fields are not correctly converted to Catalyst arrays. For example, the following Parquet schema {noformat} message root { repeated int32 f1 } {noformat} should be converted to {noformat} StructType(StructField(f1, ArrayType(IntegerType, containsNull = false), nullable = false) :: Nil) {noformat} But now it triggers an {{AnalysisException}}. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org