xwmr-max opened a new issue #2627:
URL: https://github.com/apache/iceberg/issues/2627


   When multiple pieces of data with the same primary key value are inserted in 
the same batch in iceberg, query through flink sql, the data cannot be queried, 
and the following error will be reported:
   java.lang.IllegalArgumentException: Row arity: 3, but serializer arity: 2
   Caused by: java.lang.IllegalArgumentException: Row arity: 3, but serializer
   arity: 2
           at
   
org.apache.flink.table.runtime.typeutils.RowDataSerializer.copy(RowDataSerializer.java:117)
           at
   
org.apache.flink.table.runtime.typeutils.RowDataSerializer.copy(RowDataSerializer.java:50)
           at
   
org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.pushToOperator(OperatorChain.java:715)
           at
   
org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:692)
           at
   
org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:672)
           at
   
org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:52)
           at
   
org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:30)
           at
   
org.apache.flink.streaming.api.operators.StreamSourceContexts$ManualWatermarkContext.processAndCollect(StreamSourceContexts.java:305)
           at
   
org.apache.flink.streaming.api.operators.StreamSourceContexts$WatermarkContext.collect(StreamSourceContexts.java:394)
           at
   
org.apache.flink.streaming.api.functions.source.InputFormatSourceFunction.run(InputFormatSourceFunction.java:94)
           at
   
org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:100)
           at
   
org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:63)
           at
   
org.apache.flink.streaming.runtime.tasks.SourceStreamTask$LegacySourceFunctionThread.run(SourceStreamTask.java:201)
   
   This error means that the original data is two columns, and the schema of 
the table is also two columns, but now the schema has become three columns.
   But when the primary keys are different, inserting multiple pieces of data 
in the same batch can be queried normally, and can also be upsert


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
[email protected]



---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to