Akshay-connectwise commented on issue #2040:
URL: https://github.com/apache/iceberg/issues/2040#issuecomment-1420569509
Hi @RussellSpitzer,
Have tried setting up the table properties but it didn't worked for me below
is table creation query and error when I have an extra column "comment" . This
is using Pyspark in jupyter notebook
-->spark.sql("CREATE TABLE iceberg.employee ( \
firstname string, \
middlename string, \
lastname string, \
id string, \
gender string, \
salary bigint ) \
USING iceberg \
TBLPROPERTIES
('SPARK_WRITE_ACCEPT_ANY_SCHEMA'='write.spark.accept-any-schema')")
Error:
AnalysisException:
Cannot write to 'local_schema_evo.iceberg.employee', too many data columns:
Table columns: 'firstname', 'middlename', 'lastname', 'id', 'gender',
'salary'
Data columns: 'firstname', 'middlename', 'lastname', 'id', 'gender',
'salary', 'comment'
While checking table properties this property is enabled but still not
working. Please suggest if we have any other way for schema evolution without
using Alter queries.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]