heman026 commented on issue #1759:
URL:
https://github.com/apache/iceberg-python/issues/1759#issuecomment-2700632648
> can you share more on what code you ran? specifically what is `table`,
`pyarrow_table`, and `join_cols`
```
iceberg_table = 'test.table1'
table = catalog.load_table(iceberg_table)
data= table.scan(row_filter=E.EqualTo('id', 1)).to_arrow()
```
icerbeg table is fitlered and set to data pyarrow table). I am doing some
computation on data and modify one of the column values. I update the columns
values to the original iceberg table using upsert.
`table.upsert(data,join_cols=['id','col1','col2,'col3'])`
This is a sample code. The iceberg table which I am testing with has 144
million rows with 14 columns. And the filtered data has 500 rows with the same
number of columns. I have 4 primary key columns which I have provided in
join_cols argument of upsert method.
If the filtered data has around 400 rows, upsert method is working fine, if
it has around 500 data, it is giving the above exception.
Let me know if you need more details
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]