Hi,
I am trying to create partitioned iceberg table using scala code below based on
example in docs.
df_c.writeTo(output_table)
.partitionBy(days(col("last_updated")))
.createOrReplace()
However, this code does not compile and throws two errors:
value partitionBy is not a member of
org.apache.spark.sql.DataFrameWriterV2[org.apache.spark.sql.Row]
[error] possible cause: maybe a semicolon is missing before `value partitionBy'?
[error] .partitionBy(days(col("last_updated")))
[error] ^
[error] not found: value days
[error] .partitionBy(days(col("last_updated")))
[error] ^
[error] two errors found
Not sure where to look for a problem. Any advice appreciated.
Best regards,
Saulius Pakalka