Amogh, thanks for sharing the input. Conceptually, I agree that it is good
to fix those issues in the 1.10.0 release.

My main concern is that a couple of them are large efforts and still in
early draft status. It may take a few weeks to get them in. Since those
bugs aren't introduced by the 1.10.0 release, can they be included in the
1.11.0 release?

> we should indeed block for fixing those since it doesn't seem right to do
another release which would further amplify the problem.

I am not sure if a new release further amplifies the problem since it is
the same for users if they pick up a new release of 1.9 or 1.10, and use
Spark for V3 tables. Definitely interested in hearing others' take on this.

On Mon, Jun 23, 2025 at 7:10 AM Amogh Jahagirdar <2am...@gmail.com> wrote:

> Thanks Steven,
>
> While I definitely agree that we don't hold releases for new features, I
> feel an important aspect to consider especially now that V3 is ratified is
> to make sure we've resolved any known issues that would propagate bad V3
> metadata. My take is basically if there are known issues from 1.9 in V3
> implementation which propagate spec incompliant metadata, we should indeed
> block for fixing those since it doesn't seem right to do another release
> which would further amplify the problem.
>
> Some examples
>
>
>    -  PR <https://github.com/apache/iceberg/pull/13061> for fixing an
>    issue in row lineage propagation when distributed planning is applied in
>    the Spark integration. Without this fix, row lineage metadata could get
>    corrupted.
>    - Also as of today for default value DDLs, Spark doesn't technically
>    support them yet (I have a PR
>    <https://github.com/apache/iceberg/pull/13107> for that as well, but I
>    think it'll take a bit more time, I need to look into handling struct
>    values better). However today, the spark integration silently accepts the
>    DDL but doesn't actually do anything. Though it doesn't produce non
>    compliant metadata it still does feel like a really misleading behavior. I
>    think at minimum for the next release we should probably just fail the DDL
>    if the PR doesn't get updated in time for handling default values for
>    struct fields more cleanly
>    - The timestamp nanos fix https://github.com/apache/iceberg/pull/11775
>    which I think was already called out in this thread
>    - Preventing orphan DVs <https://github.com/apache/iceberg/pull/13245>
>    since that's required by the V3 spec
>
> So all in all, before a 1.10 release I'd encourage folks to test out parts
> of the V3 work and anything that is either a correctness issue or produces
> spec incompliant metadata should be surfaced (again, imo it's OK if there's
> feature implementation gaps but at the same time don't want to potentially
> amplify known incompliance problems by doing a release before they're fixed)
>
> Thanks,
> Amogh Jahagirdar
>
> On Thu, Jun 19, 2025 at 2:36 AM Péter Váry <peter.vary.apa...@gmail.com>
> wrote:
>
>> If possible, I would love to have the File Format API interfaces approved
>> and merged: https://github.com/apache/iceberg/pull/12774
>> The effort is ongoing for half a year now, and not much change requested
>> lately.
>>
>> On Thu, Jun 19, 2025, 00:16 Steven Wu <stevenz...@gmail.com> wrote:
>>
>>> sorry, I meant 1.10.0 release. Thanks for catching the error, JB!
>>>
>>> On Wed, Jun 18, 2025 at 2:29 PM Jean-Baptiste Onofré <j...@nanthrax.net>
>>> wrote:
>>>
>>>> Hi
>>>>
>>>> I guess you mean 1.10.0 release :)
>>>>
>>>> Regards
>>>> JB
>>>>
>>>> On Wed, Jun 18, 2025 at 11:01 PM Steven Wu <stevenz...@gmail.com>
>>>> wrote:
>>>> >
>>>> > V3 related features reference implementation don’t have much
>>>> progress, which is probably not going to change significantly in the next 1
>>>> or 2 weeks. I would propose to cut the release branch by the end of next
>>>> Friday (June 27). There are a few important features to be released like
>>>> Spark 4.0 support, Flink 2.0 support, Flink dynamic sink etc. We typically
>>>> don't want to hold back releases for extended time to wait for new feature
>>>> implementations.
>>>> >
>>>> >
>>>> > There are 11 open and 6 closed issues/PRs for the 0.10.0 milestone
>>>> >
>>>> > https://github.com/apache/iceberg/milestone/54
>>>> >
>>>> >
>>>> > For the remaining open issues
>>>> >
>>>> > Flink: Dynamic Iceberg Sink Contribution. This is a large effort.
>>>> Seems that Max and Peter have merged all breakdown PRs. So it is on track.
>>>> >
>>>> > Core: Fix numeric overflow of timestamp nano literal. Still have some
>>>> discussion on the right approach for the short term and longer term
>>>> >
>>>> > Some of the other issues/PRs may need to be pushed to the next
>>>> release.
>>>> >
>>>> >
>>>> > Feedbacks are welcomed.
>>>> >
>>>> > Thanks,
>>>> > Steven
>>>>
>>>

Reply via email to