That issue is closed, it looks like just a user error.

On Sun, Jan 24, 2021 at 12:07 PM Ryan Murray <rym...@dremio.com> wrote:

> Based on #2144 <https://github.com/apache/iceberg/issues/2144> I would
> change my vote to -1. I will raise a fix asap.
>
>
> On Sun, Jan 24, 2021 at 3:12 PM Ryan Murray <rym...@dremio.com> wrote:
>
>>
>> +1 (non-binding)
>>
>> I verified the build and ran the tests. Also verified both flink and
>> spark custom catalogs are working.
>>
>> One side note: I had to run the tests a few times to get the build to
>> pass. Flaky Hive tests.
>>
>> Best,
>>
>> Ryan Murray
>>
>> On Sat, Jan 23, 2021 at 12:26 AM Jack Ye <yezhao...@gmail.com> wrote:
>>
>>> Hi everyone,
>>>
>>> I propose the following RC to be released as the official Apache Iceberg
>>> 0.11.0 release. The RC is also reviewed and signed by Ryan Blue.
>>>
>>> The commit id is ad78cc6cf259b7a0c66ab5de6675cc005febd939
>>>
>>> This corresponds to the tag: apache-iceberg-0.11.0-rc0
>>> * https://github.com/apache/iceberg/commits/apache-iceberg-0.11.0-rc0
>>> * https://github.com/apache/iceberg/tree/apache-iceberg-0.11.0-rc0
>>>
>>> The release tarball, signature, and checksums are here:
>>> *
>>> https://dist.apache.org/repos/dist/dev/iceberg/apache-iceberg-0.11.0-rc0
>>>
>>> You can find the KEYS file here:
>>> * https://dist.apache.org/repos/dist/dev/iceberg/KEYS
>>>
>>> Convenience binary artifacts are staged in Nexus. The Maven repository
>>> URL is:
>>> *
>>> https://repository.apache.org/content/repositories/orgapacheiceberg-1015
>>>
>>> This release includes the following changes:
>>>
>>> *High-level features*
>>>
>>>    - Core API now supports partition spec and sort order evolution
>>>    - Spark 3 now supports the following SQL extensions:
>>>       - MERGE INTO
>>>       - DELETE FROM
>>>       - ALTER TABLE ... ADD/DROP PARTITION
>>>       - ALTER TABLE ... WRITE ORDERED BY
>>>       - invoke stored procedures using CALL
>>>    - Flink now supports streaming reads, CDC writes (experimental), and
>>>    filter pushdown
>>>    - AWS module is added to support better integration with AWS, with AWS
>>>    Glue catalog <https://aws.amazon.com/glue> support and dedicated S3
>>>    FileIO implementation
>>>    - Nessie module is added to support integration with project Nessie
>>>    <https://projectnessie.org>
>>>
>>> *Important bug fixes*
>>>
>>>    - #1981 fixes date and timestamp transforms
>>>    - #2091 fixes Parquet vectorized reads when column types are promoted
>>>    - #1962 fixes Parquet vectorized position reader
>>>    - #1991 fixes Avro schema conversions to preserve field docs
>>>    - #1811 makes refreshing Spark cache optional
>>>    - #1798 fixes read failure when encountering duplicate entries of
>>>    data files
>>>    - #1785 fixes invalidation of metadata tables in CachingCatalog
>>>    - #1784 fixes resolving of SparkSession table's metadata tables
>>>
>>> *Other notable changes*
>>>
>>>    - NaN counter is added to format v2 metrics
>>>    - Shared catalog properties are added in core library to standardize
>>>    catalog level configurations
>>>    - Spark and Flink now supports dynamically loading customized
>>>    `Catalog` and `FileIO` implementations
>>>    - Spark now supports loading tables with file paths via HadoopTables
>>>    - Spark 2 now supports loading tables from other catalogs, like
>>>    Spark 3
>>>    - Spark 3 now supports catalog names in DataFrameReader when using
>>>    Iceberg as a format
>>>    - Hive now supports INSERT INTO, case insensitive query, projection
>>>    pushdown, create DDL with schema and auto type conversion
>>>    - ORC now supports reading tinyint, smallint, char, varchar types
>>>    - Hadoop catalog now supports role-based access of table listing
>>>
>>> Please download, verify, and test.
>>>
>>> Please vote in the next 72 hours.
>>>
>>> [ ] +1 Release this as Apache Iceberg 0.11.0
>>> [ ] +0
>>> [ ] -1 Do not release this because...
>>>
>>

Reply via email to