+1 non-binding on RC4. I tested out the Hive read path on a distributed
cluster using HadoopTables.

On Thu, 5 Nov 2020 at 04:46, Dongjoon Hyun <dongjoon.h...@gmail.com> wrote:

> +1 for 0.10.0 RC4.
>
> Bests,
> Dongjoon.
>
> On Wed, Nov 4, 2020 at 7:17 PM Jingsong Li <jingsongl...@gmail.com> wrote:
>
>> +1
>>
>> 1. Download the source tarball, signature (.asc), and checksum
>> (.sha512):   OK
>> 2. Import gpg keys: download KEYS and run gpg --import
>> /path/to/downloaded/KEYS (optional if this hasn’t changed) :  OK
>> 3. Verify the signature by running: gpg --verify
>> apache-iceberg-xx.tar.gz.asc:  OK
>> 4. Verify the checksum by running: sha512sum -c
>> apache-iceberg-xx.tar.gz.sha512 :  OK
>> 5. Untar the archive and go into the source directory: tar xzf
>> apache-iceberg-xx.tar.gz && cd apache-iceberg-xx:  OK
>> 6. Run RAT checks to validate license headers: dev/check-license: OK
>> 7. Build and test the project: ./gradlew build (use Java 8) :   OK
>>
>> Best,
>> Jingsong
>>
>> On Thu, Nov 5, 2020 at 7:38 AM Ryan Blue <rb...@netflix.com.invalid>
>> wrote:
>>
>>> +1
>>>
>>>    - Validated checksum and signature
>>>    - Ran license checks
>>>    - Built and ran tests
>>>    - Queried a Hadoop FS table created with 0.9.0 in Spark 3.0.1
>>>    - Created a Hive table from Spark 3.0.1
>>>    - Tested metadata tables from Spark
>>>    - Tested Hive and Hadoop table reads in Hive 2.3.7
>>>
>>> I was able to read both Hadoop and Hive tables created in Spark from
>>> Hive using:
>>>
>>> add jar /home/blue/Downloads/iceberg-hive-runtime-0.10.0.jar;
>>> create external table hadoop_table
>>>   stored by 'org.apache.iceberg.mr.hive.HiveIcebergStorageHandler'
>>>   location 'file:/home/blue/tmp/hadoop-warehouse/default/test';
>>> select * from hadoop_table;
>>>
>>> set iceberg.mr.catalog=hive;
>>> select * from hive_table;
>>>
>>> The hive_table needed engine.hive.enabled=true set in table properties
>>> by Spark using:
>>>
>>> alter table hive_table set tblproperties ('engine.hive.enabled'='true')
>>>
>>> Hive couldn’t read the #snapshots metadata table for Hadoop. It failed
>>> with this error:
>>>
>>> Failed with exception 
>>> java.io.IOException:org.apache.hadoop.hive.ql.metadata.HiveException: 
>>> java.lang.ClassCastException: java.lang.Long cannot be cast to 
>>> java.time.OffsetDateTime
>>>
>>> I also couldn’t read the Hadoop table once iceberg.mr.catalog was set
>>> in my environment, so I think we have a bit more work to do to clean up
>>> Hive table configuration.
>>>
>>> On Wed, Nov 4, 2020 at 12:54 AM Ryan Murray <rym...@dremio.com> wrote:
>>>
>>>> +1 (non-binding)
>>>>
>>>> 1. Download the source tarball, signature (.asc), and checksum
>>>> (.sha512):   OK
>>>> 2. Import gpg keys: download KEYS and run gpg --import
>>>> /path/to/downloaded/KEYS (optional if this hasn’t changed) :  OK
>>>> 3. Verify the signature by running: gpg --verify
>>>> apache-iceberg-xx.tar.gz.asc:  I got a warning "gpg: WARNING: This key is
>>>> not certified with a trusted signature! gpg:          There is no
>>>> indication that the signature belongs to the owner." but it passed
>>>> 4. Verify the checksum by running: sha512sum -c
>>>> apache-iceberg-xx.tar.gz.sha512 :  OK
>>>> 5. Untar the archive and go into the source directory: tar xzf
>>>> apache-iceberg-xx.tar.gz && cd apache-iceberg-xx:  OK
>>>> 6. Run RAT checks to validate license headers: dev/check-license: OK
>>>> 7. Build and test the project: ./gradlew build (use Java 8 & Java 11)
>>>> :   OK
>>>>
>>>>
>>>> On Wed, Nov 4, 2020 at 2:56 AM OpenInx <open...@gmail.com> wrote:
>>>>
>>>>> +1 for 0.10.0 RC4
>>>>>
>>>>> 1. Download the source tarball, signature (.asc), and checksum
>>>>> (.sha512):   OK
>>>>> 2. Import gpg keys: download KEYS and run gpg --import
>>>>> /path/to/downloaded/KEYS (optional if this hasn’t changed) :  OK
>>>>> 3. Verify the signature by running: gpg --verify
>>>>> apache-iceberg-xx.tar.gz.asc:  OK
>>>>> 4. Verify the checksum by running: sha512sum -c
>>>>> apache-iceberg-xx.tar.gz.sha512 :  OK
>>>>> 5. Untar the archive and go into the source directory: tar xzf
>>>>> apache-iceberg-xx.tar.gz && cd apache-iceberg-xx:  OK
>>>>> 6. Run RAT checks to validate license headers: dev/check-license: OK
>>>>> 7. Build and test the project: ./gradlew build (use Java 8) :   OK
>>>>>
>>>>> On Wed, Nov 4, 2020 at 8:25 AM Anton Okolnychyi
>>>>> <aokolnyc...@apple.com.invalid> wrote:
>>>>>
>>>>>> Hi everyone,
>>>>>>
>>>>>> I propose the following RC to be released as official Apache Iceberg
>>>>>> 0.10.0 release.
>>>>>>
>>>>>> The commit id is d39fad00b7dded98121368309f381473ec21e85f
>>>>>> * This corresponds to the tag: apache-iceberg-0.10.0-rc4
>>>>>> * https://github.com/apache/iceberg/commits/apache-iceberg-0.10.0-rc4
>>>>>> *
>>>>>> https://github.com/apache/iceberg/tree/d39fad00b7dded98121368309f381473ec21e85f
>>>>>>
>>>>>> The release tarball, signature, and checksums are here:
>>>>>> *
>>>>>> https://dist.apache.org/repos/dist/dev/iceberg/apache-iceberg-0.10.0-rc4/
>>>>>>
>>>>>> You can find the KEYS file here (make sure to import the new key that
>>>>>> was used to sign the release):
>>>>>> * https://dist.apache.org/repos/dist/dev/iceberg/KEYS
>>>>>>
>>>>>> Convenience binary artifacts are staged in Nexus. The Maven
>>>>>> repository URL is:
>>>>>> *
>>>>>> https://repository.apache.org/content/repositories/orgapacheiceberg-1012
>>>>>>
>>>>>> This release includes important changes:
>>>>>>
>>>>>> * Flink support
>>>>>> * Hive read support
>>>>>> * ORC support fixes and improvements
>>>>>> * Application of row-level delete files on read
>>>>>> * Snapshot partition summary
>>>>>> * Ability to load LocationProvider dynamically
>>>>>> * Sort spec
>>>>>>
>>>>>> Please download, verify, and test.
>>>>>>
>>>>>> Please vote in the next 72 hours.
>>>>>>
>>>>>> [ ] +1 Release this as Apache Iceberg 0.10.0
>>>>>> [ ] +0
>>>>>> [ ] -1 Do not release this because…
>>>>>>
>>>>>> Thanks,
>>>>>> Anton
>>>>>>
>>>>>
>>>
>>> --
>>> Ryan Blue
>>> Software Engineer
>>> Netflix
>>>
>>
>>
>> --
>> Best, Jingsong Lee
>>
>

Reply via email to