+1 (non-binding) I verified the build and ran the tests. Also verified both flink and spark custom catalogs are working.
One side note: I had to run the tests a few times to get the build to pass. Flaky Hive tests. Best, Ryan Murray On Sat, Jan 23, 2021 at 12:26 AM Jack Ye <yezhao...@gmail.com> wrote: > Hi everyone, > > I propose the following RC to be released as the official Apache Iceberg > 0.11.0 release. The RC is also reviewed and signed by Ryan Blue. > > The commit id is ad78cc6cf259b7a0c66ab5de6675cc005febd939 > > This corresponds to the tag: apache-iceberg-0.11.0-rc0 > * https://github.com/apache/iceberg/commits/apache-iceberg-0.11.0-rc0 > * https://github.com/apache/iceberg/tree/apache-iceberg-0.11.0-rc0 > > The release tarball, signature, and checksums are here: > * https://dist.apache.org/repos/dist/dev/iceberg/apache-iceberg-0.11.0-rc0 > > You can find the KEYS file here: > * https://dist.apache.org/repos/dist/dev/iceberg/KEYS > > Convenience binary artifacts are staged in Nexus. The Maven repository URL > is: > * https://repository.apache.org/content/repositories/orgapacheiceberg-1015 > > This release includes the following changes: > > *High-level features* > > - Core API now supports partition spec and sort order evolution > - Spark 3 now supports the following SQL extensions: > - MERGE INTO > - DELETE FROM > - ALTER TABLE ... ADD/DROP PARTITION > - ALTER TABLE ... WRITE ORDERED BY > - invoke stored procedures using CALL > - Flink now supports streaming reads, CDC writes (experimental), and > filter pushdown > - AWS module is added to support better integration with AWS, with AWS > Glue catalog <https://aws.amazon.com/glue> support and dedicated S3 > FileIO implementation > - Nessie module is added to support integration with project Nessie > <https://projectnessie.org> > > *Important bug fixes* > > - #1981 fixes date and timestamp transforms > - #2091 fixes Parquet vectorized reads when column types are promoted > - #1962 fixes Parquet vectorized position reader > - #1991 fixes Avro schema conversions to preserve field docs > - #1811 makes refreshing Spark cache optional > - #1798 fixes read failure when encountering duplicate entries of data > files > - #1785 fixes invalidation of metadata tables in CachingCatalog > - #1784 fixes resolving of SparkSession table's metadata tables > > *Other notable changes* > > - NaN counter is added to format v2 metrics > - Shared catalog properties are added in core library to standardize > catalog level configurations > - Spark and Flink now supports dynamically loading customized > `Catalog` and `FileIO` implementations > - Spark now supports loading tables with file paths via HadoopTables > - Spark 2 now supports loading tables from other catalogs, like Spark 3 > - Spark 3 now supports catalog names in DataFrameReader when using > Iceberg as a format > - Hive now supports INSERT INTO, case insensitive query, projection > pushdown, create DDL with schema and auto type conversion > - ORC now supports reading tinyint, smallint, char, varchar types > - Hadoop catalog now supports role-based access of table listing > > Please download, verify, and test. > > Please vote in the next 72 hours. > > [ ] +1 Release this as Apache Iceberg 0.11.0 > [ ] +0 > [ ] -1 Do not release this because... >