I agree. I saw similar errors while testing this afternoon. Will do more testing tomorrow but I think it's ok...
On Sun, 24 Jan 2021, 22:01 Ryan Blue, <rb...@netflix.com.invalid> wrote: > Looks like the problem in #2144 is that the error message isn't quite > right. The problem was that `hadoop` was used for the Catalog > implementation rather than the type, which is a short-cut to not need to > specify the full class name. The error message should be better and note > that the class was actually not found, instead of the no-arg constructor. > But I don't think that's a problem that needs to be fixed for the release. > > On Sun, Jan 24, 2021 at 12:07 PM Ryan Murray <rym...@dremio.com> wrote: > >> Based on #2144 <https://github.com/apache/iceberg/issues/2144> I would >> change my vote to -1. I will raise a fix asap. >> >> >> On Sun, Jan 24, 2021 at 3:12 PM Ryan Murray <rym...@dremio.com> wrote: >> >>> >>> +1 (non-binding) >>> >>> I verified the build and ran the tests. Also verified both flink and >>> spark custom catalogs are working. >>> >>> One side note: I had to run the tests a few times to get the build to >>> pass. Flaky Hive tests. >>> >>> Best, >>> >>> Ryan Murray >>> >>> On Sat, Jan 23, 2021 at 12:26 AM Jack Ye <yezhao...@gmail.com> wrote: >>> >>>> Hi everyone, >>>> >>>> I propose the following RC to be released as the official Apache >>>> Iceberg 0.11.0 release. The RC is also reviewed and signed by Ryan Blue. >>>> >>>> The commit id is ad78cc6cf259b7a0c66ab5de6675cc005febd939 >>>> >>>> This corresponds to the tag: apache-iceberg-0.11.0-rc0 >>>> * https://github.com/apache/iceberg/commits/apache-iceberg-0.11.0-rc0 >>>> * https://github.com/apache/iceberg/tree/apache-iceberg-0.11.0-rc0 >>>> >>>> The release tarball, signature, and checksums are here: >>>> * >>>> https://dist.apache.org/repos/dist/dev/iceberg/apache-iceberg-0.11.0-rc0 >>>> >>>> You can find the KEYS file here: >>>> * https://dist.apache.org/repos/dist/dev/iceberg/KEYS >>>> >>>> Convenience binary artifacts are staged in Nexus. The Maven repository >>>> URL is: >>>> * >>>> https://repository.apache.org/content/repositories/orgapacheiceberg-1015 >>>> >>>> This release includes the following changes: >>>> >>>> *High-level features* >>>> >>>> - Core API now supports partition spec and sort order evolution >>>> - Spark 3 now supports the following SQL extensions: >>>> - MERGE INTO >>>> - DELETE FROM >>>> - ALTER TABLE ... ADD/DROP PARTITION >>>> - ALTER TABLE ... WRITE ORDERED BY >>>> - invoke stored procedures using CALL >>>> - Flink now supports streaming reads, CDC writes (experimental), >>>> and filter pushdown >>>> - AWS module is added to support better integration with AWS, with AWS >>>> Glue catalog <https://aws.amazon.com/glue> support and dedicated S3 >>>> FileIO implementation >>>> - Nessie module is added to support integration with project Nessie >>>> <https://projectnessie.org> >>>> >>>> *Important bug fixes* >>>> >>>> - #1981 fixes date and timestamp transforms >>>> - #2091 fixes Parquet vectorized reads when column types are >>>> promoted >>>> - #1962 fixes Parquet vectorized position reader >>>> - #1991 fixes Avro schema conversions to preserve field docs >>>> - #1811 makes refreshing Spark cache optional >>>> - #1798 fixes read failure when encountering duplicate entries of >>>> data files >>>> - #1785 fixes invalidation of metadata tables in CachingCatalog >>>> - #1784 fixes resolving of SparkSession table's metadata tables >>>> >>>> *Other notable changes* >>>> >>>> - NaN counter is added to format v2 metrics >>>> - Shared catalog properties are added in core library to >>>> standardize catalog level configurations >>>> - Spark and Flink now supports dynamically loading customized >>>> `Catalog` and `FileIO` implementations >>>> - Spark now supports loading tables with file paths via HadoopTables >>>> - Spark 2 now supports loading tables from other catalogs, like >>>> Spark 3 >>>> - Spark 3 now supports catalog names in DataFrameReader when using >>>> Iceberg as a format >>>> - Hive now supports INSERT INTO, case insensitive query, projection >>>> pushdown, create DDL with schema and auto type conversion >>>> - ORC now supports reading tinyint, smallint, char, varchar types >>>> - Hadoop catalog now supports role-based access of table listing >>>> >>>> Please download, verify, and test. >>>> >>>> Please vote in the next 72 hours. >>>> >>>> [ ] +1 Release this as Apache Iceberg 0.11.0 >>>> [ ] +0 >>>> [ ] -1 Do not release this because... >>>> >>> > > -- > Ryan Blue > Software Engineer > Netflix >