+1 (Binding)

Tested against hbase-1.3 and hbase-1.6

* Build from source (mvn clean install -DskipTests
-Dhbase.profile=1.3/1.6): OK
* Green build: OK (thanks for triggering this Viraj)
* Did some basic DDL, queries, upserts, deletes and everything looked fine:
OK
* Did some upgrade testing: Create tables, views, indices from an old
client, query, upsert. Then upgrade to 4.16 metadata, query, upsert from an
old client, then upgrade the client and query, upsert from a new client: OK
* Verified checksums: OK
* Verified signatures: OK
* mvn clean apache-rat:check: OK

On Sun, Feb 7, 2021 at 10:03 PM Viraj Jasani <vjas...@apache.org> wrote:

> +1 (non-binding)
>
> Clean build:
> https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/4.16/29/
>
> Tested against HBase-1.6 profile:
>
> * Checksum : ok
> * Rat check (1.8.0_171): ok
>  - mvn clean apache-rat:check
> * Built from source (1.8.0_171): ok
>  - mvn clean install  -DskipTests
> * Basic testing with mini cluster: ok
> * Unit tests pass (1.8.0_171): failed (passing eventually)
>  - mvn clean package  && mvn verify  -Dskip.embedded
>
>
> [ERROR] Tests run: 23, Failures: 0, Errors: 1, Skipped: 0, Time elapsed:
> 197.428 s <<< FAILURE! - in org.apache.phoenix.end2end.AggregateIT
> [ERROR]
> testOrderByOptimizeForClientAggregatePlanBug4820(org.apache.phoenix.end2end.AggregateIT)
> Time elapsed: 9.055 s  <<< ERROR!
> java.lang.RuntimeException: java.lang.OutOfMemoryError: unable to create
> new native thread
>         at
> org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithoutRetries(RpcRetryingCaller.java:239)
>         at
> org.apache.hadoop.hbase.client.ClientScanner.call(ClientScanner.java:273)
>         at
> org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:434)
>         at
> org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:308)
>
>
> [ERROR] Tests run: 37, Failures: 0, Errors: 1, Skipped: 0, Time elapsed:
> 204.243 s <<< FAILURE! - in org.apache.phoenix.end2end.ArrayAppendFunctionIT
> [ERROR]
> testUpsertArrayAppendFunctionVarchar(org.apache.phoenix.end2end.ArrayAppendFunctionIT)
> Time elapsed: 4.286 s  <<< ERROR!
> org.apache.phoenix.exception.PhoenixIOException:
> org.apache.hadoop.hbase.DoNotRetryIOException: N000065:
> java.lang.RuntimeException: java.lang.OutOfMemoryError: unable to create
> new native thread
>         at
> org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:122)
>         at
> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:2151)
>         at
> org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:17317)
>
>
> [ERROR] Tests run: 28, Failures: 0, Errors: 1, Skipped: 0, Time elapsed:
> 147.854 s <<< FAILURE! - in org.apache.phoenix.end2end.ArrayRemoveFunctionIT
> [ERROR]
> testArrayRemoveFunctionWithNull(org.apache.phoenix.end2end.ArrayRemoveFunctionIT)
> Time elapsed: 2.519 s  <<< ERROR!
> org.apache.phoenix.exception.PhoenixIOException:
> java.util.concurrent.ExecutionException: java.lang.OutOfMemoryError: unable
> to create new native thread
>         at
> org.apache.phoenix.util.ServerUtil.parseServerException(ServerUtil.java:146)
>         at
> org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:1511)
>         at
> org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1901)
>         at
> org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:3065)
>
>
> [ERROR] Tests run: 10, Failures: 0, Errors: 1, Skipped: 0, Time elapsed:
> 5,069.234 s <<< FAILURE! - in
> org.apache.phoenix.end2end.PermissionNSDisabledWithCustomAccessControllerIT
> [ERROR]
> testAutomaticGrantWithIndexAndView(org.apache.phoenix.end2end.PermissionNSDisabledWithCustomAccessControllerIT)
> Time elapsed: 2,572.586 s  <<< ERROR!
> java.lang.reflect.UndeclaredThrowableException
>         at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1862)
>         at
> org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:340)
>         at
> org.apache.phoenix.end2end.BasePermissionsIT.verifyAllowed(BasePermissionsIT.java:776)
>         at
> org.apache.phoenix.end2end.BasePermissionsIT.verifyAllowed(BasePermissionsIT.java:769)
>
> Tests are passing in subsequent runs.
>
>
> On 2021/02/06 04:53:48, Xinyi Yan <yanxi...@apache.org> wrote:
> > Hello Everyone,
> >
> > This is a call for a vote on Apache Phoenix 4.16.0 RC2. This is the next
> > minor release of Phoenix 4, compatible with Apache HBase 1.3, 1.4, 1.5
> > and 1.6.
> >
> > The VOTE will remain open for at least 72 hours.
> >
> > [ ] +1 Release this package as Apache phoenix 4.16.0
> > [ ] -1 Do not release this package because ...
> >
> > The tag to be voted on is 4.16.0RC2
> > https://github.com/apache/phoenix/tree/4.16.0RC2
> >
> > The release files, including signatures, digests, as well as CHANGES.md
> > and RELEASENOTES.md included in this RC can be found at:
> > https://dist.apache.org/repos/dist/dev/phoenix/phoenix-4.16.0RC2/
> >
> > For a complete list of changes, see:
> >
> https://dist.apache.org/repos/dist/dev/phoenix/phoenix-4.16.0RC2/CHANGES.md
> >
> > Artifacts are signed with my "CODE SIGNING KEY":
> > E4882DD3AB711587
> >
> > KEYS file available here:
> > https://dist.apache.org/repos/dist/dev/phoenix/KEYS
> >
> >
> > Thanks,
> > Xinyi
> >
> > <
> https://dist.apache.org/repos/dist/dev/phoenix/phoenix-4.16.0RC1/CHANGES.md
> >
> >
>


-- 
Chinmay Kulkarni

Reply via email to