See 
<https://builds.apache.org/job/Phoenix-4.x-HBase-1.2/634/display/redirect?page=changes>

Changes:

[ankitsinghal59] PHOENIX-5178 SYSTEM schema is not getting cached at MetaData 
server

------------------------------------------
[...truncated 1.03 MB...]
[ERROR] 
testCreateViewMappedToExistingHbaseTableWithNamespaceMappingEnabled[ViewIT_transactionProvider=null,
 columnEncoded=true](org.apache.phoenix.end2end.ViewIT)  Time elapsed: 0.042 s  
<<< ERROR!
org.apache.phoenix.exception.PhoenixIOException: 
org.apache.hadoop.hbase.DoNotRetryIOException: NS_N000460: Timestamp cannot be 
negative. minStamp:0, maxStamp:-9223372036854775808
        at 
org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:120)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createSchema(MetaDataEndpointImpl.java:4730)
        at 
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:17063)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:7873)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2008)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:1990)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33652)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2188)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
        at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
        at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.IllegalArgumentException: Timestamp cannot be negative. 
minStamp:0, maxStamp:-9223372036854775808
        at org.apache.hadoop.hbase.io.TimeRange.<init>(TimeRange.java:84)
        at org.apache.hadoop.hbase.client.Scan.setTimeRange(Scan.java:331)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.buildSchemas(MetaDataEndpointImpl.java:1088)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.loadSchema(MetaDataEndpointImpl.java:1957)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createSchema(MetaDataEndpointImpl.java:4690)
        ... 10 more

        at 
org.apache.phoenix.end2end.ViewIT.testCreateViewMappedToExistingHbaseTableWithNamespaceMappingEnabled(ViewIT.java:916)
Caused by: org.apache.hadoop.hbase.DoNotRetryIOException: 
org.apache.hadoop.hbase.DoNotRetryIOException: NS_N000460: Timestamp cannot be 
negative. minStamp:0, maxStamp:-9223372036854775808
        at 
org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:120)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createSchema(MetaDataEndpointImpl.java:4730)
        at 
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:17063)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:7873)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2008)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:1990)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33652)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2188)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
        at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
        at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.IllegalArgumentException: Timestamp cannot be negative. 
minStamp:0, maxStamp:-9223372036854775808
        at org.apache.hadoop.hbase.io.TimeRange.<init>(TimeRange.java:84)
        at org.apache.hadoop.hbase.client.Scan.setTimeRange(Scan.java:331)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.buildSchemas(MetaDataEndpointImpl.java:1088)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.loadSchema(MetaDataEndpointImpl.java:1957)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createSchema(MetaDataEndpointImpl.java:4690)
        ... 10 more

Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException: 
org.apache.hadoop.hbase.DoNotRetryIOException: NS_N000460: Timestamp cannot be 
negative. minStamp:0, maxStamp:-9223372036854775808
        at 
org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:120)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createSchema(MetaDataEndpointImpl.java:4730)
        at 
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:17063)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:7873)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2008)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:1990)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33652)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2188)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
        at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
        at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.IllegalArgumentException: Timestamp cannot be negative. 
minStamp:0, maxStamp:-9223372036854775808
        at org.apache.hadoop.hbase.io.TimeRange.<init>(TimeRange.java:84)
        at org.apache.hadoop.hbase.client.Scan.setTimeRange(Scan.java:331)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.buildSchemas(MetaDataEndpointImpl.java:1088)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.loadSchema(MetaDataEndpointImpl.java:1957)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createSchema(MetaDataEndpointImpl.java:4690)
        ... 10 more


[ERROR] 
testViewAndTableInDifferentSchemasWithNamespaceMappingEnabled[ViewIT_transactionProvider=null,
 columnEncoded=true](org.apache.phoenix.end2end.ViewIT)  Time elapsed: 0.011 s  
<<< ERROR!
org.apache.phoenix.exception.PhoenixIOException: 
org.apache.hadoop.hbase.DoNotRetryIOException: SCHEMA1: Timestamp cannot be 
negative. minStamp:0, maxStamp:-9223372036854775808
        at 
org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:120)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createSchema(MetaDataEndpointImpl.java:4730)
        at 
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:17063)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:7873)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2008)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:1990)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33652)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2188)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
        at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
        at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.IllegalArgumentException: Timestamp cannot be negative. 
minStamp:0, maxStamp:-9223372036854775808
        at org.apache.hadoop.hbase.io.TimeRange.<init>(TimeRange.java:84)
        at org.apache.hadoop.hbase.client.Scan.setTimeRange(Scan.java:331)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.buildSchemas(MetaDataEndpointImpl.java:1088)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.loadSchema(MetaDataEndpointImpl.java:1957)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createSchema(MetaDataEndpointImpl.java:4690)
        ... 10 more

        at 
org.apache.phoenix.end2end.ViewIT.testViewAndTableInDifferentSchemas(ViewIT.java:405)
        at 
org.apache.phoenix.end2end.ViewIT.testViewAndTableInDifferentSchemasWithNamespaceMappingEnabled(ViewIT.java:382)
Caused by: org.apache.hadoop.hbase.DoNotRetryIOException: 
org.apache.hadoop.hbase.DoNotRetryIOException: SCHEMA1: Timestamp cannot be 
negative. minStamp:0, maxStamp:-9223372036854775808
        at 
org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:120)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createSchema(MetaDataEndpointImpl.java:4730)
        at 
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:17063)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:7873)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2008)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:1990)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33652)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2188)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
        at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
        at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.IllegalArgumentException: Timestamp cannot be negative. 
minStamp:0, maxStamp:-9223372036854775808
        at org.apache.hadoop.hbase.io.TimeRange.<init>(TimeRange.java:84)
        at org.apache.hadoop.hbase.client.Scan.setTimeRange(Scan.java:331)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.buildSchemas(MetaDataEndpointImpl.java:1088)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.loadSchema(MetaDataEndpointImpl.java:1957)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createSchema(MetaDataEndpointImpl.java:4690)
        ... 10 more

Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException: 
org.apache.hadoop.hbase.DoNotRetryIOException: SCHEMA1: Timestamp cannot be 
negative. minStamp:0, maxStamp:-9223372036854775808
        at 
org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:120)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createSchema(MetaDataEndpointImpl.java:4730)
        at 
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:17063)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:7873)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2008)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:1990)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33652)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2188)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
        at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
        at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.IllegalArgumentException: Timestamp cannot be negative. 
minStamp:0, maxStamp:-9223372036854775808
        at org.apache.hadoop.hbase.io.TimeRange.<init>(TimeRange.java:84)
        at org.apache.hadoop.hbase.client.Scan.setTimeRange(Scan.java:331)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.buildSchemas(MetaDataEndpointImpl.java:1088)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.loadSchema(MetaDataEndpointImpl.java:1957)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createSchema(MetaDataEndpointImpl.java:4690)
        ... 10 more


[INFO] Tests run: 112, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 
1,158.327 s - in org.apache.phoenix.end2end.AlterTableWithViewsIT
[INFO] 
[INFO] Results:
[INFO] 
[ERROR] Errors: 
[ERROR]   
TenantSpecificViewIndexIT.testMultiCFViewIndexWithNamespaceMapping:86->testMultiCFViewIndex:124->createTableAndValidate:102
 » PhoenixIO
[ERROR]   
ViewIT.testCreateViewMappedToExistingHbaseTableWithNamespaceMappingEnabled:916 
» PhoenixIO
[ERROR]   
ViewIT.testCreateViewMappedToExistingHbaseTableWithNamespaceMappingEnabled:916 
» PhoenixIO
[ERROR]   
ViewIT.testCreateViewMappedToExistingHbaseTableWithNamespaceMappingEnabled:916 
» PhoenixIO
[ERROR]   
ViewIT.testCreateViewMappedToExistingHbaseTableWithNamespaceMappingEnabled:916 
» PhoenixIO
[ERROR]   
ViewIT.testCreateViewMappedToExistingHbaseTableWithNamespaceMappingEnabled:916 
» PhoenixIO
[ERROR]   
ViewIT.testViewAndTableInDifferentSchemasWithNamespaceMappingEnabled:382->testViewAndTableInDifferentSchemas:405
 » PhoenixIO
[ERROR]   
ViewIT.testViewAndTableInDifferentSchemasWithNamespaceMappingEnabled:382->testViewAndTableInDifferentSchemas:405
 » PhoenixIO
[ERROR]   
ViewIT.testViewAndTableInDifferentSchemasWithNamespaceMappingEnabled:382->testViewAndTableInDifferentSchemas:405
 » PhoenixIO
[ERROR]   
ViewIT.testViewAndTableInDifferentSchemasWithNamespaceMappingEnabled:382->testViewAndTableInDifferentSchemas:405
 » PhoenixIO
[ERROR]   
ViewIT.testViewAndTableInDifferentSchemasWithNamespaceMappingEnabled:382->testViewAndTableInDifferentSchemas:405
 » PhoenixIO
[ERROR]   
ViewIndexIT.testCreatingIndexOnViewBuiltOnTableWithOnlyNamedColumnFamilies:513 
» PhoenixIO
[ERROR]   ViewIndexIT.testDeleteViewIndexSequences:148->createBaseTable:77 » 
PhoenixIO o...
[ERROR]   
ViewIndexIT.testGlobalAndTenantViewIndexesHaveDifferentIndexIds:559->createBaseTable:77
 » PhoenixIO
[ERROR]   ViewIndexIT.testMultiTenantViewLocalIndex:178->createBaseTable:77 » 
PhoenixIO ...
[ERROR]   ViewIndexIT.testUpdateOnTenantViewWithGlobalView:388 » PhoenixIO 
org.apache.ha...
[INFO] 
[ERROR] Tests run: 320, Failures: 0, Errors: 16, Skipped: 2
[INFO] 
[INFO] 
[INFO] --- maven-failsafe-plugin:2.20:verify (ParallelStatsEnabledTest) @ 
phoenix-core ---
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary for Apache Phoenix 4.15.0-HBase-1.2-SNAPSHOT:
[INFO] 
[INFO] Apache Phoenix ..................................... SUCCESS [  2.596 s]
[INFO] Phoenix Core ....................................... FAILURE [  03:02 h]
[INFO] Phoenix - Flume .................................... SKIPPED
[INFO] Phoenix - Kafka .................................... SKIPPED
[INFO] Phoenix - Pig ...................................... SKIPPED
[INFO] Phoenix - Pherf .................................... SKIPPED
[INFO] Phoenix - Spark .................................... SKIPPED
[INFO] Phoenix - Hive ..................................... SKIPPED
[INFO] Phoenix Client ..................................... SKIPPED
[INFO] Phoenix Server ..................................... SKIPPED
[INFO] Phoenix Assembly ................................... SKIPPED
[INFO] Phoenix - Tracing Web Application .................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time:  03:02 h
[INFO] Finished at: 2019-03-12T03:18:20Z
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal 
org.apache.maven.plugins:maven-failsafe-plugin:2.20:verify 
(ParallelStatsEnabledTest) on project phoenix-core: There are test failures.
[ERROR] 
[ERROR] Please refer to 
<https://builds.apache.org/job/Phoenix-4.x-HBase-1.2/ws/phoenix-core/target/failsafe-reports>
 for the individual test results.
[ERROR] Please refer to dump files (if any exist) [date]-jvmRun[N].dump, 
[date].dumpstream and [date]-jvmRun[N].dumpstream.
[ERROR] org.apache.maven.surefire.booter.SurefireBooterForkException: 
ExecutionException The forked VM terminated without properly saying goodbye. VM 
crash or System.exit called?
[ERROR] Command was /bin/sh -c cd 
<https://builds.apache.org/job/Phoenix-4.x-HBase-1.2/ws/phoenix-core> && 
/usr/local/asfpackages/java/jdk1.8.0_191/jre/bin/java -enableassertions 
-Xmx2000m -XX:MaxPermSize=256m -Djava.security.egd=file:/dev/./urandom 
'-Djava.library.path=${hadoop.library.path}:/usr/java/packages/lib/amd64:/usr/lib64:/lib64:/lib:/usr/lib'
 -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=./target/ -jar 
<https://builds.apache.org/job/Phoenix-4.x-HBase-1.2/634/artifact/phoenix-core/target/surefire/surefirebooter4614620807726568700.jar>
 
<https://builds.apache.org/job/Phoenix-4.x-HBase-1.2/ws/phoenix-core/target/surefire>
 2019-03-12T00-17-43_009-jvmRun4 surefire5574660628718352477tmp 
surefire_1251629939449490682556tmp
[ERROR] Process Exit Code: 0
[ERROR]         at 
org.apache.maven.plugin.surefire.booterclient.ForkStarter.awaitResultsDone(ForkStarter.java:494)
[ERROR]         at 
org.apache.maven.plugin.surefire.booterclient.ForkStarter.runSuitesForkPerTestSet(ForkStarter.java:441)
[ERROR]         at 
org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:292)
[ERROR]         at 
org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:243)
[ERROR]         at 
org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1077)
[ERROR]         at 
org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:907)
[ERROR]         at 
org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:785)
[ERROR]         at 
org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)
[ERROR]         at 
org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:210)
[ERROR]         at 
org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:156)
[ERROR]         at 
org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:148)
[ERROR]         at 
org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)
[ERROR]         at 
org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:81)
[ERROR]         at 
org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build(SingleThreadedBuilder.java:56)
[ERROR]         at 
org.apache.maven.lifecycle.internal.LifecycleStarter.execute(LifecycleStarter.java:128)
[ERROR]         at 
org.apache.maven.DefaultMaven.doExecute(DefaultMaven.java:305)
[ERROR]         at 
org.apache.maven.DefaultMaven.doExecute(DefaultMaven.java:192)
[ERROR]         at org.apache.maven.DefaultMaven.execute(DefaultMaven.java:105)
[ERROR]         at org.apache.maven.cli.MavenCli.execute(MavenCli.java:956)
[ERROR]         at org.apache.maven.cli.MavenCli.doMain(MavenCli.java:288)
[ERROR]         at org.apache.maven.cli.MavenCli.main(MavenCli.java:192)
[ERROR]         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
[ERROR]         at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
[ERROR]         at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
[ERROR]         at java.lang.reflect.Method.invoke(Method.java:498)
[ERROR]         at 
org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced(Launcher.java:289)
[ERROR]         at 
org.codehaus.plexus.classworlds.launcher.Launcher.launch(Launcher.java:229)
[ERROR]         at 
org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode(Launcher.java:415)
[ERROR]         at 
org.codehaus.plexus.classworlds.launcher.Launcher.main(Launcher.java:356)
[ERROR] Caused by: 
org.apache.maven.surefire.booter.SurefireBooterForkException: The forked VM 
terminated without properly saying goodbye. VM crash or System.exit called?
[ERROR] Command was /bin/sh -c cd 
<https://builds.apache.org/job/Phoenix-4.x-HBase-1.2/ws/phoenix-core> && 
/usr/local/asfpackages/java/jdk1.8.0_191/jre/bin/java -enableassertions 
-Xmx2000m -XX:MaxPermSize=256m -Djava.security.egd=file:/dev/./urandom 
'-Djava.library.path=${hadoop.library.path}:/usr/java/packages/lib/amd64:/usr/lib64:/lib64:/lib:/usr/lib'
 -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=./target/ -jar 
<https://builds.apache.org/job/Phoenix-4.x-HBase-1.2/634/artifact/phoenix-core/target/surefire/surefirebooter4614620807726568700.jar>
 
<https://builds.apache.org/job/Phoenix-4.x-HBase-1.2/ws/phoenix-core/target/surefire>
 2019-03-12T00-17-43_009-jvmRun4 surefire5574660628718352477tmp 
surefire_1251629939449490682556tmp
[ERROR] Process Exit Code: 0
[ERROR]         at 
org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:679)
[ERROR]         at 
org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:533)
[ERROR]         at 
org.apache.maven.plugin.surefire.booterclient.ForkStarter.access$600(ForkStarter.java:117)
[ERROR]         at 
org.apache.maven.plugin.surefire.booterclient.ForkStarter$2.call(ForkStarter.java:429)
[ERROR]         at 
org.apache.maven.plugin.surefire.booterclient.ForkStarter$2.call(ForkStarter.java:406)
[ERROR]         at java.util.concurrent.FutureTask.run(FutureTask.java:266)
[ERROR]         at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
[ERROR]         at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
[ERROR]         at java.lang.Thread.run(Thread.java:748)
[ERROR] 
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please 
read the following articles:
[ERROR] [Help 1] 
http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :phoenix-core
Build step 'Invoke top-level Maven targets' marked build as failure
Archiving artifacts
Recording test results
Not sending mail to unregistered user [email protected]

Reply via email to