[ https://issues.apache.org/jira/browse/FLINK-17576?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Dian Fu updated FLINK-17576: ---------------------------- Labels: test-stability (was: ) > HiveTableSinkTest and TableEnvHiveConnectorTest are instable > ------------------------------------------------------------ > > Key: FLINK-17576 > URL: https://issues.apache.org/jira/browse/FLINK-17576 > Project: Flink > Issue Type: Bug > Components: Connectors / Hive, Tests > Affects Versions: 1.11.0 > Reporter: Dian Fu > Priority: Major > Labels: test-stability > > HiveTableSinkTest and TableEnvHiveConnectorTest failed with the following > exception: > {code:java} > 2020-05-08T09:38:44.5916441Z [ERROR] > testWriteComplexType(org.apache.flink.connectors.hive.HiveTableSinkTest) > Time elapsed: 1.362 s <<< ERROR! > 2020-05-08T09:38:44.5932270Z java.util.concurrent.ExecutionException: > org.apache.flink.runtime.messages.FlinkJobNotFoundException: Could not find > Flink job (e27d50c5a780264a576aa8a21a6dd6c6) > 2020-05-08T09:38:44.5938598Z at > java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357) > 2020-05-08T09:38:44.5939435Z at > java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908) > 2020-05-08T09:38:44.5939970Z at > org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:1663) > 2020-05-08T09:38:44.5940551Z at > org.apache.flink.streaming.api.environment.LocalStreamEnvironment.execute(LocalStreamEnvironment.java:74) > 2020-05-08T09:38:44.5941188Z at > org.apache.flink.table.planner.delegation.ExecutorBase.execute(ExecutorBase.java:52) > 2020-05-08T09:38:44.5941834Z at > org.apache.flink.table.api.internal.TableEnvironmentImpl.execute(TableEnvironmentImpl.java:916) > 2020-05-08T09:38:44.5945405Z at > org.apache.flink.connectors.hive.HiveTableSinkTest.testWriteComplexType(HiveTableSinkTest.java:143) > 2020-05-08T09:38:44.5946105Z at > sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > 2020-05-08T09:38:44.5946628Z at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > 2020-05-08T09:38:44.5947106Z at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > 2020-05-08T09:38:44.5947770Z at > java.lang.reflect.Method.invoke(Method.java:498) > 2020-05-08T09:38:44.5948393Z at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) > 2020-05-08T09:38:44.5949102Z at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) > 2020-05-08T09:38:44.5949853Z at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) > 2020-05-08T09:38:44.5950587Z at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) > 2020-05-08T09:38:44.5951763Z at > org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runTestMethod(FlinkStandaloneHiveRunner.java:169) > 2020-05-08T09:38:44.5952660Z at > org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runChild(FlinkStandaloneHiveRunner.java:154) > 2020-05-08T09:38:44.5953829Z at > org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runChild(FlinkStandaloneHiveRunner.java:92) > 2020-05-08T09:38:44.5966233Z at > org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) > 2020-05-08T09:38:44.5967051Z at > org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71) > 2020-05-08T09:38:44.5968062Z at > org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) > 2020-05-08T09:38:44.5968949Z at > org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) > 2020-05-08T09:38:44.5969824Z at > org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) > 2020-05-08T09:38:44.5970751Z at > org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) > 2020-05-08T09:38:44.5971584Z at > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) > 2020-05-08T09:38:44.5972386Z at > org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48) > 2020-05-08T09:38:44.5973469Z at > org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48) > 2020-05-08T09:38:44.5974147Z at > org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48) > 2020-05-08T09:38:44.5974684Z at > org.junit.rules.RunRules.evaluate(RunRules.java:20) > 2020-05-08T09:38:44.5975227Z at > org.junit.runners.ParentRunner.run(ParentRunner.java:363) > 2020-05-08T09:38:44.5975827Z at > org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365) > 2020-05-08T09:38:44.5976533Z at > org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273) > 2020-05-08T09:38:44.5977627Z at > org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238) > 2020-05-08T09:38:44.5978552Z at > org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159) > 2020-05-08T09:38:44.5979527Z at > org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384) > 2020-05-08T09:38:44.5980556Z at > org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345) > 2020-05-08T09:38:44.5981696Z at > org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126) > 2020-05-08T09:38:44.5982559Z at > org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418) > 2020-05-08T09:38:44.5983422Z Caused by: > org.apache.flink.runtime.messages.FlinkJobNotFoundException: Could not find > Flink job (e27d50c5a780264a576aa8a21a6dd6c6) > 2020-05-08T09:38:44.5984215Z at > org.apache.flink.runtime.dispatcher.Dispatcher.requestJobResult(Dispatcher.java:563) > 2020-05-08T09:38:44.5984798Z at > sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > 2020-05-08T09:38:44.5985384Z at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > 2020-05-08T09:38:44.5986037Z at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > 2020-05-08T09:38:44.5986721Z at > java.lang.reflect.Method.invoke(Method.java:498) > 2020-05-08T09:38:44.5987383Z at > org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcInvocation(AkkaRpcActor.java:284) > 2020-05-08T09:38:44.5988199Z at > org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:199) > 2020-05-08T09:38:44.5989080Z at > org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:74) > 2020-05-08T09:38:44.5989914Z at > org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:152) > 2020-05-08T09:38:44.5990781Z at > akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26) > 2020-05-08T09:38:44.5991545Z at > akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21) > 2020-05-08T09:38:44.5992368Z at > scala.PartialFunction$class.applyOrElse(PartialFunction.scala:123) > 2020-05-08T09:38:44.5993122Z at > akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21) > 2020-05-08T09:38:44.5993826Z at > scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:170) > 2020-05-08T09:38:44.5994473Z at > scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171) > 2020-05-08T09:38:44.5995161Z at > scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171) > 2020-05-08T09:38:44.5995794Z at > akka.actor.Actor$class.aroundReceive(Actor.scala:517) > 2020-05-08T09:38:44.5996445Z at > akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225) > 2020-05-08T09:38:44.5997053Z at > akka.actor.ActorCell.receiveMessage(ActorCell.scala:592) > 2020-05-08T09:38:44.5997658Z at > akka.actor.ActorCell.invoke(ActorCell.scala:561) > 2020-05-08T09:38:44.5998227Z at > akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258) > 2020-05-08T09:38:44.5998805Z at akka.dispatch.Mailbox.run(Mailbox.scala:225) > 2020-05-08T09:38:44.5999332Z at akka.dispatch.Mailbox.exec(Mailbox.scala:235) > 2020-05-08T09:38:44.5999920Z at > akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) > 2020-05-08T09:38:44.6000703Z at > akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) > 2020-05-08T09:38:44.6001454Z at > akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) > 2020-05-08T09:38:44.6002040Z at > akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) > 2020-05-08T09:38:44.6002538Z > 2020-05-08T09:38:44.6003181Z [ERROR] > testInsertIntoNonPartitionTable(org.apache.flink.connectors.hive.HiveTableSinkTest) > Time elapsed: 0.05 s <<< ERROR! > 2020-05-08T09:38:44.6004907Z > org.apache.flink.table.catalog.exceptions.TableAlreadyExistException: Table > (or view) default.dest already exists in Catalog test-catalog. > 2020-05-08T09:38:44.6005671Z at > org.apache.flink.table.catalog.hive.HiveCatalog.createTable(HiveCatalog.java:381) > 2020-05-08T09:38:44.6006731Z at > org.apache.flink.connectors.hive.HiveTableSinkTest.createHiveDestTable(HiveTableSinkTest.java:221) > 2020-05-08T09:38:44.6007458Z at > org.apache.flink.connectors.hive.HiveTableSinkTest.createHiveDestTable(HiveTableSinkTest.java:233) > 2020-05-08T09:38:44.6008204Z at > org.apache.flink.connectors.hive.HiveTableSinkTest.testInsertIntoNonPartitionTable(HiveTableSinkTest.java:91) > 2020-05-08T09:38:44.6008828Z at > sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > 2020-05-08T09:38:44.6009371Z at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > 2020-05-08T09:38:44.6010558Z at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > 2020-05-08T09:38:44.6011208Z at > java.lang.reflect.Method.invoke(Method.java:498) > 2020-05-08T09:38:44.6011809Z at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) > 2020-05-08T09:38:44.6012473Z at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) > 2020-05-08T09:38:44.6013197Z at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) > 2020-05-08T09:38:44.6013851Z at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) > 2020-05-08T09:38:44.6014545Z at > org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runTestMethod(FlinkStandaloneHiveRunner.java:169) > 2020-05-08T09:38:44.6015457Z at > org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runChild(FlinkStandaloneHiveRunner.java:154) > 2020-05-08T09:38:44.6016216Z at > org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runChild(FlinkStandaloneHiveRunner.java:92) > 2020-05-08T09:38:44.6017023Z at > org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) > 2020-05-08T09:38:44.6017701Z at > org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71) > 2020-05-08T09:38:44.6018172Z at > org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) > 2020-05-08T09:38:44.6018703Z at > org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) > 2020-05-08T09:38:44.6019284Z at > org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) > 2020-05-08T09:38:44.6019938Z at > org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) > 2020-05-08T09:38:44.6020713Z at > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) > 2020-05-08T09:38:44.6021472Z at > org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48) > 2020-05-08T09:38:44.6022069Z at > org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48) > 2020-05-08T09:38:44.6022685Z at > org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48) > 2020-05-08T09:38:44.6023296Z at > org.junit.rules.RunRules.evaluate(RunRules.java:20) > 2020-05-08T09:38:44.6023904Z at > org.junit.runners.ParentRunner.run(ParentRunner.java:363) > 2020-05-08T09:38:44.6024527Z at > org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365) > 2020-05-08T09:38:44.6025235Z at > org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273) > 2020-05-08T09:38:44.6025945Z at > org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238) > 2020-05-08T09:38:44.6026745Z at > org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159) > 2020-05-08T09:38:44.6027625Z at > org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384) > 2020-05-08T09:38:44.6028430Z at > org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345) > 2020-05-08T09:38:44.6029179Z at > org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126) > 2020-05-08T09:38:44.6029875Z at > org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418) > 2020-05-08T09:38:44.6030663Z Caused by: AlreadyExistsException(message:Table > dest already exists) > 2020-05-08T09:38:44.6031755Z at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$create_table_with_environment_context_result$create_table_with_environment_context_resultStandardScheme.read(ThriftHiveMetastore.java:42052) > 2020-05-08T09:38:44.6033379Z at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$create_table_with_environment_context_result$create_table_with_environment_context_resultStandardScheme.read(ThriftHiveMetastore.java:42038) > 2020-05-08T09:38:44.6034566Z at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$create_table_with_environment_context_result.read(ThriftHiveMetastore.java:41964) > 2020-05-08T09:38:44.6035344Z at > org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:86) > 2020-05-08T09:38:44.6036352Z at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_create_table_with_environment_context(ThriftHiveMetastore.java:1199) > 2020-05-08T09:38:44.6037302Z at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.create_table_with_environment_context(ThriftHiveMetastore.java:1185) > 2020-05-08T09:38:44.6038089Z at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.create_table_with_environment_context(HiveMetaStoreClient.java:2399) > 2020-05-08T09:38:44.6038849Z at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:752) > 2020-05-08T09:38:44.6039507Z at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:740) > 2020-05-08T09:38:44.6040073Z at > sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > 2020-05-08T09:38:44.6040752Z at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > 2020-05-08T09:38:44.6041435Z at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > 2020-05-08T09:38:44.6041899Z at > java.lang.reflect.Method.invoke(Method.java:498) > 2020-05-08T09:38:44.6042490Z at > org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:169) > 2020-05-08T09:38:44.6042926Z at com.sun.proxy.$Proxy34.createTable(Unknown > Source) > 2020-05-08T09:38:44.6043476Z at > org.apache.flink.table.catalog.hive.client.HiveMetastoreClientWrapper.createTable(HiveMetastoreClientWrapper.java:142) > 2020-05-08T09:38:44.6044191Z at > org.apache.flink.table.catalog.hive.HiveCatalog.createTable(HiveCatalog.java:378) > 2020-05-08T09:38:44.6044593Z ... 34 more > 2020-05-08T09:38:44.6044759Z > 2020-05-08T09:38:45.4944094Z [INFO] Running > org.apache.flink.connectors.hive.TableEnvHiveConnectorTest > 2020-05-08T09:38:57.2455895Z OK > 2020-05-08T09:38:57.5074904Z OK > 2020-05-08T09:39:54.4485216Z WARNING: Hive-on-MR is deprecated in Hive 2 and > may not be available in the future versions. Consider using a different > execution engine (i.e. spark, tez) or using Hive 1.X releases. > 2020-05-08T09:39:54.4486112Z Query ID = > agent02_azpcontainer_20200508093954_5b7c9f45-4c3c-4baa-9192-cec3724fd063 > 2020-05-08T09:39:54.4486420Z Total jobs = 3 > 2020-05-08T09:39:54.4486631Z Launching Job 1 out of 3 > 2020-05-08T09:39:54.4493419Z Number of reduce tasks is set to 0 since there's > no reduce operator > 2020-05-08T09:39:54.6913698Z Job running in-process (local Hadoop) > 2020-05-08T09:39:54.6931724Z 2020-05-08 09:39:54,692 Stage-1 map = 0%, > reduce = 0% > 2020-05-08T09:39:54.7230751Z 2020-05-08 09:39:54,722 Stage-1 map = 100%, > reduce = 0% > 2020-05-08T09:39:54.7268752Z Ended Job = job_local1575498928_0003 > 2020-05-08T09:39:54.7282475Z Stage-3 is selected by condition resolver. > 2020-05-08T09:39:54.7283011Z Stage-2 is filtered out by condition resolver. > 2020-05-08T09:39:54.7283698Z Stage-4 is filtered out by condition resolver. > 2020-05-08T09:39:54.7293228Z Moving data to directory > file:/tmp/junit1597676277387724188/warehouse/db1.db/src1/.hive-staging_hive_2020-05-08_09-39-54_359_5169339369873513430-1/-ext-10000 > 2020-05-08T09:39:54.7619636Z Loading data to table db1.src1 > 2020-05-08T09:39:54.8385179Z MapReduce Jobs Launched: > 2020-05-08T09:39:54.8386671Z Stage-Stage-1: HDFS Read: 0 HDFS Write: 0 > SUCCESS > 2020-05-08T09:39:54.8387003Z Total MapReduce CPU Time Spent: 0 msec > 2020-05-08T09:39:54.8387337Z OK > 2020-05-08T09:39:56.4934855Z OK > 2020-05-08T09:39:56.5937839Z OK > 2020-05-08T09:39:58.5593808Z OK > 2020-05-08T09:39:58.6535469Z OK > 2020-05-08T09:39:58.7554994Z OK > 2020-05-08T09:39:58.7831835Z OK > 2020-05-08T09:39:58.8160236Z OK > 2020-05-08T09:39:58.8509517Z OK > 2020-05-08T09:39:58.8633482Z Loading data to table db1.src > 2020-05-08T09:39:58.9445381Z OK > 2020-05-08T09:40:02.0134976Z FAILED: Execution Error, return code 1 from > org.apache.hadoop.hive.ql.exec.DDLTask. Database db1 already exists > 2020-05-08T09:40:02.0813385Z FAILED: Execution Error, return code 1 from > org.apache.hadoop.hive.ql.exec.DDLTask. Database db1 already exists > 2020-05-08T09:40:02.1006755Z FAILED: Execution Error, return code 1 from > org.apache.hadoop.hive.ql.exec.DDLTask. Database db1 already exists > 2020-05-08T09:40:02.1427929Z OK > 2020-05-08T09:40:02.1877136Z [ERROR] Tests run: 21, Failures: 0, Errors: 4, > Skipped: 2, Time elapsed: 76.685 s <<< FAILURE! - in > org.apache.flink.connectors.hive.TableEnvHiveConnectorTest > 2020-05-08T09:40:02.1878103Z [ERROR] > testDefaultPartitionName(org.apache.flink.connectors.hive.TableEnvHiveConnectorTest) > Time elapsed: 3.191 s <<< ERROR! > 2020-05-08T09:40:02.1878675Z java.util.concurrent.ExecutionException: > org.apache.flink.runtime.client.JobExecutionException: Job execution failed. > 2020-05-08T09:40:02.1879172Z at > java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357) > 2020-05-08T09:40:02.1879747Z at > java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908) > 2020-05-08T09:40:02.1880286Z at > org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:1663) > 2020-05-08T09:40:02.1881369Z at > org.apache.flink.streaming.api.environment.LocalStreamEnvironment.execute(LocalStreamEnvironment.java:74) > 2020-05-08T09:40:02.1881929Z at > org.apache.flink.table.planner.delegation.ExecutorBase.execute(ExecutorBase.java:52) > 2020-05-08T09:40:02.1882689Z at > org.apache.flink.table.api.internal.TableEnvironmentImpl.execute(TableEnvironmentImpl.java:916) > 2020-05-08T09:40:02.1883216Z at > org.apache.flink.table.api.TableUtils.collectToList(TableUtils.java:85) > 2020-05-08T09:40:02.1884413Z at > org.apache.flink.connectors.hive.TableEnvHiveConnectorTest.testDefaultPartitionName(TableEnvHiveConnectorTest.java:106) > 2020-05-08T09:40:02.1885239Z at > sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > 2020-05-08T09:40:02.1886075Z at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > 2020-05-08T09:40:02.1886761Z at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > 2020-05-08T09:40:02.1887300Z at > java.lang.reflect.Method.invoke(Method.java:498) > 2020-05-08T09:40:02.1887851Z at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) > 2020-05-08T09:40:02.1888403Z at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) > 2020-05-08T09:40:02.1888874Z at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) > 2020-05-08T09:40:02.1889500Z at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) > 2020-05-08T09:40:02.1890020Z at > org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runTestMethod(FlinkStandaloneHiveRunner.java:169) > 2020-05-08T09:40:02.1894803Z at > org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runChild(FlinkStandaloneHiveRunner.java:154) > 2020-05-08T09:40:02.1895796Z at > org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runChild(FlinkStandaloneHiveRunner.java:92) > 2020-05-08T09:40:02.1896503Z at > org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) > 2020-05-08T09:40:02.1897131Z at > org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71) > 2020-05-08T09:40:02.1897749Z at > org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) > 2020-05-08T09:40:02.1898405Z at > org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) > 2020-05-08T09:40:02.1899264Z at > org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) > 2020-05-08T09:40:02.1899935Z at > org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) > 2020-05-08T09:40:02.1900618Z at > org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48) > 2020-05-08T09:40:02.1901425Z at > org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48) > 2020-05-08T09:40:02.1902191Z at > org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48) > 2020-05-08T09:40:02.1902841Z at > org.junit.rules.RunRules.evaluate(RunRules.java:20) > 2020-05-08T09:40:02.1903518Z at > org.junit.runners.ParentRunner.run(ParentRunner.java:363) > 2020-05-08T09:40:02.1904171Z at > org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365) > 2020-05-08T09:40:02.1905065Z at > org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273) > 2020-05-08T09:40:02.1905855Z at > org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238) > 2020-05-08T09:40:02.1906654Z at > org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159) > 2020-05-08T09:40:02.1907706Z at > org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384) > 2020-05-08T09:40:02.1908493Z at > org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345) > 2020-05-08T09:40:02.1909217Z at > org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126) > 2020-05-08T09:40:02.1909955Z at > org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418) > 2020-05-08T09:40:02.1910663Z Caused by: > org.apache.flink.runtime.client.JobExecutionException: Job execution failed. > 2020-05-08T09:40:02.1911517Z at > org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:147) > 2020-05-08T09:40:02.1912462Z at > org.apache.flink.client.program.PerJobMiniClusterFactory$PerJobMiniClusterJobClient.lambda$getJobExecutionResult$2(PerJobMiniClusterFactory.java:179) > 2020-05-08T09:40:02.1913762Z at > java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:616) > 2020-05-08T09:40:02.1914637Z at > java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:591) > 2020-05-08T09:40:02.1915388Z at > java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488) > 2020-05-08T09:40:02.1916119Z at > java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975) > 2020-05-08T09:40:02.1917085Z at > org.apache.flink.runtime.rpc.akka.AkkaInvocationHandler.lambda$invokeRpc$0(AkkaInvocationHandler.java:229) > 2020-05-08T09:40:02.1917933Z at > java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:774) > 2020-05-08T09:40:02.1918701Z at > java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:750) > 2020-05-08T09:40:02.1919480Z at > java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488) > 2020-05-08T09:40:02.1920131Z at > java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975) > 2020-05-08T09:40:02.1920772Z at > org.apache.flink.runtime.concurrent.FutureUtils$1.onComplete(FutureUtils.java:890) > 2020-05-08T09:40:02.1921625Z at > akka.dispatch.OnComplete.internal(Future.scala:264) > 2020-05-08T09:40:02.1922178Z at > akka.dispatch.OnComplete.internal(Future.scala:261) > 2020-05-08T09:40:02.1922718Z at > akka.dispatch.japi$CallbackBridge.apply(Future.scala:191) > 2020-05-08T09:40:02.1923375Z at > akka.dispatch.japi$CallbackBridge.apply(Future.scala:188) > 2020-05-08T09:40:02.1924014Z at > scala.concurrent.impl.CallbackRunnable.run(Promise.scala:36) > 2020-05-08T09:40:02.1924841Z at > org.apache.flink.runtime.concurrent.Executors$DirectExecutionContext.execute(Executors.java:74) > 2020-05-08T09:40:02.1925627Z at > scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:44) > 2020-05-08T09:40:02.1926319Z at > scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:252) > 2020-05-08T09:40:02.1926996Z at > akka.pattern.PromiseActorRef.$bang(AskSupport.scala:572) > 2020-05-08T09:40:02.1927720Z at > akka.pattern.PipeToSupport$PipeableFuture$$anonfun$pipeTo$1.applyOrElse(PipeToSupport.scala:22) > 2020-05-08T09:40:02.1928567Z at > akka.pattern.PipeToSupport$PipeableFuture$$anonfun$pipeTo$1.applyOrElse(PipeToSupport.scala:21) > 2020-05-08T09:40:02.1929329Z at > scala.concurrent.Future$$anonfun$andThen$1.apply(Future.scala:436) > 2020-05-08T09:40:02.1929984Z at > scala.concurrent.Future$$anonfun$andThen$1.apply(Future.scala:435) > 2020-05-08T09:40:02.1930641Z at > scala.concurrent.impl.CallbackRunnable.run(Promise.scala:36) > 2020-05-08T09:40:02.1931456Z at > akka.dispatch.BatchingExecutor$AbstractBatch.processBatch(BatchingExecutor.scala:55) > 2020-05-08T09:40:02.1932304Z at > akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply$mcV$sp(BatchingExecutor.scala:91) > 2020-05-08T09:40:02.1933210Z at > akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply(BatchingExecutor.scala:91) > 2020-05-08T09:40:02.1934044Z at > akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply(BatchingExecutor.scala:91) > 2020-05-08T09:40:02.1934899Z at > scala.concurrent.BlockContext$.withBlockContext(BlockContext.scala:72) > 2020-05-08T09:40:02.1935581Z at > akka.dispatch.BatchingExecutor$BlockableBatch.run(BatchingExecutor.scala:90) > 2020-05-08T09:40:02.1936203Z at > akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:40) > 2020-05-08T09:40:02.1936956Z at > akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(ForkJoinExecutorConfigurator.scala:44) > 2020-05-08T09:40:02.1937707Z at > akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) > 2020-05-08T09:40:02.1938395Z at > akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) > 2020-05-08T09:40:02.1939096Z at > akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) > 2020-05-08T09:40:02.1939777Z at > akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) > 2020-05-08T09:40:02.1940540Z Caused by: > org.apache.flink.runtime.JobException: Recovery is suppressed by > NoRestartBackoffTimeStrategy > 2020-05-08T09:40:02.1941702Z at > org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.handleFailure(ExecutionFailureHandler.java:112) > 2020-05-08T09:40:02.1942721Z at > org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.getFailureHandlingResult(ExecutionFailureHandler.java:78) > 2020-05-08T09:40:02.1943727Z at > org.apache.flink.runtime.scheduler.DefaultScheduler.handleTaskFailure(DefaultScheduler.java:189) > 2020-05-08T09:40:02.1946503Z at > org.apache.flink.runtime.scheduler.DefaultScheduler.maybeHandleTaskFailure(DefaultScheduler.java:183) > 2020-05-08T09:40:02.1947357Z at > org.apache.flink.runtime.scheduler.DefaultScheduler.updateTaskExecutionStateInternal(DefaultScheduler.java:177) > 2020-05-08T09:40:02.1948239Z at > org.apache.flink.runtime.scheduler.SchedulerBase.updateTaskExecutionState(SchedulerBase.java:497) > 2020-05-08T09:40:02.1949045Z at > org.apache.flink.runtime.jobmaster.JobMaster.updateTaskExecutionState(JobMaster.java:384) > 2020-05-08T09:40:02.1949693Z at > sun.reflect.GeneratedMethodAccessor46.invoke(Unknown Source) > 2020-05-08T09:40:02.1950329Z at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > 2020-05-08T09:40:02.1950963Z at > java.lang.reflect.Method.invoke(Method.java:498) > 2020-05-08T09:40:02.1951755Z at > org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcInvocation(AkkaRpcActor.java:284) > 2020-05-08T09:40:02.1952473Z at > org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:199) > 2020-05-08T09:40:02.1953417Z at > org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:74) > 2020-05-08T09:40:02.1954212Z at > org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:152) > 2020-05-08T09:40:02.1955031Z at > akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26) > 2020-05-08T09:40:02.1955651Z at > akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21) > 2020-05-08T09:40:02.1956318Z at > scala.PartialFunction$class.applyOrElse(PartialFunction.scala:123) > 2020-05-08T09:40:02.1956946Z at > akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21) > 2020-05-08T09:40:02.1957628Z at > scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:170) > 2020-05-08T09:40:02.1958290Z at > scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171) > 2020-05-08T09:40:02.1958959Z at > scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171) > 2020-05-08T09:40:02.1959598Z at > akka.actor.Actor$class.aroundReceive(Actor.scala:517) > 2020-05-08T09:40:02.1960197Z at > akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225) > 2020-05-08T09:40:02.1960845Z at > akka.actor.ActorCell.receiveMessage(ActorCell.scala:592) > 2020-05-08T09:40:02.1961540Z at > akka.actor.ActorCell.invoke(ActorCell.scala:561) > 2020-05-08T09:40:02.1962127Z at > akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258) > 2020-05-08T09:40:02.1962677Z at akka.dispatch.Mailbox.run(Mailbox.scala:225) > 2020-05-08T09:40:02.1963303Z at akka.dispatch.Mailbox.exec(Mailbox.scala:235) > 2020-05-08T09:40:02.1963710Z ... 4 more > 2020-05-08T09:40:02.1964182Z Caused by: java.lang.OutOfMemoryError: unable to > create new native thread > 2020-05-08T09:40:02.1964816Z at java.lang.Thread.start0(Native Method) > 2020-05-08T09:40:02.1965294Z at java.lang.Thread.start(Thread.java:717) > 2020-05-08T09:40:02.1965949Z at > org.apache.flink.streaming.runtime.tasks.SourceStreamTask.processInput(SourceStreamTask.java:133) > 2020-05-08T09:40:02.1966867Z at > org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxStep(MailboxProcessor.java:206) > 2020-05-08T09:40:02.1967767Z at > org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:196) > 2020-05-08T09:40:02.1968639Z at > org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:503) > 2020-05-08T09:40:02.1969406Z at > org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:483) > 2020-05-08T09:40:02.1970296Z at > org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:713) > 2020-05-08T09:40:02.1970944Z at > org.apache.flink.runtime.taskmanager.Task.run(Task.java:539) > 2020-05-08T09:40:02.1971568Z at java.lang.Thread.run(Thread.java:748) > 2020-05-08T09:40:02.1971844Z > 2020-05-08T09:40:02.1972463Z [ERROR] > testInsertOverwrite(org.apache.flink.connectors.hive.TableEnvHiveConnectorTest) > Time elapsed: 0.116 s <<< ERROR! > 2020-05-08T09:40:02.1973835Z java.lang.IllegalArgumentException: Failed to > executeQuery Hive query create database db1: Error while processing > statement: FAILED: Execution Error, return code 1 from > org.apache.hadoop.hive.ql.exec.DDLTask. Database db1 already exists > 2020-05-08T09:40:02.1975322Z at > com.klarna.hiverunner.HiveServerContainer.executeStatement(HiveServerContainer.java:143) > 2020-05-08T09:40:02.1976209Z at > com.klarna.hiverunner.builder.HiveShellBase.executeStatementsWithCommandShellEmulation(HiveShellBase.java:121) > 2020-05-08T09:40:02.1977145Z at > com.klarna.hiverunner.builder.HiveShellBase.executeScriptWithCommandShellEmulation(HiveShellBase.java:110) > 2020-05-08T09:40:02.1977986Z at > com.klarna.hiverunner.builder.HiveShellBase.execute(HiveShellBase.java:129) > 2020-05-08T09:40:02.1978847Z at > org.apache.flink.connectors.hive.TableEnvHiveConnectorTest.testInsertOverwrite(TableEnvHiveConnectorTest.java:238) > 2020-05-08T09:40:02.1979610Z at > sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > 2020-05-08T09:40:02.1980251Z at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > 2020-05-08T09:40:02.1981013Z at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > 2020-05-08T09:40:02.1981814Z at > java.lang.reflect.Method.invoke(Method.java:498) > 2020-05-08T09:40:02.1982562Z at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) > 2020-05-08T09:40:02.1983409Z at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) > 2020-05-08T09:40:02.1984113Z at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) > 2020-05-08T09:40:02.1984996Z at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) > 2020-05-08T09:40:02.1985742Z at > org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runTestMethod(FlinkStandaloneHiveRunner.java:169) > 2020-05-08T09:40:02.1986654Z at > org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runChild(FlinkStandaloneHiveRunner.java:154) > 2020-05-08T09:40:02.1987532Z at > org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runChild(FlinkStandaloneHiveRunner.java:92) > 2020-05-08T09:40:02.1988310Z at > org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) > 2020-05-08T09:40:02.1988950Z at > org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71) > 2020-05-08T09:40:02.1989594Z at > org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) > 2020-05-08T09:40:02.1990247Z at > org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) > 2020-05-08T09:40:02.1990883Z at > org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) > 2020-05-08T09:40:02.1991751Z at > org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) > 2020-05-08T09:40:02.1992455Z at > org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48) > 2020-05-08T09:40:02.1993245Z at > org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48) > 2020-05-08T09:40:02.1993917Z at > org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48) > 2020-05-08T09:40:02.1994676Z at > org.junit.rules.RunRules.evaluate(RunRules.java:20) > 2020-05-08T09:40:02.1995261Z at > org.junit.runners.ParentRunner.run(ParentRunner.java:363) > 2020-05-08T09:40:02.1995958Z at > org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365) > 2020-05-08T09:40:02.1996725Z at > org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273) > 2020-05-08T09:40:02.1997542Z at > org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238) > 2020-05-08T09:40:02.1998499Z at > org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159) > 2020-05-08T09:40:02.1999330Z at > org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384) > 2020-05-08T09:40:02.2000177Z at > org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345) > 2020-05-08T09:40:02.2000934Z at > org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126) > 2020-05-08T09:40:02.2001791Z at > org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418) > 2020-05-08T09:40:02.2002957Z Caused by: > org.apache.hive.service.cli.HiveSQLException: Error while processing > statement: FAILED: Execution Error, return code 1 from > org.apache.hadoop.hive.ql.exec.DDLTask. Database db1 already exists > 2020-05-08T09:40:02.2004116Z at > org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:380) > 2020-05-08T09:40:02.2005005Z at > org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:257) > 2020-05-08T09:40:02.2005745Z at > org.apache.hive.service.cli.operation.SQLOperation.runInternal(SQLOperation.java:293) > 2020-05-08T09:40:02.2006515Z at > org.apache.hive.service.cli.operation.Operation.run(Operation.java:320) > 2020-05-08T09:40:02.2007317Z at > org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:530) > 2020-05-08T09:40:02.2008209Z at > org.apache.hive.service.cli.session.HiveSessionImpl.executeStatement(HiveSessionImpl.java:500) > 2020-05-08T09:40:02.2008976Z at > org.apache.hive.service.cli.CLIService.executeStatement(CLIService.java:265) > 2020-05-08T09:40:02.2009739Z at > com.klarna.hiverunner.HiveServerContainer.executeStatement(HiveServerContainer.java:116) > 2020-05-08T09:40:02.2010277Z ... 34 more > 2020-05-08T09:40:02.2010807Z Caused by: > org.apache.hadoop.hive.ql.metadata.HiveException: Database db1 already exists > 2020-05-08T09:40:02.2011697Z at > org.apache.hadoop.hive.ql.exec.DDLTask.createDatabase(DDLTask.java:4247) > 2020-05-08T09:40:02.2012387Z at > org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:319) > 2020-05-08T09:40:02.2013040Z at > org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:199) > 2020-05-08T09:40:02.2013834Z at > org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:100) > 2020-05-08T09:40:02.2014628Z at > org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2183) > 2020-05-08T09:40:02.2015266Z at > org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1839) > 2020-05-08T09:40:02.2015878Z at > org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1526) > 2020-05-08T09:40:02.2016484Z at > org.apache.hadoop.hive.ql.Driver.run(Driver.java:1237) > 2020-05-08T09:40:02.2017073Z at > org.apache.hadoop.hive.ql.Driver.run(Driver.java:1232) > 2020-05-08T09:40:02.2017723Z at > org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:255) > 2020-05-08T09:40:02.2018263Z ... 40 more > 2020-05-08T09:40:02.2018703Z Caused by: > AlreadyExistsException(message:Database db1 already exists) > 2020-05-08T09:40:02.2019631Z at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$create_database_result$create_database_resultStandardScheme.read(ThriftHiveMetastore.java:26487) > 2020-05-08T09:40:02.2020824Z at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$create_database_result$create_database_resultStandardScheme.read(ThriftHiveMetastore.java:26473) > 2020-05-08T09:40:02.2022044Z at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$create_database_result.read(ThriftHiveMetastore.java:26407) > 2020-05-08T09:40:02.2022859Z at > org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:86) > 2020-05-08T09:40:02.2023761Z at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_create_database(ThriftHiveMetastore.java:749) > 2020-05-08T09:40:02.2024792Z at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.create_database(ThriftHiveMetastore.java:736) > 2020-05-08T09:40:02.2025682Z at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createDatabase(HiveMetaStoreClient.java:727) > 2020-05-08T09:40:02.2026541Z at > sun.reflect.GeneratedMethodAccessor56.invoke(Unknown Source) > 2020-05-08T09:40:02.2027193Z at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > 2020-05-08T09:40:02.2027866Z at > java.lang.reflect.Method.invoke(Method.java:498) > 2020-05-08T09:40:02.2028536Z at > org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:173) > 2020-05-08T09:40:02.2029209Z at > com.sun.proxy.$Proxy33.createDatabase(Unknown Source) > 2020-05-08T09:40:02.2029824Z at > sun.reflect.GeneratedMethodAccessor56.invoke(Unknown Source) > 2020-05-08T09:40:02.2030493Z at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > 2020-05-08T09:40:02.2031244Z at > java.lang.reflect.Method.invoke(Method.java:498) > 2020-05-08T09:40:02.2031998Z at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient$SynchronizedHandler.invoke(HiveMetaStoreClient.java:2330) > 2020-05-08T09:40:02.2032716Z at > com.sun.proxy.$Proxy33.createDatabase(Unknown Source) > 2020-05-08T09:40:02.2033414Z at > org.apache.hadoop.hive.ql.metadata.Hive.createDatabase(Hive.java:427) > 2020-05-08T09:40:02.2034174Z at > org.apache.hadoop.hive.ql.exec.DDLTask.createDatabase(DDLTask.java:4243) > 2020-05-08T09:40:02.2034959Z ... 49 more > 2020-05-08T09:40:02.2035232Z > 2020-05-08T09:40:02.2036012Z [ERROR] > testDifferentFormats(org.apache.flink.connectors.hive.TableEnvHiveConnectorTest) > Time elapsed: 0.031 s <<< ERROR! > 2020-05-08T09:40:02.2037188Z java.lang.IllegalArgumentException: Failed to > executeQuery Hive query create database db1: Error while processing > statement: FAILED: Execution Error, return code 1 from > org.apache.hadoop.hive.ql.exec.DDLTask. Database db1 already exists > 2020-05-08T09:40:02.2038329Z at > com.klarna.hiverunner.HiveServerContainer.executeStatement(HiveServerContainer.java:143) > 2020-05-08T09:40:02.2039176Z at > com.klarna.hiverunner.builder.HiveShellBase.executeStatementsWithCommandShellEmulation(HiveShellBase.java:121) > 2020-05-08T09:40:02.2040066Z at > com.klarna.hiverunner.builder.HiveShellBase.executeScriptWithCommandShellEmulation(HiveShellBase.java:110) > 2020-05-08T09:40:02.2040964Z at > com.klarna.hiverunner.builder.HiveShellBase.execute(HiveShellBase.java:129) > 2020-05-08T09:40:02.2042010Z at > org.apache.flink.connectors.hive.TableEnvHiveConnectorTest.readWriteFormat(TableEnvHiveConnectorTest.java:142) > 2020-05-08T09:40:02.2042941Z at > org.apache.flink.connectors.hive.TableEnvHiveConnectorTest.testDifferentFormats(TableEnvHiveConnectorTest.java:135) > 2020-05-08T09:40:02.2043764Z at > sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > 2020-05-08T09:40:02.2044514Z at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > 2020-05-08T09:40:02.2045198Z at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > 2020-05-08T09:40:02.2045812Z at > java.lang.reflect.Method.invoke(Method.java:498) > 2020-05-08T09:40:02.2046423Z at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) > 2020-05-08T09:40:02.2047134Z at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) > 2020-05-08T09:40:02.2047860Z at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) > 2020-05-08T09:40:02.2048579Z at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) > 2020-05-08T09:40:02.2049372Z at > org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runTestMethod(FlinkStandaloneHiveRunner.java:169) > 2020-05-08T09:40:02.2050242Z at > org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runChild(FlinkStandaloneHiveRunner.java:154) > 2020-05-08T09:40:02.2051230Z at > org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runChild(FlinkStandaloneHiveRunner.java:92) > 2020-05-08T09:40:02.2051916Z at > org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) > 2020-05-08T09:40:02.2052516Z at > org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71) > 2020-05-08T09:40:02.2053144Z at > org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) > 2020-05-08T09:40:02.2054136Z at > org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) > 2020-05-08T09:40:02.2055027Z at > org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) > 2020-05-08T09:40:02.2055775Z at > org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) > 2020-05-08T09:40:02.2056731Z at > org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48) > 2020-05-08T09:40:02.2057369Z at > org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48) > 2020-05-08T09:40:02.2058109Z at > org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48) > 2020-05-08T09:40:02.2058580Z at > org.junit.rules.RunRules.evaluate(RunRules.java:20) > 2020-05-08T09:40:02.2059127Z at > org.junit.runners.ParentRunner.run(ParentRunner.java:363) > 2020-05-08T09:40:02.2059523Z at > org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365) > 2020-05-08T09:40:02.2060161Z at > org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273) > 2020-05-08T09:40:02.2060798Z at > org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238) > 2020-05-08T09:40:02.2061500Z at > org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159) > 2020-05-08T09:40:02.2062191Z at > org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384) > 2020-05-08T09:40:02.2062826Z at > org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345) > 2020-05-08T09:40:02.2063446Z at > org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126) > 2020-05-08T09:40:02.2063921Z at > org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418) > 2020-05-08T09:40:02.2064783Z Caused by: > org.apache.hive.service.cli.HiveSQLException: Error while processing > statement: FAILED: Execution Error, return code 1 from > org.apache.hadoop.hive.ql.exec.DDLTask. Database db1 already exists > 2020-05-08T09:40:02.2065630Z at > org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:380) > 2020-05-08T09:40:02.2066238Z at > org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:257) > 2020-05-08T09:40:02.2066729Z at > org.apache.hive.service.cli.operation.SQLOperation.runInternal(SQLOperation.java:293) > 2020-05-08T09:40:02.2067415Z at > org.apache.hive.service.cli.operation.Operation.run(Operation.java:320) > 2020-05-08T09:40:02.2068080Z at > org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:530) > 2020-05-08T09:40:02.2068798Z at > org.apache.hive.service.cli.session.HiveSessionImpl.executeStatement(HiveSessionImpl.java:500) > 2020-05-08T09:40:02.2069312Z at > org.apache.hive.service.cli.CLIService.executeStatement(CLIService.java:265) > 2020-05-08T09:40:02.2069907Z at > com.klarna.hiverunner.HiveServerContainer.executeStatement(HiveServerContainer.java:116) > 2020-05-08T09:40:02.2070342Z ... 35 more > 2020-05-08T09:40:02.2070713Z Caused by: > org.apache.hadoop.hive.ql.metadata.HiveException: Database db1 already exists > 2020-05-08T09:40:02.2071379Z at > org.apache.hadoop.hive.ql.exec.DDLTask.createDatabase(DDLTask.java:4247) > 2020-05-08T09:40:02.2071954Z at > org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:319) > 2020-05-08T09:40:02.2072449Z at > org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:199) > 2020-05-08T09:40:02.2073014Z at > org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:100) > 2020-05-08T09:40:02.2073720Z at > org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2183) > 2020-05-08T09:40:02.2074206Z at > org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1839) > 2020-05-08T09:40:02.2074833Z at > org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1526) > 2020-05-08T09:40:02.2075300Z at > org.apache.hadoop.hive.ql.Driver.run(Driver.java:1237) > 2020-05-08T09:40:02.2075837Z at > org.apache.hadoop.hive.ql.Driver.run(Driver.java:1232) > 2020-05-08T09:40:02.2076320Z at > org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:255) > 2020-05-08T09:40:02.2076713Z ... 41 more > 2020-05-08T09:40:02.2077221Z Caused by: > AlreadyExistsException(message:Database db1 already exists) > 2020-05-08T09:40:02.2078000Z at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$create_database_result$create_database_resultStandardScheme.read(ThriftHiveMetastore.java:26487) > 2020-05-08T09:40:02.2078840Z at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$create_database_result$create_database_resultStandardScheme.read(ThriftHiveMetastore.java:26473) > 2020-05-08T09:40:02.2079803Z at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$create_database_result.read(ThriftHiveMetastore.java:26407) > 2020-05-08T09:40:02.2080296Z at > org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:86) > 2020-05-08T09:40:02.2080897Z at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_create_database(ThriftHiveMetastore.java:749) > 2020-05-08T09:40:02.2081643Z at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.create_database(ThriftHiveMetastore.java:736) > 2020-05-08T09:40:02.2082363Z at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createDatabase(HiveMetaStoreClient.java:727) > 2020-05-08T09:40:02.2082848Z at > sun.reflect.GeneratedMethodAccessor56.invoke(Unknown Source) > 2020-05-08T09:40:02.2083299Z at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > 2020-05-08T09:40:02.2083946Z at > java.lang.reflect.Method.invoke(Method.java:498) > 2020-05-08T09:40:02.2084542Z at > org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:173) > 2020-05-08T09:40:02.2084952Z at > com.sun.proxy.$Proxy33.createDatabase(Unknown Source) > 2020-05-08T09:40:02.2085384Z at > sun.reflect.GeneratedMethodAccessor56.invoke(Unknown Source) > 2020-05-08T09:40:02.2086078Z at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > 2020-05-08T09:40:02.2086655Z at > java.lang.reflect.Method.invoke(Method.java:498) > 2020-05-08T09:40:02.2087354Z at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient$SynchronizedHandler.invoke(HiveMetaStoreClient.java:2330) > 2020-05-08T09:40:02.2087790Z at > com.sun.proxy.$Proxy33.createDatabase(Unknown Source) > 2020-05-08T09:40:02.2088356Z at > org.apache.hadoop.hive.ql.metadata.Hive.createDatabase(Hive.java:427) > 2020-05-08T09:40:02.2088784Z at > org.apache.hadoop.hive.ql.exec.DDLTask.createDatabase(DDLTask.java:4243) > 2020-05-08T09:40:02.2089306Z ... 50 more > 2020-05-08T09:40:02.2089426Z > 2020-05-08T09:40:02.2089939Z [ERROR] > testUpdatePartitionSD(org.apache.flink.connectors.hive.TableEnvHiveConnectorTest) > Time elapsed: 0.015 s <<< ERROR! > 2020-05-08T09:40:02.2090950Z java.lang.IllegalArgumentException: Failed to > executeQuery Hive query create database db1: Error while processing > statement: FAILED: Execution Error, return code 1 from > org.apache.hadoop.hive.ql.exec.DDLTask. Database db1 already exists > 2020-05-08T09:40:02.2091981Z at > com.klarna.hiverunner.HiveServerContainer.executeStatement(HiveServerContainer.java:143) > 2020-05-08T09:40:02.2092712Z at > com.klarna.hiverunner.builder.HiveShellBase.executeStatementsWithCommandShellEmulation(HiveShellBase.java:121) > 2020-05-08T09:40:02.2093480Z at > com.klarna.hiverunner.builder.HiveShellBase.executeScriptWithCommandShellEmulation(HiveShellBase.java:110) > 2020-05-08T09:40:02.2094185Z at > com.klarna.hiverunner.builder.HiveShellBase.execute(HiveShellBase.java:129) > 2020-05-08T09:40:02.2094909Z at > org.apache.flink.connectors.hive.TableEnvHiveConnectorTest.testUpdatePartitionSD(TableEnvHiveConnectorTest.java:588) > 2020-05-08T09:40:02.2095525Z at > sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > 2020-05-08T09:40:02.2095998Z at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > 2020-05-08T09:40:02.2096642Z at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > 2020-05-08T09:40:02.2097088Z at > java.lang.reflect.Method.invoke(Method.java:498) > 2020-05-08T09:40:02.2097619Z at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) > 2020-05-08T09:40:02.2098340Z at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) > 2020-05-08T09:40:02.2098795Z at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) > 2020-05-08T09:40:02.2099326Z at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) > 2020-05-08T09:40:02.2099978Z at > org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runTestMethod(FlinkStandaloneHiveRunner.java:169) > 2020-05-08T09:40:02.2100828Z at > org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runChild(FlinkStandaloneHiveRunner.java:154) > 2020-05-08T09:40:02.2101591Z at > org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runChild(FlinkStandaloneHiveRunner.java:92) > 2020-05-08T09:40:02.2102040Z at > org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) > 2020-05-08T09:40:02.2102668Z at > org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71) > 2020-05-08T09:40:02.2103233Z at > org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) > 2020-05-08T09:40:02.2103755Z at > org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) > 2020-05-08T09:40:02.2104461Z at > org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) > 2020-05-08T09:40:02.2105023Z at > org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) > 2020-05-08T09:40:02.2105600Z at > org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48) > 2020-05-08T09:40:02.2106187Z at > org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48) > 2020-05-08T09:40:02.2106792Z at > org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48) > 2020-05-08T09:40:02.2107217Z at > org.junit.rules.RunRules.evaluate(RunRules.java:20) > 2020-05-08T09:40:02.2107586Z at > org.junit.runners.ParentRunner.run(ParentRunner.java:363) > 2020-05-08T09:40:02.2108205Z at > org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365) > 2020-05-08T09:40:02.2108794Z at > org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273) > 2020-05-08T09:40:02.2109434Z at > org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238) > 2020-05-08T09:40:02.2109943Z at > org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159) > 2020-05-08T09:40:02.2110695Z at > org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384) > 2020-05-08T09:40:02.2111342Z at > org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345) > 2020-05-08T09:40:02.2111801Z at > org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126) > 2020-05-08T09:40:02.2112443Z at > org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418) > 2020-05-08T09:40:02.2113300Z Caused by: > org.apache.hive.service.cli.HiveSQLException: Error while processing > statement: FAILED: Execution Error, return code 1 from > org.apache.hadoop.hive.ql.exec.DDLTask. Database db1 already exists > 2020-05-08T09:40:02.2114079Z at > org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:380) > 2020-05-08T09:40:02.2114738Z at > org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:257) > 2020-05-08T09:40:02.2115208Z at > org.apache.hive.service.cli.operation.SQLOperation.runInternal(SQLOperation.java:293) > 2020-05-08T09:40:02.2115635Z at > org.apache.hive.service.cli.operation.Operation.run(Operation.java:320) > 2020-05-08T09:40:02.2116251Z at > org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:530) > 2020-05-08T09:40:02.2117029Z at > org.apache.hive.service.cli.session.HiveSessionImpl.executeStatement(HiveSessionImpl.java:500) > 2020-05-08T09:40:02.2117612Z at > org.apache.hive.service.cli.CLIService.executeStatement(CLIService.java:265) > 2020-05-08T09:40:02.2118212Z at > com.klarna.hiverunner.HiveServerContainer.executeStatement(HiveServerContainer.java:116) > 2020-05-08T09:40:02.2118692Z ... 34 more > 2020-05-08T09:40:02.2119010Z Caused by: > org.apache.hadoop.hive.ql.metadata.HiveException: Database db1 already exists > 2020-05-08T09:40:02.2119697Z at > org.apache.hadoop.hive.ql.exec.DDLTask.createDatabase(DDLTask.java:4247) > 2020-05-08T09:40:02.2120268Z at > org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:319) > 2020-05-08T09:40:02.2120643Z at > org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:199) > 2020-05-08T09:40:02.2121352Z at > org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:100) > 2020-05-08T09:40:02.2121812Z at > org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2183) > 2020-05-08T09:40:02.2122461Z at > org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1839) > 2020-05-08T09:40:02.2122824Z at > org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1526) > 2020-05-08T09:40:02.2123422Z at > org.apache.hadoop.hive.ql.Driver.run(Driver.java:1237) > 2020-05-08T09:40:02.2123756Z at > org.apache.hadoop.hive.ql.Driver.run(Driver.java:1232) > 2020-05-08T09:40:02.2124548Z at > org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:255) > 2020-05-08T09:40:02.2125000Z ... 40 more > 2020-05-08T09:40:02.2125286Z Caused by: > AlreadyExistsException(message:Database db1 already exists) > 2020-05-08T09:40:02.2125956Z at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$create_database_result$create_database_resultStandardScheme.read(ThriftHiveMetastore.java:26487) > 2020-05-08T09:40:02.2126801Z at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$create_database_result$create_database_resultStandardScheme.read(ThriftHiveMetastore.java:26473) > 2020-05-08T09:40:02.2127553Z at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$create_database_result.read(ThriftHiveMetastore.java:26407) > 2020-05-08T09:40:02.2128061Z at > org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:86) > 2020-05-08T09:40:02.2128553Z at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_create_database(ThriftHiveMetastore.java:749) > 2020-05-08T09:40:02.2129172Z at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.create_database(ThriftHiveMetastore.java:736) > 2020-05-08T09:40:02.2129821Z at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createDatabase(HiveMetaStoreClient.java:727) > 2020-05-08T09:40:02.2130291Z at > sun.reflect.GeneratedMethodAccessor56.invoke(Unknown Source) > 2020-05-08T09:40:02.2130753Z at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > 2020-05-08T09:40:02.2131444Z at > java.lang.reflect.Method.invoke(Method.java:498) > 2020-05-08T09:40:02.2131903Z at > org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:173) > 2020-05-08T09:40:02.2132399Z at > com.sun.proxy.$Proxy33.createDatabase(Unknown Source) > 2020-05-08T09:40:02.2132829Z at > sun.reflect.GeneratedMethodAccessor56.invoke(Unknown Source) > 2020-05-08T09:40:02.2133338Z at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > 2020-05-08T09:40:02.2133848Z at > java.lang.reflect.Method.invoke(Method.java:498) > 2020-05-08T09:40:02.2134353Z at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient$SynchronizedHandler.invoke(HiveMetaStoreClient.java:2330) > 2020-05-08T09:40:02.2134876Z at > com.sun.proxy.$Proxy33.createDatabase(Unknown Source) > 2020-05-08T09:40:02.2135338Z at > org.apache.hadoop.hive.ql.metadata.Hive.createDatabase(Hive.java:427) > 2020-05-08T09:40:02.2135847Z at > org.apache.hadoop.hive.ql.exec.DDLTask.createDatabase(DDLTask.java:4243) > 2020-05-08T09:40:02.2136195Z ... 49 more > 2020-05-08T09:40:02.2136313Z > 2020-05-08T09:40:03.2246042Z [INFO] Running > org.apache.flink.connectors.hive.HiveDialectTest > 2020-05-08T09:40:16.1437654Z [INFO] Tests run: 2, Failures: 0, Errors: 0, > Skipped: 0, Time elapsed: 12.914 s - in > org.apache.flink.connectors.hive.HiveDialectTest > 2020-05-08T09:40:16.5569091Z [INFO] > 2020-05-08T09:40:16.5569441Z [INFO] Results: > 2020-05-08T09:40:16.5569645Z [INFO] > 2020-05-08T09:40:16.5569962Z [ERROR] Errors: > 2020-05-08T09:40:16.5571473Z [ERROR] > HiveTableSinkTest.testInsertIntoNonPartitionTable:91->createHiveDestTable:233->createHiveDestTable:221 > » TableAlreadyExist > 2020-05-08T09:40:16.5572957Z [ERROR] > HiveTableSinkTest.testWriteComplexType:143 » Execution > org.apache.flink.runtim... > 2020-05-08T09:40:16.5573995Z [ERROR] > TableEnvHiveConnectorTest.testDefaultPartitionName:106 » Execution > org.apache.... > 2020-05-08T09:40:16.5574692Z [ERROR] > TableEnvHiveConnectorTest.testDifferentFormats:135->readWriteFormat:142 » > IllegalArgument > 2020-05-08T09:40:16.5575454Z [ERROR] > TableEnvHiveConnectorTest.testInsertOverwrite:238 » IllegalArgument Failed > to ... > 2020-05-08T09:40:16.5576562Z [ERROR] > TableEnvHiveConnectorTest.testUpdatePartitionSD:588 » IllegalArgument Failed > t... > {code} > instance: > [https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_apis/build/builds/806/logs/127] -- This message was sent by Atlassian Jira (v8.3.4#803005)