[ 
https://issues.apache.org/jira/browse/HIVE-27153?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aman Raj updated HIVE-27153:
----------------------------
    Fix Version/s: 3.2.0

> Revert "HIVE-20182: Backport HIVE-20067 to branch-3"
> ----------------------------------------------------
>
>                 Key: HIVE-27153
>                 URL: https://issues.apache.org/jira/browse/HIVE-27153
>             Project: Hive
>          Issue Type: Sub-task
>            Reporter: Aman Raj
>            Assignee: Aman Raj
>            Priority: Major
>              Labels: pull-request-available
>             Fix For: 3.2.0
>
>          Time Spent: 50m
>  Remaining Estimate: 0h
>
> The mm_all.q test is failing because of this commit. This commit was not 
> validated before committing.
> There is no stack trace for this exception. Link to the exception : 
> [http://ci.hive.apache.org/blue/organizations/jenkins/hive-precommit/detail/PR-4126/2/tests]
>  
> {code:java}
> java.lang.AssertionError: Client execution failed with error code = 1 running 
> "insert into table part_mm_n0 partition(key_mm=455) select key from 
> intermediate_n0" fname=mm_all.q See ./ql/target/tmp/log/hive.log or 
> ./itests/qtest/target/tmp/log/hive.log, or check ./ql/target/surefire-reports 
> or ./itests/qtest/target/surefire-reports/ for specific test cases logs.  at 
> org.junit.Assert.fail(Assert.java:88)        at 
> org.apache.hadoop.hive.ql.QTestUtil.failed(QTestUtil.java:2232)      at 
> org.apache.hadoop.hive.cli.control.CoreCliDriver.runTest(CoreCliDriver.java:180)
>      at 
> org.apache.hadoop.hive.cli.control.CliAdapter.runTest(CliAdapter.java:104)   
> at 
> org.apache.hadoop.hive.cli.split1.TestMiniLlapCliDriver.testCliDriver(TestMiniLlapCliDriver.java:62)
>  at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)  at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
>        at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:498) {code}
>  
>  
> Found the actual error :
> {code:java}
> 2023-03-19T15:18:07,705 DEBUG [699603ee-f4a1-43b7-b160-7faf858ca4b4 main] 
> converters.ArrayConverter: Converting 'java.net.URL[]' value 
> '[Ljava.net.URL;@7535f28' to type 'java.net.URL[]'
> 2023-03-19T15:18:07,705 DEBUG [699603ee-f4a1-43b7-b160-7faf858ca4b4 main] 
> converters.ArrayConverter:     No conversion required, value is already a 
> java.net.URL[]
> 2023-03-19T15:18:07,819  INFO [699603ee-f4a1-43b7-b160-7faf858ca4b4 main] 
> beanutils.FluentPropertyBeanIntrospector: Error when creating 
> PropertyDescriptor for public final void 
> org.apache.commons.configuration2.AbstractConfiguration.setProperty(java.lang.String,java.lang.Object)!
>  Ignoring this property.
> 2023-03-19T15:18:07,819 DEBUG [699603ee-f4a1-43b7-b160-7faf858ca4b4 main] 
> beanutils.FluentPropertyBeanIntrospector: Exception is:
> java.beans.IntrospectionException: bad write method arg count: public final 
> void 
> org.apache.commons.configuration2.AbstractConfiguration.setProperty(java.lang.String,java.lang.Object)
>     at 
> java.beans.PropertyDescriptor.findPropertyType(PropertyDescriptor.java:657) 
> ~[?:1.8.0_342]
>     at 
> java.beans.PropertyDescriptor.setWriteMethod(PropertyDescriptor.java:327) 
> ~[?:1.8.0_342]
>     at java.beans.PropertyDescriptor.<init>(PropertyDescriptor.java:139) 
> ~[?:1.8.0_342]
>     at 
> org.apache.commons.beanutils.FluentPropertyBeanIntrospector.createFluentPropertyDescritor(FluentPropertyBeanIntrospector.java:178)
>  ~[commons-beanutils-1.9.3.jar:1.9.3]
>     at 
> org.apache.commons.beanutils.FluentPropertyBeanIntrospector.introspect(FluentPropertyBeanIntrospector.java:141)
>  [commons-beanutils-1.9.3.jar:1.9.3]
>     at 
> org.apache.commons.beanutils.PropertyUtilsBean.fetchIntrospectionData(PropertyUtilsBean.java:2245)
>  [commons-beanutils-1.9.3.jar:1.9.3]
>     at 
> org.apache.commons.beanutils.PropertyUtilsBean.getIntrospectionData(PropertyUtilsBean.java:2226)
>  [commons-beanutils-1.9.3.jar:1.9.3]
>     at 
> org.apache.commons.beanutils.PropertyUtilsBean.getPropertyDescriptor(PropertyUtilsBean.java:954)
>  [commons-beanutils-1.9.3.jar:1.9.3]
>     at 
> org.apache.commons.beanutils.PropertyUtilsBean.isWriteable(PropertyUtilsBean.java:1478)
>  [commons-beanutils-1.9.3.jar:1.9.3]
>     at 
> org.apache.commons.configuration2.beanutils.BeanHelper.isPropertyWriteable(BeanHelper.java:521)
>  [commons-configuration2-2.1.1.jar:2.1.1]
>     at 
> org.apache.commons.configuration2.beanutils.BeanHelper.initProperty(BeanHelper.java:357)
>  [commons-configuration2-2.1.1.jar:2.1.1]
>     at 
> org.apache.commons.configuration2.beanutils.BeanHelper.initBeanProperties(BeanHelper.java:273)
>  [commons-configuration2-2.1.1.jar:2.1.1]
>     at 
> org.apache.commons.configuration2.beanutils.BeanHelper.initBean(BeanHelper.java:192)
>  [commons-configuration2-2.1.1.jar:2.1.1]
>     at 
> org.apache.commons.configuration2.beanutils.BeanHelper$BeanCreationContextImpl.initBean(BeanHelper.java:669)
>  [commons-configuration2-2.1.1.jar:2.1.1]
>     at 
> org.apache.commons.configuration2.beanutils.DefaultBeanFactory.initBeanInstance(DefaultBeanFactory.java:162)
>  [commons-configuration2-2.1.1.jar:2.1.1]
>     at 
> org.apache.commons.configuration2.beanutils.DefaultBeanFactory.createBean(DefaultBeanFactory.java:116)
>  [commons-configuration2-2.1.1.jar:2.1.1]
>     at 
> org.apache.commons.configuration2.beanutils.BeanHelper.createBean(BeanHelper.java:459)
>  [commons-configuration2-2.1.1.jar:2.1.1]
>     at 
> org.apache.commons.configuration2.beanutils.BeanHelper.createBean(BeanHelper.java:479)
>  [commons-configuration2-2.1.1.jar:2.1.1]
>     at 
> org.apache.commons.configuration2.beanutils.BeanHelper.createBean(BeanHelper.java:492)
>  [commons-configuration2-2.1.1.jar:2.1.1]
>     at 
> org.apache.commons.configuration2.builder.BasicConfigurationBuilder.createResultInstance(BasicConfigurationBuilder.java:447)
>  [commons-configuration2-2.1.1.jar:2.1.1]
>     at 
> org.apache.commons.configuration2.builder.BasicConfigurationBuilder.createResult(BasicConfigurationBuilder.java:417)
>  [commons-configuration2-2.1.1.jar:2.1.1]
>     at 
> org.apache.commons.configuration2.builder.BasicConfigurationBuilder.getConfiguration(BasicConfigurationBuilder.java:285)
>  [commons-configuration2-2.1.1.jar:2.1.1]
>     at 
> org.apache.hadoop.metrics2.impl.MetricsConfig.loadFirst(MetricsConfig.java:119)
>  [hadoop-common-3.1.0.jar:?]
>     at 
> org.apache.hadoop.metrics2.impl.MetricsConfig.create(MetricsConfig.java:98) 
> [hadoop-common-3.1.0.jar:?]
>     at 
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl.configure(MetricsSystemImpl.java:478)
>  [hadoop-common-3.1.0.jar:?]
>     at 
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl.start(MetricsSystemImpl.java:188)
>  [hadoop-common-3.1.0.jar:?]
>     at 
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl.init(MetricsSystemImpl.java:163)
>  [hadoop-common-3.1.0.jar:?]
>     at 
> org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.init(DefaultMetricsSystem.java:62)
>  [hadoop-common-3.1.0.jar:?]
>     at 
> org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.initialize(DefaultMetricsSystem.java:58)
>  [hadoop-common-3.1.0.jar:?]
>     at 
> org.apache.hadoop.mapred.LocalJobRunnerMetrics.create(LocalJobRunnerMetrics.java:45)
>  [hadoop-mapreduce-client-common-3.1.0.jar:?]
>     at 
> org.apache.hadoop.mapred.LocalJobRunner.<init>(LocalJobRunner.java:771) 
> [hadoop-mapreduce-client-common-3.1.0.jar:?]
>     at 
> org.apache.hadoop.mapred.LocalJobRunner.<init>(LocalJobRunner.java:764) 
> [hadoop-mapreduce-client-common-3.1.0.jar:?]
>     at 
> org.apache.hadoop.mapred.LocalClientProtocolProvider.create(LocalClientProtocolProvider.java:42)
>  [hadoop-mapreduce-client-common-3.1.0.jar:?]
>     at org.apache.hadoop.mapreduce.Cluster.initialize(Cluster.java:130) 
> [hadoop-mapreduce-client-core-3.1.0.jar:?]
>     at org.apache.hadoop.mapreduce.Cluster.<init>(Cluster.java:109) 
> [hadoop-mapreduce-client-core-3.1.0.jar:?]
>     at org.apache.hadoop.mapreduce.Cluster.<init>(Cluster.java:102) 
> [hadoop-mapreduce-client-core-3.1.0.jar:?]
>     at org.apache.hadoop.mapred.JobClient.init(JobClient.java:475) 
> [hadoop-mapreduce-client-core-3.1.0.jar:?]
>     at org.apache.hadoop.mapred.JobClient.<init>(JobClient.java:454) 
> [hadoop-mapreduce-client-core-3.1.0.jar:?]
>     at 
> org.apache.hadoop.hive.ql.exec.mr.ExecDriver.execute(ExecDriver.java:381) 
> [classes/:?]
>     at 
> org.apache.hadoop.hive.ql.exec.mr.MapRedTask.execute(MapRedTask.java:149) 
> [classes/:?]
>     at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:210) 
> [classes/:?]
>     at 
> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97) 
> [classes/:?]
>     at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2692) 
> [classes/:?]
>     at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2363) [classes/:?]
>     at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2039) 
> [classes/:?]
>     at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1737) [classes/:?]
>     at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1731) [classes/:?]
>     at 
> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157) 
> [classes/:?]
>     at 
> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218) 
> [classes/:?]
>     at 
> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239) 
> [classes/:?]
>     at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188) 
> [classes/:?]
>     at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402) 
> [classes/:?]
>     at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:335) 
> [classes/:?]
>     at 
> org.apache.hadoop.hive.ql.QTestUtil.initFromScript(QTestUtil.java:1216) 
> [classes/:?]
>     at org.apache.hadoop.hive.ql.QTestUtil.createSources(QTestUtil.java:1201) 
> [classes/:?]
>     at org.apache.hadoop.hive.ql.QTestUtil.createSources(QTestUtil.java:1188) 
> [classes/:?]
>     at 
> org.apache.hadoop.hive.cli.control.CoreCliDriver$3.invokeInternal(CoreCliDriver.java:83)
>  [classes/:?]
>     at 
> org.apache.hadoop.hive.cli.control.CoreCliDriver$3.invokeInternal(CoreCliDriver.java:80)
>  [classes/:?]
>     at 
> org.apache.hadoop.hive.util.ElapsedTimeLoggingWrapper.invoke(ElapsedTimeLoggingWrapper.java:33)
>  [classes/:?]
>     at 
> org.apache.hadoop.hive.cli.control.CoreCliDriver.beforeClass(CoreCliDriver.java:86)
>  [classes/:?]
>     at 
> org.apache.hadoop.hive.cli.control.CliAdapter$1$1.evaluate(CliAdapter.java:71)
>  [classes/:?]
>     at org.junit.rules.RunRules.evaluate(RunRules.java:20) [junit-4.11.jar:?]
>     at org.junit.runners.ParentRunner.run(ParentRunner.java:309) 
> [junit-4.11.jar:?]
>     at 
> org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
>  [surefire-junit4-2.21.0.jar:2.21.0]
>     at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273)
>  [surefire-junit4-2.21.0.jar:2.21.0]
>     at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
>  [surefire-junit4-2.21.0.jar:2.21.0]
>     at 
> org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
>  [surefire-junit4-2.21.0.jar:2.21.0]
>     at 
> org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:379)
>  [surefire-booter-2.21.0.jar:2.21.0]
>     at 
> org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:340)
>  [surefire-booter-2.21.0.jar:2.21.0]
>     at 
> org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:125) 
> [surefire-booter-2.21.0.jar:2.21.0]
>     at 
> org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:413) 
> [surefire-booter-2.21.0.jar:2.21.0] {code}
>  
> Also this can be an issue :
> {code:java}
> 2023-03-19T15:19:06,187 ERROR [699603ee-f4a1-43b7-b160-7faf858ca4b4 main] 
> ql.Driver: FAILED: Hive Internal Error: 
> org.apache.hadoop.hive.ql.metadata.HiveException(Error while invoking 
> PreHook. hooks: java.lang.RuntimeException: Cannot overwrite read-only table: 
> src
>     at 
> org.apache.hadoop.hive.ql.hooks.EnforceReadOnlyTables.run(EnforceReadOnlyTables.java:64)
>     at 
> org.apache.hadoop.hive.ql.hooks.EnforceReadOnlyTables.run(EnforceReadOnlyTables.java:44)
>     at 
> org.apache.hadoop.hive.ql.HookRunner.invokeGeneralHook(HookRunner.java:296)
>     at org.apache.hadoop.hive.ql.HookRunner.runPreHooks(HookRunner.java:273)
>     at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2304)
>     at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2039)
>     at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1737)
>     at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1731)
>     at 
> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157)
>     at 
> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218)
>     at 
> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239)
>     at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188)
>     at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402) 
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to