Hi shaofeng: Yes, I can run the command well in hive shell.
I can’t find calcite-core version 1.13.0-kylin-r4. Best wishes. > 在 2018年10月15日,下午2:20,ShaoFeng Shi <shaofeng...@apache.org> 写道: > > Hive version 2.3.3 can work well with HDP 3? Can you try the HiveQL that > Kylin executed out of Kylin, if it works, then there should be something > wrong in Kylin. > > liuzhixin <liuz...@163.com <mailto:liuz...@163.com>> 于2018年10月15日周一 下午1:47写道: > >> Thank you for the answer! >> >> I can’t decide the hive version. >> >> And the hive version 2.3.3 can work well with HDP 3. >> >> Perhaps you can test the Kylin with hive version 2.3.3. >> >> Maybe it’s other error. Thanks! >> >> Best wishes! >> >> >> 在 2018年10月15日,下午1:24,ShaoFeng Shi <shaofeng...@apache.org> 写道: >> >> Hi zhixin, >> >> I think the problem is how to run Hive 2 with HDP 3, no relation with >> Kylin. >> >> Usually, we don't encourage user to customize the component version in a >> release, because that may bring dependency conflicts. >> >> I suggest you use the original Hive version in HDP 3. >> >> liuzhixin <liuz...@163.com <mailto:liuz...@163.com>> 于2018年10月15日周一 >> 上午11:25写道: >> >>> Hi ShaoFeng Shi >>> >>> Yes, the error from hive version 2.3.3, >>> >>> And Kylin need hive version 3.1.0. >>> >>> So how to solve the question? >>> >>> Best wishes! >>> >>>> 在 2018年10月15日,上午11:10,ShaoFeng Shi <shaofeng...@apache.org >>>> <mailto:shaofeng...@apache.org>> 写道: >>>> >>>> Hi Zhixin, >>>> >>>> The error log is thrown from Hive, not from Kylin I think. Please verify >>>> your hive is properly installed; You can manually run that hive command >>> : >>>> >>>> hive -e "use default; xxx" >>>> >>>> Lijun Cao <641507...@qq.com <mailto:641507...@qq.com>> 于2018年10月15日周一 >>>> 上午11:01写道: >>>> >>>>> Hi liuzhixin: >>>>> >>>>> As I remember, the Hive version in HDP 3 is 3.1.0 . >>>>> >>>>> You can update Hive to 3.1.0 and then have another try. >>>>> >>>>> And according to my previous test, the binary package >>>>> apache-kylin-2.5.0-bin-hadoop3.tar.gz can work properly on HDP 3. You >>> can >>>>> get it form official site. >>>>> >>>>> Best Regards >>>>> >>>>> Lijun Cao >>>>> >>>>>> 在 2018年10月15日,10:22,liuzhixin <liuz...@163.com <mailto:liuz...@163.com>> >>>>>> 写道: >>>>>> >>>>>> hi cao lijun, >>>>>> # >>>>>> the platform is ambari hdp3.0, and hive is 2.3.3, hbase version is 2.0 >>>>>> >>>>>> I have compile the source code with hive 2.3.3, >>>>>> >>>>>> but the module atopcalcite depends on protobuf 3.1.0, >>>>>> >>>>>> other module depends on protobuf 2.5.0. >>>>>> >>>>>> >>>>>>> 在 2018年10月15日,上午8:40,Lijun Cao <641507...@qq.com >>>>>>> <mailto:641507...@qq.com>> 写道: >>>>>>> >>>>>>> Hi liuzhixin: >>>>>>> >>>>>>> Which platform did you use? >>>>>>> >>>>>>> The CDH 6.0.x or HDP 3.0 ? >>>>>>> >>>>>>> Best Regards >>>>>>> >>>>>>> Lijun Cao >>>>>>> >>>>>>>> 在 2018年10月12日,21:14,liuzhixin <liuz...@163.com >>>>>>>> <mailto:liuz...@163.com>> 写道: >>>>>>>> >>>>>>>> Logging initialized using configuration in >>>>> >>> file:/data/hadoop-enviorment/apache-hive-2.3.3/conf/hive-log4j2.properties >>>>> Async: true >>>>>>>> OK >>>>>>>> Time taken: 4.512 seconds >>>>>>>> OK >>>>>>>> Time taken: 1.511 seconds >>>>>>>> OK >>>>>>>> Time taken: 0.272 seconds >>>>>>>> OK >>>>>>>> Time taken: 0.185 seconds >>>>>>>> Exception in thread "main" java.lang.NoSuchMethodError: >>>>> com.google.protobuf.Descriptors$Descriptor.getOneofs()Ljava/util/List; >>>>>>>> at >>>>> >>> com.google.protobuf.GeneratedMessageV3$FieldAccessorTable.<init>(GeneratedMessageV3.java:1704) >>>>>>>> at >>>>> org.apache.calcite.avatica.proto.Common.<clinit>(Common.java:18927) >>>>>>>> at >>>>> >>> org.apache.calcite.avatica.proto.Common$ConnectionProperties.getDescriptor(Common.java:1264) >>>>>>>> at >>>>> >>> org.apache.calcite.avatica.ConnectionPropertiesImpl.<clinit>(ConnectionPropertiesImpl.java:38) >>>>>>>> at org.apache.calcite.avatica.MetaImpl.<init>(MetaImpl.java:72) >>>>>>>> at >>>>> org.apache.calcite.jdbc.CalciteMetaImpl.<init>(CalciteMetaImpl.java:88) >>>>>>>> at org.apache.calcite.jdbc.Driver.createMeta(Driver.java:169) >>>>>>>> at >>>>> >>> org.apache.calcite.avatica.AvaticaConnection.<init>(AvaticaConnection.java:121) >>>>>>>> at >>>>> >>> org.apache.calcite.jdbc.CalciteConnectionImpl.<init>(CalciteConnectionImpl.java:113) >>>>>>>> at >>>>> >>> org.apache.calcite.jdbc.CalciteJdbc41Factory$CalciteJdbc41Connection.<init>(CalciteJdbc41Factory.java:114) >>>>>>>> at >>>>> >>> org.apache.calcite.jdbc.CalciteJdbc41Factory.newConnection(CalciteJdbc41Factory.java:59) >>>>>>>> at >>>>> >>> org.apache.calcite.jdbc.CalciteJdbc41Factory.newConnection(CalciteJdbc41Factory.java:44) >>>>>>>> at >>>>> >>> org.apache.calcite.jdbc.CalciteFactory.newConnection(CalciteFactory.java:53) >>>>>>>> at >>>>> >>> org.apache.calcite.avatica.UnregisteredDriver.connect(UnregisteredDriver.java:138) >>>>>>>> at java.sql.DriverManager.getConnection(DriverManager.java:664) >>>>>>>> at java.sql.DriverManager.getConnection(DriverManager.java:208) >>>>>>>> at >>>>> org.apache.calcite.tools.Frameworks.withPrepare(Frameworks.java:145) >>>>>>>> at >>>>> org.apache.calcite.tools.Frameworks.withPlanner(Frameworks.java:106) >>>>>>>> at >>>>> >>> org.apache.hadoop.hive.ql.parse.CalcitePlanner.logicalPlan(CalcitePlanner.java:1069) >>>>>>>> at >>>>> >>> org.apache.hadoop.hive.ql.parse.CalcitePlanner.getOptimizedAST(CalcitePlanner.java:1085) >>>>>>>> at >>>>> >>> org.apache.hadoop.hive.ql.parse.CalcitePlanner.genOPTree(CalcitePlanner.java:364) >>>>>>>> at >>>>> >>> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:11138) >>>>>>>> at >>>>> >>> org.apache.hadoop.hive.ql.parse.CalcitePlanner.analyzeInternal(CalcitePlanner.java:286) >>>>>>>> at >>>>> >>> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:258) >>>>>>>> at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:512) >>>>>>>> at >>>>> org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1317) >>>>>>>> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1457) >>>>>>>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1237) >>>>>>>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1227) >>>>>>>> at >>>>> >>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:233) >>>>>>>> at >>>>> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:184) >>>>>>>> at >>>>> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:403) >>>>>>>> at >>>>> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:336) >>>>>>>> at >>>>> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:787) >>>>>>>> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759) >>>>>>>> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:686) >>>>>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>>>>>> at >>>>> >>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >>>>>>>> at >>>>> >>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>>>>>> at java.lang.reflect.Method.invoke(Method.java:498) >>>>>>>> at org.apache.hadoop.util.RunJar.run(RunJar.java:318) >>>>>>>> at org.apache.hadoop.util.RunJar.main(RunJar.java:232) >>>>>>>> The command is: >>>>>>>> hive -e "USE default; >>>>>>> >>>>>>> >>>>>> >>>>> >>>>> >>>> >>>> -- >>>> Best regards, >>>> >>>> Shaofeng Shi 史少锋 >>> >>> >>> >> >> -- >> Best regards, >> >> Shaofeng Shi 史少锋 >> >> >> > > -- > Best regards, > > Shaofeng Shi 史少锋