Thanks ShaoFeng Shi. Its working after I renamed the hive table. Yes TRANSACTIONS is a non-reserved keyword in hive.
On Mon, Jul 11, 2016 at 8:28 PM, ShaoFeng Shi <shaofeng...@gmail.com> wrote: > (Continued. ) did you try to rename the fact table? > > Regards, > > Shaofeng Shi > > shaofeng...@gmail.com > > From Outlook Mobile > > > > > On Mon, Jul 11, 2016 at 10:55 PM +0800, "ShaoFeng Shi" < > shaofeng...@gmail.com> wrote: > > > > > > > > > > > FAILED: ParseException line 10:29 mismatched input > 'TRANSACTIONS' expecting Identifier near 'as' in table source > Interesting, is "transactions" a keyword in hive? We used to use > "fact_table", "lookup_1", "lookup_2" as the alias, but changed to use table > name for better readability; could you please open a JIRA for tracking? > To bypass it, > Regards, > > Shaofeng Shi > > shaofeng...@gmail.com > > From Outlook Mobile > > > > > On Mon, Jul 11, 2016 at 2:30 PM +0800, "Karthigeyan K" < > karthigeyan.t...@gmail.com> wrote: > > > > > > > > > > > Hi , > I was able to build the cube while using only fact table without lookup > tables. > > But Its failing when adding lookup tables. > I have one fact table and 2 lookup tables. > > The problem is AS keyword used with JOIN conditions. > Because The same query ran successfully when I run it manually in hive > after removing those table alias. > How to fix this in Kylin? > > pasted entire log below. King help is appreciated. > > Thanks, > Karthigeyan. > > > OS command error exit with 64 -- hive -e "USE default; > DROP TABLE IF EXISTS > > kylin_intermediate_transactions_demo_cube_19700101000000_2922789940817071255; > > CREATE EXTERNAL TABLE IF NOT EXISTS > > kylin_intermediate_transactions_demo_cube_19700101000000_2922789940817071255 > ( > DEFAULT_TRANSACTIONS_CUSTOMERID string > ,DEFAULT_TRANSACTIONS_PRODUCTID string > ,DEFAULT_TRANSACTIONS_PURCHASEDATE date > ,DEFAULT_TRANSACTIONS_QUANTITY int > ,DEFAULT_TRANSACTIONS_PRICE double > ,DEFAULT_TRANSACTIONS_SALE double > ) > ROW FORMAT DELIMITED FIELDS TERMINATED BY '\177' > STORED AS SEQUENCEFILE > LOCATION > '/kylin/kylin_metadata/kylin-e6854c92-1e73-41e1-b0da-0e33f18dbfec/kylin_intermediate_transactions_demo_cube_19700101000000_2922789940817071255'; > > SET dfs.replication=2; > SET dfs.block.size=33554432; > SET hive.exec.compress.output=true; > SET hive.auto.convert.join.noconditionaltask=true; > SET hive.auto.convert.join.noconditionaltask.size=300000000; > SET > mapreduce.map.output.compress.codec=org.apache.hadoop.io.compress.SnappyCodec; > SET > mapreduce.output.fileoutputformat.compress.codec=org.apache.hadoop.io.compress.SnappyCodec; > SET hive.merge.mapfiles=true; > SET hive.merge.mapredfiles=true; > SET mapred.output.compression.type=BLOCK; > SET hive.merge.size.per.task=256000000; > SET hive.support.concurrency=false; > SET mapreduce.job.split.metainfo.maxsize=-1; > > INSERT OVERWRITE TABLE > > kylin_intermediate_transactions_demo_cube_19700101000000_2922789940817071255 > SELECT > TRANSACTIONS.CUSTOMERID > ,TRANSACTIONS.PRODUCTID > ,TRANSACTIONS.PURCHASEDATE > ,TRANSACTIONS.QUANTITY > ,TRANSACTIONS.PRICE > ,TRANSACTIONS.SALE > FROM DEFAULT.TRANSACTIONS as TRANSACTIONS > LEFT JOIN DEFAULT.PRODUCT as PRODUCT > ON TRANSACTIONS.PRODUCTID = PRODUCT.PRODUCTID > LEFT JOIN DEFAULT.CUSTOMER as CUSTOMER > ON TRANSACTIONS.CUSTOMERID = CUSTOMER.CUSTOMERID > ; > > " > SLF4J: Class path contains multiple SLF4J bindings. > SLF4J: Found binding in > > [jar:file:/usr/hdp/2.3.2.0-2950/hadoop/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: Found binding in > > [jar:file:/usr/hdp/2.3.2.1-12/spark-1.5.2-bin-hadoop2.6/lib/spark-assembly-1.5.2-hadoop2.6.0.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an > explanation. > SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] > WARNING: Use "yarn jar" to launch YARN applications. > SLF4J: Class path contains multiple SLF4J bindings. > SLF4J: Found binding in > > [jar:file:/usr/hdp/2.3.2.0-2950/hadoop/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: Found binding in > > [jar:file:/usr/hdp/2.3.2.1-12/spark-1.5.2-bin-hadoop2.6/lib/spark-assembly-1.5.2-hadoop2.6.0.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an > explanation. > SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] > 16/07/09 17:09:08 WARN conf.HiveConf: HiveConf of name > hive.optimize.mapjoin.mapreduce does not exist > 16/07/09 17:09:08 WARN conf.HiveConf: HiveConf of name hive.heapsize > does not exist > 16/07/09 17:09:08 WARN conf.HiveConf: HiveConf of name > hive.metastore.local does not exist > 16/07/09 17:09:08 WARN conf.HiveConf: HiveConf of name > hive.auto.convert.sortmerge.join.noconditionaltask does not exist > ivysettings.xml file not found in HIVE_HOME or > > HIVE_CONF_DIR,file:/usr/hdp/2.3.2.0-2950/hadoop/lib/hadoop-lzo-0.6.0.2.3.2.0-2950-sources.jar!/ivysettings.xml > will be used > > Logging initialized using configuration in > > jar:file:/usr/hdp/2.3.2.0-2950/hive/lib/hive-common-1.2.1.2.3.2.0-2950.jar!/hive-log4j.properties > OK > Time taken: 2.28 seconds > OK > Time taken: 0.497 seconds > OK > Time taken: 0.527 seconds > MismatchedTokenException(262!=26) > at > org.antlr.runtime.BaseRecognizer.recoverFromMismatchedToken(BaseRecognizer.java:617) > at org.antlr.runtime.BaseRecognizer.match(BaseRecognizer.java:115) > at > org.apache.hadoop.hive.ql.parse.HiveParser_FromClauseParser.tableSource(HiveParser_FromClauseParser.java:4608) > at > org.apache.hadoop.hive.ql.parse.HiveParser_FromClauseParser.fromSource(HiveParser_FromClauseParser.java:3729) > at > org.apache.hadoop.hive.ql.parse.HiveParser_FromClauseParser.joinSource(HiveParser_FromClauseParser.java:1873) > at > org.apache.hadoop.hive.ql.parse.HiveParser_FromClauseParser.fromClause(HiveParser_FromClauseParser.java:1518) > at > org.apache.hadoop.hive.ql.parse.HiveParser.fromClause(HiveParser.java:45857) > at > org.apache.hadoop.hive.ql.parse.HiveParser.selectStatement(HiveParser.java:41519) > at > org.apache.hadoop.hive.ql.parse.HiveParser.regularBody(HiveParser.java:41233) > at > org.apache.hadoop.hive.ql.parse.HiveParser.queryStatementExpressionBody(HiveParser.java:40416) > at > org.apache.hadoop.hive.ql.parse.HiveParser.queryStatementExpression(HiveParser.java:40286) > at > org.apache.hadoop.hive.ql.parse.HiveParser.execStatement(HiveParser.java:1593) > at > org.apache.hadoop.hive.ql.parse.HiveParser.statement(HiveParser.java:1112) > at > org.apache.hadoop.hive.ql.parse.ParseDriver.parse(ParseDriver.java:202) > at > org.apache.hadoop.hive.ql.parse.ParseDriver.parse(ParseDriver.java:166) > at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:396) > at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:308) > at > org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1122) > at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1170) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1059) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1049) > at > org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:213) > at > org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:165) > at > org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376) > at > org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:311) > at > org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:708) > at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:681) > at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:621) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:606) > at org.apache.hadoop.util.RunJar.run(RunJar.java:221) > at org.apache.hadoop.util.RunJar.main(RunJar.java:136) > FAILED: ParseException line 10:29 mismatched input 'TRANSACTIONS' > expecting Identifier near 'as' in table source > > > > > > > > > > >