[jira] [Assigned] (KYLIN-3709) Pass the wrong parameter to addResource function, kylin can not load the configuration file hbase.hdfs.xml .

2018-12-04 Thread Shaofeng SHI (JIRA)


 [ 
https://issues.apache.org/jira/browse/KYLIN-3709?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shaofeng SHI reassigned KYLIN-3709:
---

Assignee: Chao Long  (was: Shaofeng SHI)

Assign to Chao for this issue, the previous change was made by him.

> Pass the wrong parameter to addResource function, kylin can not load the 
> configuration file hbase.hdfs.xml .
> 
>
> Key: KYLIN-3709
> URL: https://issues.apache.org/jira/browse/KYLIN-3709
> Project: Kylin
>  Issue Type: Bug
>  Components: Job Engine
>Affects Versions: v2.5.1, v2.5.2
>Reporter: Hongtao He
>Assignee: Chao Long
>Priority: Critical
> Attachments: 
> method-1-Pass-the-wrong-parameter-to-addResource-function-kyl.patch, 
> method-2-Pass-the-wrong-parameter-to-addResource-function-kyl.patch
>
>
> When calling Configuration.addResource, KYLIN-3648 use a Path as the 
> parameter instead of a string in Kylin's HBaseConnection.
> {code:java}
> String hdfsConfigFile = 
> KylinConfig.getInstanceFromEnv().getHBaseClusterHDFSConfigFile();
> if (hdfsConfigFile == null || hdfsConfigFile.isEmpty()) {
> return;
> }
> Configuration hdfsConf = new Configuration(false);
> hdfsConf.addResource(new Path(hdfsConfigFile));
> {code}
> Use a Path as the parameter of Configuration.addResource is better. 
> Unfortunately,  the parameter which passed to the addResource function is 
> wrong. The addResource function only accepts absolute paths, but the 
> parameter is just a filename. For example, the value of hdfsConfigFile is 
> "hbase.hdfs.xml", so addResource function will not work . The end result is 
> that kylin can not load the hbase configuration file hbase.hdfs.xml .
> There are two ways to fix this bug, and I think method 1 is better.
> Method-1.revert the code
> {code:java}
> String hdfsConfigFile = 
> KylinConfig.getInstanceFromEnv().getHBaseClusterHDFSConfigFile();
> if (hdfsConfigFile == null || hdfsConfigFile.isEmpty()) {
> return;
> }
> Configuration hdfsConf = new Configuration(false);
> hdfsConf.addResource(hdfsConfigFile);{code}
> Method-2.Get the absolute path of the configuration file
> {code:java}
> String hdfsConfigFile = 
> KylinConfig.getInstanceFromEnv().getHBaseClusterHDFSConfigFile();
> if (hdfsConfigFile == null || hdfsConfigFile.isEmpty()) {
> return;
> }
> Configuration hdfsConf = new Configuration(false);
> String hbaseHdfsConfigPath = System.getProperty("user.dir") + "/../conf/" + 
> hdfsConfigFile;
> hdfsConf.addResource(new Path(hbaseHdfsConfigPath));
> {code}
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (KYLIN-3709) Pass the wrong parameter to addResource function, kylin can not load the configuration file hbase.hdfs.xml .

2018-12-04 Thread Shaofeng SHI (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3709?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16709649#comment-16709649
 ] 

Shaofeng SHI commented on KYLIN-3709:
-

Hi Hongtao, we only merged KYLIN-3648 on master branch; Kylin 2.5.x doesn't 
have the code change of KYLIN-3648 , so 2.5.0 to 2.5.2 not affected, you can 
double check it.

 

If "kylin.storage.hbase.cluster-hdfs-config-file" be set to an absolute path, 
is there such problem?

> Pass the wrong parameter to addResource function, kylin can not load the 
> configuration file hbase.hdfs.xml .
> 
>
> Key: KYLIN-3709
> URL: https://issues.apache.org/jira/browse/KYLIN-3709
> Project: Kylin
>  Issue Type: Bug
>  Components: Job Engine
>Affects Versions: v2.5.1, v2.5.2
>Reporter: Hongtao He
>Assignee: Shaofeng SHI
>Priority: Critical
> Attachments: 
> method-1-Pass-the-wrong-parameter-to-addResource-function-kyl.patch, 
> method-2-Pass-the-wrong-parameter-to-addResource-function-kyl.patch
>
>
> When calling Configuration.addResource, KYLIN-3648 use a Path as the 
> parameter instead of a string in Kylin's HBaseConnection.
> {code:java}
> String hdfsConfigFile = 
> KylinConfig.getInstanceFromEnv().getHBaseClusterHDFSConfigFile();
> if (hdfsConfigFile == null || hdfsConfigFile.isEmpty()) {
> return;
> }
> Configuration hdfsConf = new Configuration(false);
> hdfsConf.addResource(new Path(hdfsConfigFile));
> {code}
> Use a Path as the parameter of Configuration.addResource is better. 
> Unfortunately,  the parameter which passed to the addResource function is 
> wrong. The addResource function only accepts absolute paths, but the 
> parameter is just a filename. For example, the value of hdfsConfigFile is 
> "hbase.hdfs.xml", so addResource function will not work . The end result is 
> that kylin can not load the hbase configuration file hbase.hdfs.xml .
> There are two ways to fix this bug, and I think method 1 is better.
> Method-1.revert the code
> {code:java}
> String hdfsConfigFile = 
> KylinConfig.getInstanceFromEnv().getHBaseClusterHDFSConfigFile();
> if (hdfsConfigFile == null || hdfsConfigFile.isEmpty()) {
> return;
> }
> Configuration hdfsConf = new Configuration(false);
> hdfsConf.addResource(hdfsConfigFile);{code}
> Method-2.Get the absolute path of the configuration file
> {code:java}
> String hdfsConfigFile = 
> KylinConfig.getInstanceFromEnv().getHBaseClusterHDFSConfigFile();
> if (hdfsConfigFile == null || hdfsConfigFile.isEmpty()) {
> return;
> }
> Configuration hdfsConf = new Configuration(false);
> String hbaseHdfsConfigPath = System.getProperty("user.dir") + "/../conf/" + 
> hdfsConfigFile;
> hdfsConf.addResource(new Path(hbaseHdfsConfigPath));
> {code}
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (KYLIN-3406) When the /execute_output/ metadata file sinking to HDFS is deleted, the monitoring page does not display any tasks.

2018-12-04 Thread XiaoXiang Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3406?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16709638#comment-16709638
 ] 

XiaoXiang Yu edited comment on KYLIN-3406 at 12/5/18 5:57 AM:
--

[5261]I have reproduced error as reporter describe.

When any single file of Kylin's metadata miss in HDFS, the "Monitor" Page will 
failed to display all job info and kylin process will throw 
FileNotFoundException as above.


The better way should be miss file should only affect specific task.  And if 
only execute output miss, task should be success, otherwise task should be 
marked as failed.


was (Author: hit_lacus):
[5261]I have reproduced error as reporter describe.

When any single file of Kylin's metadata miss in HDFS, the "Monitor" Page will 
failed to display all job info and kylin process will throw 
FileNotFoundException as above.


The better way should be miss file should only affect specific item.  And if 
only execute output miss, task should be success, otherwise task should be 
marked as failed.

> When the /execute_output/ metadata file sinking to HDFS is deleted, the 
> monitoring page does not display any tasks.
> ---
>
> Key: KYLIN-3406
> URL: https://issues.apache.org/jira/browse/KYLIN-3406
> Project: Kylin
>  Issue Type: Bug
>  Components: Metadata
>Affects Versions: v2.3.1
>Reporter: yujie.sun
>Assignee: XiaoXiang Yu
>Priority: Major
> Fix For: v2.6.0
>
> Attachments: image-2018-12-05-13-22-42-456.png
>
>
> *{color:#24292e}Issue description:{color}*
> When a single file of Hbase metadata is greater than a certain value, the 
> file is sunk into HDFS for storage. When files in execute_output/ are deleted 
> in HDFS, the monitoring page will not display any tasks.
> *{color:#24292e}Reproduction step:{color}*
> {color:#24292e}1.Use Hbase as metadata storage{color}
> {color:#24292e}2.Modify the hbase configuration ,configure 
> "hbase.client.keyvalue.maxsize" into a small value, such as 524288.{color}
> {color:#24292e}3.Create a new build task, after the build task finished,find 
> the /execute_output hbase metadata path on HDFS, 
> /\{working_dir}/\{metadata_url}/resources/execute_output/..., and delete one 
> of the files.{color}
> {color:#24292e}4.At this time,{color:#33}the monitoring page will not 
> display any tasks.{color}{color}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (KYLIN-3406) When the /execute_output/ metadata file sinking to HDFS is deleted, the monitoring page does not display any tasks.

2018-12-04 Thread XiaoXiang Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3406?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16709638#comment-16709638
 ] 

XiaoXiang Yu commented on KYLIN-3406:
-

[5261]I have reproduced error as reporter describe.

When any single file of Kylin's metadata miss in HDFS, the "Monitor" Page will 
failed to display all job info and kylin process will throw 
FileNotFoundException as above.


The better way should be miss file should only affect specific item.  And if 
only execute output miss, task should be success, otherwise task should be 
marked as failed.

> When the /execute_output/ metadata file sinking to HDFS is deleted, the 
> monitoring page does not display any tasks.
> ---
>
> Key: KYLIN-3406
> URL: https://issues.apache.org/jira/browse/KYLIN-3406
> Project: Kylin
>  Issue Type: Bug
>  Components: Metadata
>Affects Versions: v2.3.1
>Reporter: yujie.sun
>Assignee: XiaoXiang Yu
>Priority: Major
> Fix For: v2.6.0
>
> Attachments: image-2018-12-05-13-22-42-456.png
>
>
> *{color:#24292e}Issue description:{color}*
> When a single file of Hbase metadata is greater than a certain value, the 
> file is sunk into HDFS for storage. When files in execute_output/ are deleted 
> in HDFS, the monitoring page will not display any tasks.
> *{color:#24292e}Reproduction step:{color}*
> {color:#24292e}1.Use Hbase as metadata storage{color}
> {color:#24292e}2.Modify the hbase configuration ,configure 
> "hbase.client.keyvalue.maxsize" into a small value, such as 524288.{color}
> {color:#24292e}3.Create a new build task, after the build task finished,find 
> the /execute_output hbase metadata path on HDFS, 
> /\{working_dir}/\{metadata_url}/resources/execute_output/..., and delete one 
> of the files.{color}
> {color:#24292e}4.At this time,{color:#33}the monitoring page will not 
> display any tasks.{color}{color}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (KYLIN-3695) Error while creating hive table through Kylin build cube with mysql imported tables

2018-12-04 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3695?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16709636#comment-16709636
 ] 

ASF GitHub Bot commented on KYLIN-3695:
---

codecov-io commented on issue #371: KYLIN-3695, fix lose decimal scale value in 
column type decimal(a, b)
URL: https://github.com/apache/kylin/pull/371#issuecomment-444367864
 
 
   # [Codecov](https://codecov.io/gh/apache/kylin/pull/371?src=pr&el=h1) Report
   > Merging [#371](https://codecov.io/gh/apache/kylin/pull/371?src=pr&el=desc) 
into 
[master](https://codecov.io/gh/apache/kylin/commit/be5df4a8bb6939c0c62c6135c4312fe3ad56775d?src=pr&el=desc)
 will **decrease** coverage by `0.01%`.
   > The diff coverage is `n/a`.
   
   [![Impacted file tree 
graph](https://codecov.io/gh/apache/kylin/pull/371/graphs/tree.svg?width=650&token=JawVgbgsVo&height=150&src=pr)](https://codecov.io/gh/apache/kylin/pull/371?src=pr&el=tree)
   
   ```diff
   @@ Coverage Diff  @@
   ## master #371  +/-   ##
   
   - Coverage 23.64%   23.63%   -0.02% 
   + Complexity 4838 4837   -1 
   
 Files  1131 1131  
 Lines 6819368193  
 Branches   9684 9684  
   
   - Hits  1612416116   -8 
   - Misses5050350509   +6 
   - Partials   1566 1568   +2
   ```
   
   
   | [Impacted 
Files](https://codecov.io/gh/apache/kylin/pull/371?src=pr&el=tree) | Coverage Δ 
| Complexity Δ | |
   |---|---|---|---|
   | 
[...he/kylin/dict/lookup/cache/RocksDBLookupTable.java](https://codecov.io/gh/apache/kylin/pull/371/diff?src=pr&el=tree#diff-Y29yZS1kaWN0aW9uYXJ5L3NyYy9tYWluL2phdmEvb3JnL2FwYWNoZS9reWxpbi9kaWN0L2xvb2t1cC9jYWNoZS9Sb2Nrc0RCTG9va3VwVGFibGUuamF2YQ==)
 | `72.97% <0%> (-5.41%)` | `6% <0%> (-1%)` | |
   | 
[.../apache/kylin/cube/cuboid/TreeCuboidScheduler.java](https://codecov.io/gh/apache/kylin/pull/371/diff?src=pr&el=tree#diff-Y29yZS1jdWJlL3NyYy9tYWluL2phdmEvb3JnL2FwYWNoZS9reWxpbi9jdWJlL2N1Ym9pZC9UcmVlQ3Vib2lkU2NoZWR1bGVyLmphdmE=)
 | `63.84% <0%> (-2.31%)` | `0% <0%> (ø)` | |
   | 
[...rg/apache/kylin/cube/inmemcubing/MemDiskStore.java](https://codecov.io/gh/apache/kylin/pull/371/diff?src=pr&el=tree#diff-Y29yZS1jdWJlL3NyYy9tYWluL2phdmEvb3JnL2FwYWNoZS9reWxpbi9jdWJlL2lubWVtY3ViaW5nL01lbURpc2tTdG9yZS5qYXZh)
 | `70.21% <0%> (-0.92%)` | `7% <0%> (ø)` | |
   
   --
   
   [Continue to review full report at 
Codecov](https://codecov.io/gh/apache/kylin/pull/371?src=pr&el=continue).
   > **Legend** - [Click here to learn 
more](https://docs.codecov.io/docs/codecov-delta)
   > `Δ = absolute  (impact)`, `ø = not affected`, `? = missing data`
   > Powered by 
[Codecov](https://codecov.io/gh/apache/kylin/pull/371?src=pr&el=footer). Last 
update 
[be5df4a...a073222](https://codecov.io/gh/apache/kylin/pull/371?src=pr&el=lastupdated).
 Read the [comment docs](https://docs.codecov.io/docs/pull-request-comments).
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Error while creating hive table through Kylin build cube with mysql imported 
> tables
> ---
>
> Key: KYLIN-3695
> URL: https://issues.apache.org/jira/browse/KYLIN-3695
> Project: Kylin
>  Issue Type: Bug
>  Components: Integration
>Affects Versions: v2.5.0
>Reporter: Gaurav Rawat
>Assignee: XiaoXiang Yu
>Priority: Minor
> Attachments: image-2018-12-04-15-53-49-983.png
>
>
> HI I am trying to build a cube with Kylin, the data gets souced fine from 
> sqoop but the next step for creating hive tables fails . Looking at the 
> command being fired it looks weird as the create statement looks good to me .
> I think the issue is with DOUBLE types as when I remove the same the create 
> statement works fine . Can someone please help .
> I am using the stack in AWS EMR, kylin 2.5 hive 2.3.0
> The errors logs with commands as as below, the table is a msql table which 
> had columns with DOUBLE type
> Command
> {code:java}
> hive -e "USE default; DROP TABLE IF EXISTS 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368; 
> CREATE EXTERNAL TABLE IF NOT EXISTS 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368 ( 
> HOLDINGS_STOCK_INVESTOR_ID string ,STOCK_INVESTORS_CHANNEL string 
> ,STOCK_STOCK_ID string ,STOCK_DOMICILE string ,STOCK

[GitHub] codecov-io commented on issue #371: KYLIN-3695, fix lose decimal scale value in column type decimal(a, b)

2018-12-04 Thread GitBox
codecov-io commented on issue #371: KYLIN-3695, fix lose decimal scale value in 
column type decimal(a, b)
URL: https://github.com/apache/kylin/pull/371#issuecomment-444367864
 
 
   # [Codecov](https://codecov.io/gh/apache/kylin/pull/371?src=pr&el=h1) Report
   > Merging [#371](https://codecov.io/gh/apache/kylin/pull/371?src=pr&el=desc) 
into 
[master](https://codecov.io/gh/apache/kylin/commit/be5df4a8bb6939c0c62c6135c4312fe3ad56775d?src=pr&el=desc)
 will **decrease** coverage by `0.01%`.
   > The diff coverage is `n/a`.
   
   [![Impacted file tree 
graph](https://codecov.io/gh/apache/kylin/pull/371/graphs/tree.svg?width=650&token=JawVgbgsVo&height=150&src=pr)](https://codecov.io/gh/apache/kylin/pull/371?src=pr&el=tree)
   
   ```diff
   @@ Coverage Diff  @@
   ## master #371  +/-   ##
   
   - Coverage 23.64%   23.63%   -0.02% 
   + Complexity 4838 4837   -1 
   
 Files  1131 1131  
 Lines 6819368193  
 Branches   9684 9684  
   
   - Hits  1612416116   -8 
   - Misses5050350509   +6 
   - Partials   1566 1568   +2
   ```
   
   
   | [Impacted 
Files](https://codecov.io/gh/apache/kylin/pull/371?src=pr&el=tree) | Coverage Δ 
| Complexity Δ | |
   |---|---|---|---|
   | 
[...he/kylin/dict/lookup/cache/RocksDBLookupTable.java](https://codecov.io/gh/apache/kylin/pull/371/diff?src=pr&el=tree#diff-Y29yZS1kaWN0aW9uYXJ5L3NyYy9tYWluL2phdmEvb3JnL2FwYWNoZS9reWxpbi9kaWN0L2xvb2t1cC9jYWNoZS9Sb2Nrc0RCTG9va3VwVGFibGUuamF2YQ==)
 | `72.97% <0%> (-5.41%)` | `6% <0%> (-1%)` | |
   | 
[.../apache/kylin/cube/cuboid/TreeCuboidScheduler.java](https://codecov.io/gh/apache/kylin/pull/371/diff?src=pr&el=tree#diff-Y29yZS1jdWJlL3NyYy9tYWluL2phdmEvb3JnL2FwYWNoZS9reWxpbi9jdWJlL2N1Ym9pZC9UcmVlQ3Vib2lkU2NoZWR1bGVyLmphdmE=)
 | `63.84% <0%> (-2.31%)` | `0% <0%> (ø)` | |
   | 
[...rg/apache/kylin/cube/inmemcubing/MemDiskStore.java](https://codecov.io/gh/apache/kylin/pull/371/diff?src=pr&el=tree#diff-Y29yZS1jdWJlL3NyYy9tYWluL2phdmEvb3JnL2FwYWNoZS9reWxpbi9jdWJlL2lubWVtY3ViaW5nL01lbURpc2tTdG9yZS5qYXZh)
 | `70.21% <0%> (-0.92%)` | `7% <0%> (ø)` | |
   
   --
   
   [Continue to review full report at 
Codecov](https://codecov.io/gh/apache/kylin/pull/371?src=pr&el=continue).
   > **Legend** - [Click here to learn 
more](https://docs.codecov.io/docs/codecov-delta)
   > `Δ = absolute  (impact)`, `ø = not affected`, `? = missing data`
   > Powered by 
[Codecov](https://codecov.io/gh/apache/kylin/pull/371?src=pr&el=footer). Last 
update 
[be5df4a...a073222](https://codecov.io/gh/apache/kylin/pull/371?src=pr&el=lastupdated).
 Read the [comment docs](https://docs.codecov.io/docs/pull-request-comments).
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (KYLIN-3695) Error while creating hive table through Kylin build cube with mysql imported tables

2018-12-04 Thread ASF subversion and git services (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3695?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16709630#comment-16709630
 ] 

ASF subversion and git services commented on KYLIN-3695:


Commit 7f672b074a068d20779d33889f1cb63b0da172a7 in kylin's branch 
refs/heads/master from Emiya0306
[ https://gitbox.apache.org/repos/asf?p=kylin.git;h=7f672b0 ]

KYLIN-3695, fix lose decimal scale value in column type decimal(a, b)


> Error while creating hive table through Kylin build cube with mysql imported 
> tables
> ---
>
> Key: KYLIN-3695
> URL: https://issues.apache.org/jira/browse/KYLIN-3695
> Project: Kylin
>  Issue Type: Bug
>  Components: Integration
>Affects Versions: v2.5.0
>Reporter: Gaurav Rawat
>Assignee: XiaoXiang Yu
>Priority: Minor
> Attachments: image-2018-12-04-15-53-49-983.png
>
>
> HI I am trying to build a cube with Kylin, the data gets souced fine from 
> sqoop but the next step for creating hive tables fails . Looking at the 
> command being fired it looks weird as the create statement looks good to me .
> I think the issue is with DOUBLE types as when I remove the same the create 
> statement works fine . Can someone please help .
> I am using the stack in AWS EMR, kylin 2.5 hive 2.3.0
> The errors logs with commands as as below, the table is a msql table which 
> had columns with DOUBLE type
> Command
> {code:java}
> hive -e "USE default; DROP TABLE IF EXISTS 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368; 
> CREATE EXTERNAL TABLE IF NOT EXISTS 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368 ( 
> HOLDINGS_STOCK_INVESTOR_ID string ,STOCK_INVESTORS_CHANNEL string 
> ,STOCK_STOCK_ID string ,STOCK_DOMICILE string ,STOCK_STOCK_NM string 
> ,STOCK_APPROACH string ,STOCK_STOCK_TYP string ,INVESTOR_ID string 
> ,INVESTOR_NM string ,INVESTOR_DOMICILE_CNTRY string ,CLIENT_NM string 
> ,INVESTOR_HOLDINGS_GROSS_ASSETS_USD double(22) 
> ,INVESTOR_HOLDINGS_NET_ASSETS_USD double(22) ) ROW FORMAT DELIMITED FIELDS 
> TERMINATED BY '|' STORED AS TEXTFILE LOCATION 
> 's3://wfg1tst-models/kylin/kylin_metadata/kylin-4ae3b18b-831b-da66-eb8c-7318245c4448/kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368';
>  ALTER TABLE 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368 SET 
> TBLPROPERTIES('auto.purge'='true'); " --hiveconf hive.merge.mapredfiles=false 
> --hiveconf hive.auto.convert.join=true --hiveconf dfs.replication=2 
> --hiveconf hive.exec.compress.output=true --hiveconf 
> hive.auto.convert.join.noconditionaltask=true --hiveconf 
> mapreduce.job.split.metainfo.maxsize=-1 --hiveconf hive.merge.mapfiles=false 
> --hiveconf hive.auto.convert.join.noconditionaltask.size=1 --hiveconf 
> hive.stats.autogather=true{code}
> Error is as below
> {code:java}
> OK Time taken: 1.315 seconds OK Time taken: 0.09 seconds 
> MismatchedTokenException(334!=347) at 
> org.antlr.runtime.BaseRecognizer.recoverFromMismatchedToken(BaseRecognizer.java:617)
>  at org.antlr.runtime.BaseRecognizer.match(BaseRecognizer.java:115) at 
> org.apache.hadoop.hive.ql.parse.HiveParser.createTableStatement(HiveParser.java:6179)
>  at 
> org.apache.hadoop.hive.ql.parse.HiveParser.ddlStatement(HiveParser.java:3808) 
> at 
> org.apache.hadoop.hive.ql.parse.HiveParser.execStatement(HiveParser.java:2382)
>  at 
> org.apache.hadoop.hive.ql.parse.HiveParser.statement(HiveParser.java:1333) at 
> org.apache.hadoop.hive.ql.parse.ParseDriver.parse(ParseDriver.java:204) at 
> org.apache.hadoop.hive.ql.parse.ParseUtils.parse(ParseUtils.java:77) at 
> org.apache.hadoop.hive.ql.parse.ParseUtils.parse(ParseUtils.java:70) at 
> org.apache.hadoop.hive.ql.Driver.compile(Driver.java:468) at 
> org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1316) at 
> org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1456) at 
> org.apache.hadoop.hive.ql.Driver.run(Driver.java:1236) at 
> org.apache.hadoop.hive.ql.Driver.run(Driver.java:1226) at 
> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:233) at 
> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:184) at 
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:403) at 
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:336) at 
> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:787) at 
> org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759) at 
> org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:686) at 
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethod

[jira] [Commented] (KYLIN-3695) Error while creating hive table through Kylin build cube with mysql imported tables

2018-12-04 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3695?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16709629#comment-16709629
 ] 

ASF GitHub Bot commented on KYLIN-3695:
---

luguosheng1314 closed pull request #371: KYLIN-3695, fix lose decimal scale 
value in column type decimal(a, b)
URL: https://github.com/apache/kylin/pull/371
 
 
   

This is a PR merged from a forked repository.
As GitHub hides the original diff on merge, it is displayed below for
the sake of provenance:

As this is a foreign pull request (from a fork), the diff is supplied
below (as it won't show otherwise due to GitHub magic):

diff --git a/webapp/app/js/controllers/cubeMeasures.js 
b/webapp/app/js/controllers/cubeMeasures.js
index 7beb528d6b..f1821dda87 100644
--- a/webapp/app/js/controllers/cubeMeasures.js
+++ b/webapp/app/js/controllers/cubeMeasures.js
@@ -465,10 +465,10 @@ KylinApp.controller('CubeMeasuresCtrl', function ($scope, 
$modal,MetaModel,cubes
   return 'bigint';
 } else {
  if(colType.indexOf('decimal') != -1) {
-var returnRegex = new 
RegExp('(\\w+)(?:\\((\\w+?)(?:\\,(\\w+?))?\\))?')
-var returnValue = returnRegex.exec(colType)
-var precision = 19
-var scale = returnValue[3]
+var returnRegex = new 
RegExp('(\\w+)(?:\\((\\w+?)(?:\\,(\\w+?))?\\))?');
+var returnValue = returnRegex.exec(colType);
+var precision = 19;
+var scale = returnValue[3] || 0;
 return 'decimal(' + precision + ',' + scale + ')';
   }else{
 return colType;


 


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Error while creating hive table through Kylin build cube with mysql imported 
> tables
> ---
>
> Key: KYLIN-3695
> URL: https://issues.apache.org/jira/browse/KYLIN-3695
> Project: Kylin
>  Issue Type: Bug
>  Components: Integration
>Affects Versions: v2.5.0
>Reporter: Gaurav Rawat
>Assignee: XiaoXiang Yu
>Priority: Minor
> Attachments: image-2018-12-04-15-53-49-983.png
>
>
> HI I am trying to build a cube with Kylin, the data gets souced fine from 
> sqoop but the next step for creating hive tables fails . Looking at the 
> command being fired it looks weird as the create statement looks good to me .
> I think the issue is with DOUBLE types as when I remove the same the create 
> statement works fine . Can someone please help .
> I am using the stack in AWS EMR, kylin 2.5 hive 2.3.0
> The errors logs with commands as as below, the table is a msql table which 
> had columns with DOUBLE type
> Command
> {code:java}
> hive -e "USE default; DROP TABLE IF EXISTS 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368; 
> CREATE EXTERNAL TABLE IF NOT EXISTS 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368 ( 
> HOLDINGS_STOCK_INVESTOR_ID string ,STOCK_INVESTORS_CHANNEL string 
> ,STOCK_STOCK_ID string ,STOCK_DOMICILE string ,STOCK_STOCK_NM string 
> ,STOCK_APPROACH string ,STOCK_STOCK_TYP string ,INVESTOR_ID string 
> ,INVESTOR_NM string ,INVESTOR_DOMICILE_CNTRY string ,CLIENT_NM string 
> ,INVESTOR_HOLDINGS_GROSS_ASSETS_USD double(22) 
> ,INVESTOR_HOLDINGS_NET_ASSETS_USD double(22) ) ROW FORMAT DELIMITED FIELDS 
> TERMINATED BY '|' STORED AS TEXTFILE LOCATION 
> 's3://wfg1tst-models/kylin/kylin_metadata/kylin-4ae3b18b-831b-da66-eb8c-7318245c4448/kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368';
>  ALTER TABLE 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368 SET 
> TBLPROPERTIES('auto.purge'='true'); " --hiveconf hive.merge.mapredfiles=false 
> --hiveconf hive.auto.convert.join=true --hiveconf dfs.replication=2 
> --hiveconf hive.exec.compress.output=true --hiveconf 
> hive.auto.convert.join.noconditionaltask=true --hiveconf 
> mapreduce.job.split.metainfo.maxsize=-1 --hiveconf hive.merge.mapfiles=false 
> --hiveconf hive.auto.convert.join.noconditionaltask.size=1 --hiveconf 
> hive.stats.autogather=true{code}
> Error is as below
> {code:java}
> OK Time taken: 1.315 seconds OK Time taken: 0.09 seconds 
> MismatchedTokenException(334!=347) at 
> org.antlr.runtime.BaseRecognizer.recoverFromMismatchedToken(BaseRecognizer.java:617)
>  at org.antlr.runtime.BaseRecognizer.match(BaseRecognizer.java:115) at 
> org.apache.hadoop.hive.ql.parse.HiveParser.createTableStatement(HiveParser.java:6179)
>  at 
> org.apache.hadoop.hive.ql.parse.HiveParser.ddlStatement(HiveParser.java:380

[GitHub] luguosheng1314 closed pull request #371: KYLIN-3695, fix lose decimal scale value in column type decimal(a, b)

2018-12-04 Thread GitBox
luguosheng1314 closed pull request #371: KYLIN-3695, fix lose decimal scale 
value in column type decimal(a, b)
URL: https://github.com/apache/kylin/pull/371
 
 
   

This is a PR merged from a forked repository.
As GitHub hides the original diff on merge, it is displayed below for
the sake of provenance:

As this is a foreign pull request (from a fork), the diff is supplied
below (as it won't show otherwise due to GitHub magic):

diff --git a/webapp/app/js/controllers/cubeMeasures.js 
b/webapp/app/js/controllers/cubeMeasures.js
index 7beb528d6b..f1821dda87 100644
--- a/webapp/app/js/controllers/cubeMeasures.js
+++ b/webapp/app/js/controllers/cubeMeasures.js
@@ -465,10 +465,10 @@ KylinApp.controller('CubeMeasuresCtrl', function ($scope, 
$modal,MetaModel,cubes
   return 'bigint';
 } else {
  if(colType.indexOf('decimal') != -1) {
-var returnRegex = new 
RegExp('(\\w+)(?:\\((\\w+?)(?:\\,(\\w+?))?\\))?')
-var returnValue = returnRegex.exec(colType)
-var precision = 19
-var scale = returnValue[3]
+var returnRegex = new 
RegExp('(\\w+)(?:\\((\\w+?)(?:\\,(\\w+?))?\\))?');
+var returnValue = returnRegex.exec(colType);
+var precision = 19;
+var scale = returnValue[3] || 0;
 return 'decimal(' + precision + ',' + scale + ')';
   }else{
 return colType;


 


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (KYLIN-3695) Error while creating hive table through Kylin build cube with mysql imported tables

2018-12-04 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3695?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16709619#comment-16709619
 ] 

ASF GitHub Bot commented on KYLIN-3695:
---

Emiya0306 opened a new pull request #371: KYLIN-3695, fix lose decimal scale 
value in column type decimal(a, b)
URL: https://github.com/apache/kylin/pull/371
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Error while creating hive table through Kylin build cube with mysql imported 
> tables
> ---
>
> Key: KYLIN-3695
> URL: https://issues.apache.org/jira/browse/KYLIN-3695
> Project: Kylin
>  Issue Type: Bug
>  Components: Integration
>Affects Versions: v2.5.0
>Reporter: Gaurav Rawat
>Assignee: XiaoXiang Yu
>Priority: Minor
> Attachments: image-2018-12-04-15-53-49-983.png
>
>
> HI I am trying to build a cube with Kylin, the data gets souced fine from 
> sqoop but the next step for creating hive tables fails . Looking at the 
> command being fired it looks weird as the create statement looks good to me .
> I think the issue is with DOUBLE types as when I remove the same the create 
> statement works fine . Can someone please help .
> I am using the stack in AWS EMR, kylin 2.5 hive 2.3.0
> The errors logs with commands as as below, the table is a msql table which 
> had columns with DOUBLE type
> Command
> {code:java}
> hive -e "USE default; DROP TABLE IF EXISTS 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368; 
> CREATE EXTERNAL TABLE IF NOT EXISTS 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368 ( 
> HOLDINGS_STOCK_INVESTOR_ID string ,STOCK_INVESTORS_CHANNEL string 
> ,STOCK_STOCK_ID string ,STOCK_DOMICILE string ,STOCK_STOCK_NM string 
> ,STOCK_APPROACH string ,STOCK_STOCK_TYP string ,INVESTOR_ID string 
> ,INVESTOR_NM string ,INVESTOR_DOMICILE_CNTRY string ,CLIENT_NM string 
> ,INVESTOR_HOLDINGS_GROSS_ASSETS_USD double(22) 
> ,INVESTOR_HOLDINGS_NET_ASSETS_USD double(22) ) ROW FORMAT DELIMITED FIELDS 
> TERMINATED BY '|' STORED AS TEXTFILE LOCATION 
> 's3://wfg1tst-models/kylin/kylin_metadata/kylin-4ae3b18b-831b-da66-eb8c-7318245c4448/kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368';
>  ALTER TABLE 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368 SET 
> TBLPROPERTIES('auto.purge'='true'); " --hiveconf hive.merge.mapredfiles=false 
> --hiveconf hive.auto.convert.join=true --hiveconf dfs.replication=2 
> --hiveconf hive.exec.compress.output=true --hiveconf 
> hive.auto.convert.join.noconditionaltask=true --hiveconf 
> mapreduce.job.split.metainfo.maxsize=-1 --hiveconf hive.merge.mapfiles=false 
> --hiveconf hive.auto.convert.join.noconditionaltask.size=1 --hiveconf 
> hive.stats.autogather=true{code}
> Error is as below
> {code:java}
> OK Time taken: 1.315 seconds OK Time taken: 0.09 seconds 
> MismatchedTokenException(334!=347) at 
> org.antlr.runtime.BaseRecognizer.recoverFromMismatchedToken(BaseRecognizer.java:617)
>  at org.antlr.runtime.BaseRecognizer.match(BaseRecognizer.java:115) at 
> org.apache.hadoop.hive.ql.parse.HiveParser.createTableStatement(HiveParser.java:6179)
>  at 
> org.apache.hadoop.hive.ql.parse.HiveParser.ddlStatement(HiveParser.java:3808) 
> at 
> org.apache.hadoop.hive.ql.parse.HiveParser.execStatement(HiveParser.java:2382)
>  at 
> org.apache.hadoop.hive.ql.parse.HiveParser.statement(HiveParser.java:1333) at 
> org.apache.hadoop.hive.ql.parse.ParseDriver.parse(ParseDriver.java:204) at 
> org.apache.hadoop.hive.ql.parse.ParseUtils.parse(ParseUtils.java:77) at 
> org.apache.hadoop.hive.ql.parse.ParseUtils.parse(ParseUtils.java:70) at 
> org.apache.hadoop.hive.ql.Driver.compile(Driver.java:468) at 
> org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1316) at 
> org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1456) at 
> org.apache.hadoop.hive.ql.Driver.run(Driver.java:1236) at 
> org.apache.hadoop.hive.ql.Driver.run(Driver.java:1226) at 
> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:233) at 
> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:184) at 
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:403) at 
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:336) at 
> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:787) at 
> org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759) at 
> org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:686) at 
> s

[jira] [Commented] (KYLIN-3406) When the /execute_output/ metadata file sinking to HDFS is deleted, the monitoring page does not display any tasks.

2018-12-04 Thread XiaoXiang Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3406?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16709621#comment-16709621
 ] 

XiaoXiang Yu commented on KYLIN-3406:
-

{quote}KYLIN [ ERROR ] 12-05 11:01:10.223 
org.apache.kylin.job.dao.ExecutableDao.getJobOutput(ExecutableDao.java:355) 
from http-bio-7070-exec-1
 > error get job output id:3b27b7df-aece-ba28-e986-3758fe8b9a85-10

java.io.IOException: Failed to read resource at 
/execute_output/3b27b7df-aece-ba28-e986-3758fe8b9a85-10
 at 
org.apache.kylin.storage.hbase.HBaseResourceStore.getInputStream(HBaseResourceStore.java:266)
 at 
org.apache.kylin.storage.hbase.HBaseResourceStore.getResourceImpl(HBaseResourceStore.java:287)
 at 
org.apache.kylin.common.persistence.ResourceStore.getResource(ResourceStore.java:175)
 at 
org.apache.kylin.job.dao.ExecutableDao.readJobOutputResource(ExecutableDao.java:219)
 at org.apache.kylin.job.dao.ExecutableDao.getJobOutput(ExecutableDao.java:347)
 at 
org.apache.kylin.job.execution.ExecutableManager.getOutput(ExecutableManager.java:160)
 at 
org.apache.kylin.rest.service.JobService.getSingleJobInstance(JobService.java:505)
 at org.apache.kylin.rest.service.JobService.getJobInstance(JobService.java:464)
 at org.apache.kylin.rest.service.JobService.searchJobsV2(JobService.java:973)
 at org.apache.kylin.rest.controller.JobController.list(JobController.java:86)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
 at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:498)
 at 
org.springframework.web.method.support.InvocableHandlerMethod.doInvoke(InvocableHandlerMethod.java:205)
 at 
org.springframework.web.method.support.InvocableHandlerMethod.invokeForRequest(InvocableHandlerMethod.java:133)
 at 
org.springframework.web.servlet.mvc.method.annotation.ServletInvocableHandlerMethod.invokeAndHandle(ServletInvocableHandlerMethod.java:97)
 at 
org.springframework.web.servlet.mvc.method.annotation.RequestMappingHandlerAdapter.invokeHandlerMethod(RequestMappingHandlerAdapter.java:827)
 at 
org.springframework.web.servlet.mvc.method.annotation.RequestMappingHandlerAdapter.handleInternal(RequestMappingHandlerAdapter.java:738)
 at 
org.springframework.web.servlet.mvc.method.AbstractHandlerMethodAdapter.handle(AbstractHandlerMethodAdapter.java:85)
 at 
org.springframework.web.servlet.DispatcherServlet.doDispatch(DispatcherServlet.java:967)
 at 
org.springframework.web.servlet.DispatcherServlet.doService(DispatcherServlet.java:901)
 at 
org.springframework.web.servlet.FrameworkServlet.processRequest(FrameworkServlet.java:970)
 at 
org.springframework.web.servlet.FrameworkServlet.doGet(FrameworkServlet.java:861)
 at javax.servlet.http.HttpServlet.service(HttpServlet.java:624)
 at 
org.springframework.web.servlet.FrameworkServlet.service(FrameworkServlet.java:846)
 at javax.servlet.http.HttpServlet.service(HttpServlet.java:731)
 at 
org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:303)
 at 
org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208)
 at 
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:317)
 at 
org.springframework.security.web.access.intercept.FilterSecurityInterceptor.invoke(FilterSecurityInterceptor.java:127)
 at 
org.springframework.security.web.access.intercept.FilterSecurityInterceptor.doFilter(FilterSecurityInterceptor.java:91)
 at 
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
 at 
org.springframework.security.web.access.ExceptionTranslationFilter.doFilter(ExceptionTranslationFilter.java:114)
 at 
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
 at 
org.springframework.security.web.session.SessionManagementFilter.doFilter(SessionManagementFilter.java:137)
 at 
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
 at 
org.springframework.security.web.authentication.AnonymousAuthenticationFilter.doFilter(AnonymousAuthenticationFilter.java:111)
 at 
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
 at 
org.springframework.security.web.servletapi.SecurityContextHolderAwareRequestFilter.doFilter(SecurityContextHolderAwareRequestFilter.java:170)
 at 
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
 at 
org.springframework.security.web.savedrequest.RequestCacheAwareFilter.doFilter(RequestCacheAwareFilter.java:63)
 at 
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
 at 
org.springframework.security.web.auth

[jira] [Commented] (KYLIN-3695) Error while creating hive table through Kylin build cube with mysql imported tables

2018-12-04 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3695?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16709620#comment-16709620
 ] 

ASF GitHub Bot commented on KYLIN-3695:
---

asfgit commented on issue #371: KYLIN-3695, fix lose decimal scale value in 
column type decimal(a, b)
URL: https://github.com/apache/kylin/pull/371#issuecomment-444363598
 
 
   Can one of the admins verify this patch?


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Error while creating hive table through Kylin build cube with mysql imported 
> tables
> ---
>
> Key: KYLIN-3695
> URL: https://issues.apache.org/jira/browse/KYLIN-3695
> Project: Kylin
>  Issue Type: Bug
>  Components: Integration
>Affects Versions: v2.5.0
>Reporter: Gaurav Rawat
>Assignee: XiaoXiang Yu
>Priority: Minor
> Attachments: image-2018-12-04-15-53-49-983.png
>
>
> HI I am trying to build a cube with Kylin, the data gets souced fine from 
> sqoop but the next step for creating hive tables fails . Looking at the 
> command being fired it looks weird as the create statement looks good to me .
> I think the issue is with DOUBLE types as when I remove the same the create 
> statement works fine . Can someone please help .
> I am using the stack in AWS EMR, kylin 2.5 hive 2.3.0
> The errors logs with commands as as below, the table is a msql table which 
> had columns with DOUBLE type
> Command
> {code:java}
> hive -e "USE default; DROP TABLE IF EXISTS 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368; 
> CREATE EXTERNAL TABLE IF NOT EXISTS 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368 ( 
> HOLDINGS_STOCK_INVESTOR_ID string ,STOCK_INVESTORS_CHANNEL string 
> ,STOCK_STOCK_ID string ,STOCK_DOMICILE string ,STOCK_STOCK_NM string 
> ,STOCK_APPROACH string ,STOCK_STOCK_TYP string ,INVESTOR_ID string 
> ,INVESTOR_NM string ,INVESTOR_DOMICILE_CNTRY string ,CLIENT_NM string 
> ,INVESTOR_HOLDINGS_GROSS_ASSETS_USD double(22) 
> ,INVESTOR_HOLDINGS_NET_ASSETS_USD double(22) ) ROW FORMAT DELIMITED FIELDS 
> TERMINATED BY '|' STORED AS TEXTFILE LOCATION 
> 's3://wfg1tst-models/kylin/kylin_metadata/kylin-4ae3b18b-831b-da66-eb8c-7318245c4448/kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368';
>  ALTER TABLE 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368 SET 
> TBLPROPERTIES('auto.purge'='true'); " --hiveconf hive.merge.mapredfiles=false 
> --hiveconf hive.auto.convert.join=true --hiveconf dfs.replication=2 
> --hiveconf hive.exec.compress.output=true --hiveconf 
> hive.auto.convert.join.noconditionaltask=true --hiveconf 
> mapreduce.job.split.metainfo.maxsize=-1 --hiveconf hive.merge.mapfiles=false 
> --hiveconf hive.auto.convert.join.noconditionaltask.size=1 --hiveconf 
> hive.stats.autogather=true{code}
> Error is as below
> {code:java}
> OK Time taken: 1.315 seconds OK Time taken: 0.09 seconds 
> MismatchedTokenException(334!=347) at 
> org.antlr.runtime.BaseRecognizer.recoverFromMismatchedToken(BaseRecognizer.java:617)
>  at org.antlr.runtime.BaseRecognizer.match(BaseRecognizer.java:115) at 
> org.apache.hadoop.hive.ql.parse.HiveParser.createTableStatement(HiveParser.java:6179)
>  at 
> org.apache.hadoop.hive.ql.parse.HiveParser.ddlStatement(HiveParser.java:3808) 
> at 
> org.apache.hadoop.hive.ql.parse.HiveParser.execStatement(HiveParser.java:2382)
>  at 
> org.apache.hadoop.hive.ql.parse.HiveParser.statement(HiveParser.java:1333) at 
> org.apache.hadoop.hive.ql.parse.ParseDriver.parse(ParseDriver.java:204) at 
> org.apache.hadoop.hive.ql.parse.ParseUtils.parse(ParseUtils.java:77) at 
> org.apache.hadoop.hive.ql.parse.ParseUtils.parse(ParseUtils.java:70) at 
> org.apache.hadoop.hive.ql.Driver.compile(Driver.java:468) at 
> org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1316) at 
> org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1456) at 
> org.apache.hadoop.hive.ql.Driver.run(Driver.java:1236) at 
> org.apache.hadoop.hive.ql.Driver.run(Driver.java:1226) at 
> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:233) at 
> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:184) at 
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:403) at 
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:336) at 
> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:787) at 
> org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759) at 
> org.apache.hadoo

[GitHub] asfgit commented on issue #371: KYLIN-3695, fix lose decimal scale value in column type decimal(a, b)

2018-12-04 Thread GitBox
asfgit commented on issue #371: KYLIN-3695, fix lose decimal scale value in 
column type decimal(a, b)
URL: https://github.com/apache/kylin/pull/371#issuecomment-444363598
 
 
   Can one of the admins verify this patch?


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] Emiya0306 opened a new pull request #371: KYLIN-3695, fix lose decimal scale value in column type decimal(a, b)

2018-12-04 Thread GitBox
Emiya0306 opened a new pull request #371: KYLIN-3695, fix lose decimal scale 
value in column type decimal(a, b)
URL: https://github.com/apache/kylin/pull/371
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (KYLIN-3406) When the /execute_output/ metadata file sinking to HDFS is deleted, the monitoring page does not display any tasks.

2018-12-04 Thread XiaoXiang Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/KYLIN-3406?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

XiaoXiang Yu updated KYLIN-3406:

Attachment: image-2018-12-05-13-22-42-456.png

> When the /execute_output/ metadata file sinking to HDFS is deleted, the 
> monitoring page does not display any tasks.
> ---
>
> Key: KYLIN-3406
> URL: https://issues.apache.org/jira/browse/KYLIN-3406
> Project: Kylin
>  Issue Type: Bug
>  Components: Metadata
>Affects Versions: v2.3.1
>Reporter: yujie.sun
>Assignee: XiaoXiang Yu
>Priority: Major
> Fix For: v2.6.0
>
> Attachments: image-2018-12-05-13-22-42-456.png
>
>
> *{color:#24292e}Issue description:{color}*
> When a single file of Hbase metadata is greater than a certain value, the 
> file is sunk into HDFS for storage. When files in execute_output/ are deleted 
> in HDFS, the monitoring page will not display any tasks.
> *{color:#24292e}Reproduction step:{color}*
> {color:#24292e}1.Use Hbase as metadata storage{color}
> {color:#24292e}2.Modify the hbase configuration ,configure 
> "hbase.client.keyvalue.maxsize" into a small value, such as 524288.{color}
> {color:#24292e}3.Create a new build task, after the build task finished,find 
> the /execute_output hbase metadata path on HDFS, 
> /\{working_dir}/\{metadata_url}/resources/execute_output/..., and delete one 
> of the files.{color}
> {color:#24292e}4.At this time,{color:#33}the monitoring page will not 
> display any tasks.{color}{color}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (KYLIN-3406) When the /execute_output/ metadata file sinking to HDFS is deleted, the monitoring page does not display any tasks.

2018-12-04 Thread XiaoXiang Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3406?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16709615#comment-16709615
 ] 

XiaoXiang Yu commented on KYLIN-3406:
-

!image-2018-12-05-13-22-42-456.png!

> When the /execute_output/ metadata file sinking to HDFS is deleted, the 
> monitoring page does not display any tasks.
> ---
>
> Key: KYLIN-3406
> URL: https://issues.apache.org/jira/browse/KYLIN-3406
> Project: Kylin
>  Issue Type: Bug
>  Components: Metadata
>Affects Versions: v2.3.1
>Reporter: yujie.sun
>Assignee: XiaoXiang Yu
>Priority: Major
> Fix For: v2.6.0
>
> Attachments: image-2018-12-05-13-22-42-456.png
>
>
> *{color:#24292e}Issue description:{color}*
> When a single file of Hbase metadata is greater than a certain value, the 
> file is sunk into HDFS for storage. When files in execute_output/ are deleted 
> in HDFS, the monitoring page will not display any tasks.
> *{color:#24292e}Reproduction step:{color}*
> {color:#24292e}1.Use Hbase as metadata storage{color}
> {color:#24292e}2.Modify the hbase configuration ,configure 
> "hbase.client.keyvalue.maxsize" into a small value, such as 524288.{color}
> {color:#24292e}3.Create a new build task, after the build task finished,find 
> the /execute_output hbase metadata path on HDFS, 
> /\{working_dir}/\{metadata_url}/resources/execute_output/..., and delete one 
> of the files.{color}
> {color:#24292e}4.At this time,{color:#33}the monitoring page will not 
> display any tasks.{color}{color}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (KYLIN-3690) New streaming backend implementation

2018-12-04 Thread Ma Gang (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3690?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16709553#comment-16709553
 ] 

Ma Gang commented on KYLIN-3690:


Hi Hubert,

The PR should be ready next week, because I did some refactor work after merge 
the existing code to the master, and I need further test after the refactoring. 
Sorry it comes a little bit late.

> New streaming backend implementation
> 
>
> Key: KYLIN-3690
> URL: https://issues.apache.org/jira/browse/KYLIN-3690
> Project: Kylin
>  Issue Type: Sub-task
>Reporter: Ma Gang
>Assignee: Ma Gang
>Priority: Major
>
> New streaming backend implementation



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] shaofengshi commented on a change in pull request #348: added release notes, modified kylin and coprocessor upgrade guides

2018-12-04 Thread GitBox
shaofengshi commented on a change in pull request #348: added release notes, 
modified kylin and coprocessor upgrade guides
URL: https://github.com/apache/kylin/pull/348#discussion_r238900904
 
 

 ##
 File path: website/_docs/howto/howto_upgrade.md
 ##
 @@ -1,112 +1,146 @@
 ---
 layout: docs
-title:  Upgrade From Old Versions
+title:  Upgrade Kylin From Old Versions
 categories: howto
 permalink: /docs/howto/howto_upgrade.html
 since: v1.5.1
 ---
 
-Running as a Hadoop client, Apache Kylin's metadata and Cube data are 
persistended in Hadoop (HBase and HDFS), so the upgrade is relatively easy and 
user does not need worry about data loss. The upgrade can be performed in the 
following steps:
+Running as a Hadoop client, Apache Kylin's metadata and Cube data are 
persistent in Hadoop (HBase and HDFS), so the upgrade is relatively easy and 
user does not need worry about data loss. The upgrade can be performed in the 
following steps:
 
 Review comment:
   "persistent" is incorrect in this context, it is adjective, while here we 
need "be persisted" which is a passive voice;  (a typo in the previous version).


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] shaofengshi commented on a change in pull request #348: added release notes, modified kylin and coprocessor upgrade guides

2018-12-04 Thread GitBox
shaofengshi commented on a change in pull request #348: added release notes, 
modified kylin and coprocessor upgrade guides
URL: https://github.com/apache/kylin/pull/348#discussion_r238900533
 
 

 ##
 File path: website/_docs/howto/howto_upgrade.md
 ##
 @@ -1,112 +1,146 @@
 ---
 layout: docs
-title:  Upgrade From Old Versions
+title:  Upgrade Kylin From Old Versions
 
 Review comment:
   Of course it is "Kylin", this looks redundant ... 


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] shaofengshi commented on a change in pull request #348: added release notes, modified kylin and coprocessor upgrade guides

2018-12-04 Thread GitBox
shaofengshi commented on a change in pull request #348: added release notes, 
modified kylin and coprocessor upgrade guides
URL: https://github.com/apache/kylin/pull/348#discussion_r238900295
 
 

 ##
 File path: website/_data/docs.yml
 ##
 @@ -78,6 +77,10 @@
   - howto/howto_use_cli
   - howto/howto_ldap_and_sso
   - howto/howto_use_beeline
-  - howto/howto_update_coprocessor
+  - howto/howto_upgrade_coprocessor
 
 Review comment:
   This change will change the URL endpoint. We hope the URL can be kept as 
there are many external references.  


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] shaofengshi commented on a change in pull request #348: added release notes, modified kylin and coprocessor upgrade guides

2018-12-04 Thread GitBox
shaofengshi commented on a change in pull request #348: added release notes, 
modified kylin and coprocessor upgrade guides
URL: https://github.com/apache/kylin/pull/348#discussion_r238900055
 
 

 ##
 File path: website/download/index.md
 ##
 @@ -20,7 +20,7 @@ You can verify the download by following these 
[procedures](https://www.apache.o
 
  v2.5.0
 - This is a major release after 2.4, with 96 bug fixes and enhancement. Check 
the [v2.5.0 release announcement](/blog/2018/09/20/release-v2.5.0/) and the 
release notes. 
-- [Release notes](/docs/release_notes.html) and [upgrade 
guide](/docs/howto/howto_upgrade.html)
+- [Kylin Release Notes](/docs/releases/kylin_release_notes.html) and [Upgrade 
Kylin From Old Versions](/docs/howto/howto_upgrade.html)
 
 Review comment:
   I think the "upgrade guide" in lower case is more friendly than "Upgrade 
Kylin From Old Versions" in hump...


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] shaofengshi commented on a change in pull request #348: added release notes, modified kylin and coprocessor upgrade guides

2018-12-04 Thread GitBox
shaofengshi commented on a change in pull request #348: added release notes, 
modified kylin and coprocessor upgrade guides
URL: https://github.com/apache/kylin/pull/348#discussion_r238899806
 
 

 ##
 File path: website/download/index.cn.md
 ##
 @@ -21,7 +21,7 @@ title: 下载
 
  v2.5.0
 - 这是2.4版本后的一个主要发布版本,包含了96 个以及各种改进。关于具体内容请查看[v2.5.0 
正式发布](/blog/2018/09/20/release-v2.5.0/). 
-- [发布说明](/docs/release_notes.html) and [升级指南](/docs/howto/howto_upgrade.html)
+- [发布声明](/cn/docs/releases/kylin_release_notes.html) and [Kylin 
升级](/cn/docs/howto/howto_upgrade.html)
 
 Review comment:
   We don't need to provide multiple language for release notes...


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] shaofengshi commented on a change in pull request #348: added release notes, modified kylin and coprocessor upgrade guides

2018-12-04 Thread GitBox
shaofengshi commented on a change in pull request #348: added release notes, 
modified kylin and coprocessor upgrade guides
URL: https://github.com/apache/kylin/pull/348#discussion_r238899445
 
 

 ##
 File path: website/_docs/index.cn.md
 ##
 @@ -9,21 +9,31 @@ permalink: /cn/docs/index.html
   
 > Extreme OLAP Engine for Big Data
 
-Apache Kylin™是一个开源的分布式分析引擎,提供Hadoop之上的SQL查询接口及多维分析(OLAP)能力以支持超大规模数据,最初由eBay 
Inc.开发并贡献至开源社区。
+Apache Kylin™ 是一个开源 OLAP(On-Line Analytics 
Processing)分析引擎,针对数据库与亿万级记录提供亚秒级查询响应能力。2014 年由 eBay Inc.开发并贡献至开源社区。2015 年毕业于 
Apache 顶级项目并迅速被全球数千家公司与组织应用于数据分析业务,同时也在 2015 与 2016 年赢得 InfoWorld Bossie Awards 
年度最佳大数据开源工具。 
 
 Review comment:
   "Online" is a word, not "On-Line", please check the terms: 
https://en.wikipedia.org/wiki/Online_analytical_processing


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] shaofengshi commented on a change in pull request #348: added release notes, modified kylin and coprocessor upgrade guides

2018-12-04 Thread GitBox
shaofengshi commented on a change in pull request #348: added release notes, 
modified kylin and coprocessor upgrade guides
URL: https://github.com/apache/kylin/pull/348#discussion_r238899442
 
 

 ##
 File path: website/_docs/index.cn.md
 ##
 @@ -9,21 +9,31 @@ permalink: /cn/docs/index.html
   
 > Extreme OLAP Engine for Big Data
 
-Apache Kylin™是一个开源的分布式分析引擎,提供Hadoop之上的SQL查询接口及多维分析(OLAP)能力以支持超大规模数据,最初由eBay 
Inc.开发并贡献至开源社区。
+Apache Kylin™ 是一个开源 OLAP(On-Line Analytics 
Processing)分析引擎,针对数据库与亿万级记录提供亚秒级查询响应能力。2014 年由 eBay Inc.开发并贡献至开源社区。2015 年毕业于 
Apache 顶级项目并迅速被全球数千家公司与组织应用于数据分析业务,同时也在 2015 与 2016 年赢得 InfoWorld Bossie Awards 
年度最佳大数据开源工具。 
 
 Review comment:
   "Online" is a word, not "On-Line", please check the terms: 
https://en.wikipedia.org/wiki/Online_analytical_processing


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (KYLIN-3686) Top_N metric code requires cube storage type to be ID_SHARDED_HBASE, but the Web UI defaults to ID_HBASE and provides no safeguards against storage type mismatch

2018-12-04 Thread Vsevolod Ostapenko (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3686?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16709162#comment-16709162
 ] 

Vsevolod Ostapenko commented on KYLIN-3686:
---

Hi Chao, 
the "kylin.storage.default" parameter is not set in kylin.properties in our 
environment, so it does default ti ID_HBASE, as I understand.
As far as I can see, the fix for KYLIN-3636 changes only cube defaults to be 
ID_SHARDED_HBASE.
However, it does not address the misalignment and lack of safety checks between 
cube storage type and implied pre-requisites of Top_N metric. It's still 
possible to load cube with ID_HBASE from JSON and define Topn_N metric and get 
failing cube build with no clear explanations for the failure reasons.

Thanks,
Vsevolod.

> Top_N metric code requires cube storage type to be ID_SHARDED_HBASE, but the 
> Web UI defaults to ID_HBASE and provides no safeguards against storage type 
> mismatch
> -
>
> Key: KYLIN-3686
> URL: https://issues.apache.org/jira/browse/KYLIN-3686
> Project: Kylin
>  Issue Type: Improvement
>  Components: Measure - TopN, Metadata, Web 
>Affects Versions: v2.5.0
> Environment: HDP 2.5.6, Kylin 2.5
>Reporter: Vsevolod Ostapenko
>Assignee: Shaofeng SHI
>Priority: Major
> Fix For: v2.6.0
>
>
> When new cube is defined via Kylin 2.5 UI, the default cube storage type is 
> set to 0 (ID_HBASE).
>  Top_N metric support is currently hard coded to expect cube storage type 2 
> (ID_SHARDED_HBASE), and it *_does not_* check if the cube storage type is the 
> "sharded HBASE".
>  UI provides no safeguards either to prevent a user from defining a cube with 
> Top_N metric that would blow up on the cube building stage with a perplexing 
> stack trace like the following:
> {quote}2018-10-22 16:15:50,388 ERROR [main] 
> org.apache.kylin.engine.mr.KylinMapper:
>  java.lang.ArrayIndexOutOfBoundsException
>  at java.lang.System.arraycopy(Native Method)
>  at 
> org.apache.kylin.engine.mr.common.NDCuboidBuilder.buildKeyInternal(NDCuboidBuilder.java:106)
>  at 
> org.apache.kylin.engine.mr.common.NDCuboidBuilder.buildKey(NDCuboidBuilder.java:71)
>  at 
> org.apache.kylin.engine.mr.steps.NDCuboidMapper.doMap(NDCuboidMapper.java:112)
>  at 
> org.apache.kylin.engine.mr.steps.NDCuboidMapper.doMap(NDCuboidMapper.java:47)
>  at org.apache.kylin.engine.mr.KylinMapper.map(KylinMapper.java:77)
>  at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:146)
> {quote}
> Please, either:
> – modify Top_N code to support all cube storage types (not only 
> ID_SHARDED_HBASE),
>  or 
>  – modify Top_N code to perform explicit check for cube storage type and 
> raise descriptive exception, when cube storage is not the one that is 
> expected. Plus update the UI to prevent the user from creating cube 
> definitions that are incompatible with the storage type compatible with Top_N 
> measure
> PS: NDCCuboidBuilder,java contains the following line:
> {quote}int offset = RowConstants.ROWKEY_SHARDID_LEN + 
> RowConstants.ROWKEY_CUBOIDID_LEN; // skip shard and cuboidId{quote}
> If cube storage type is not ID_SHARDED_HBASE, offset is calculated 
> incorrectly, which leads to ArrayIndexOutOfBounds exception.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (KYLIN-3686) Top_N metric code requires cube storage type to be ID_SHARDED_HBASE, but the Web UI defaults to ID_HBASE and provides no safeguards against storage type mismatch

2018-12-04 Thread Chao Long (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3686?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16708623#comment-16708623
 ] 

Chao Long edited comment on KYLIN-3686 at 12/4/18 12:21 PM:


Hi Vsevolod, The front-end will get the value of parameter 
"kylin.storage.default"(default ID_SHARDED_HBASE=2) from the kylin.properties 
and will use it as the cube's storage type. Can you check this parameter in 
your kylin.properties?

Another situation is the front-end does not get this value for some unknown 
reason, it will pass a null value to the back-end, and the back-end will use 
ID_HBASE as the default value of storage type.

If the latter, KYLIN-3636 has fixed it and will release soon in kylin-2.5.2.


was (Author: wayne0101):
Hi Vsevolod, The front-end will get the value of parameter 
"kylin.storage.default"(default ID_SHARDED_HBASE=2) from the kylin.properties 
and will use it as the cube's storage type. Can you check this parameter in 
your kylin.properties?

Another situation is the front-end does not get this value for some unknown 
reason, it will pass a null value to the back-end, and the back-end will use 
ID_HBASE as the default value of storage type.

If the latter, [KYLIN-3636|https://issues.apache.org/jira/browse/KYLIN-3636] 
has fixed it and will release soon.

> Top_N metric code requires cube storage type to be ID_SHARDED_HBASE, but the 
> Web UI defaults to ID_HBASE and provides no safeguards against storage type 
> mismatch
> -
>
> Key: KYLIN-3686
> URL: https://issues.apache.org/jira/browse/KYLIN-3686
> Project: Kylin
>  Issue Type: Improvement
>  Components: Measure - TopN, Metadata, Web 
>Affects Versions: v2.5.0
> Environment: HDP 2.5.6, Kylin 2.5
>Reporter: Vsevolod Ostapenko
>Assignee: Shaofeng SHI
>Priority: Major
> Fix For: v2.6.0
>
>
> When new cube is defined via Kylin 2.5 UI, the default cube storage type is 
> set to 0 (ID_HBASE).
>  Top_N metric support is currently hard coded to expect cube storage type 2 
> (ID_SHARDED_HBASE), and it *_does not_* check if the cube storage type is the 
> "sharded HBASE".
>  UI provides no safeguards either to prevent a user from defining a cube with 
> Top_N metric that would blow up on the cube building stage with a perplexing 
> stack trace like the following:
> {quote}2018-10-22 16:15:50,388 ERROR [main] 
> org.apache.kylin.engine.mr.KylinMapper:
>  java.lang.ArrayIndexOutOfBoundsException
>  at java.lang.System.arraycopy(Native Method)
>  at 
> org.apache.kylin.engine.mr.common.NDCuboidBuilder.buildKeyInternal(NDCuboidBuilder.java:106)
>  at 
> org.apache.kylin.engine.mr.common.NDCuboidBuilder.buildKey(NDCuboidBuilder.java:71)
>  at 
> org.apache.kylin.engine.mr.steps.NDCuboidMapper.doMap(NDCuboidMapper.java:112)
>  at 
> org.apache.kylin.engine.mr.steps.NDCuboidMapper.doMap(NDCuboidMapper.java:47)
>  at org.apache.kylin.engine.mr.KylinMapper.map(KylinMapper.java:77)
>  at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:146)
> {quote}
> Please, either:
> – modify Top_N code to support all cube storage types (not only 
> ID_SHARDED_HBASE),
>  or 
>  – modify Top_N code to perform explicit check for cube storage type and 
> raise descriptive exception, when cube storage is not the one that is 
> expected. Plus update the UI to prevent the user from creating cube 
> definitions that are incompatible with the storage type compatible with Top_N 
> measure
> PS: NDCCuboidBuilder,java contains the following line:
> {quote}int offset = RowConstants.ROWKEY_SHARDID_LEN + 
> RowConstants.ROWKEY_CUBOIDID_LEN; // skip shard and cuboidId{quote}
> If cube storage type is not ID_SHARDED_HBASE, offset is calculated 
> incorrectly, which leads to ArrayIndexOutOfBounds exception.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (KYLIN-3686) Top_N metric code requires cube storage type to be ID_SHARDED_HBASE, but the Web UI defaults to ID_HBASE and provides no safeguards against storage type mismatch

2018-12-04 Thread Chao Long (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3686?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16708623#comment-16708623
 ] 

Chao Long commented on KYLIN-3686:
--

Hi Vsevolod, The front-end will get the value of parameter 
"kylin.storage.default"(default ID_SHARDED_HBASE=2) from the kylin.properties 
and will use it as the cube's storage type. Can you check this parameter in 
your kylin.properties?

Another situation is the front-end does not get this value for some unknown 
reason, it will pass a null value to the back-end, and the back-end will use 
ID_HBASE as the default value of storage type.

If the latter, [KYLIN-3636|https://issues.apache.org/jira/browse/KYLIN-3636] 
has fixed it and will release soon.

> Top_N metric code requires cube storage type to be ID_SHARDED_HBASE, but the 
> Web UI defaults to ID_HBASE and provides no safeguards against storage type 
> mismatch
> -
>
> Key: KYLIN-3686
> URL: https://issues.apache.org/jira/browse/KYLIN-3686
> Project: Kylin
>  Issue Type: Improvement
>  Components: Measure - TopN, Metadata, Web 
>Affects Versions: v2.5.0
> Environment: HDP 2.5.6, Kylin 2.5
>Reporter: Vsevolod Ostapenko
>Assignee: Shaofeng SHI
>Priority: Major
> Fix For: v2.6.0
>
>
> When new cube is defined via Kylin 2.5 UI, the default cube storage type is 
> set to 0 (ID_HBASE).
>  Top_N metric support is currently hard coded to expect cube storage type 2 
> (ID_SHARDED_HBASE), and it *_does not_* check if the cube storage type is the 
> "sharded HBASE".
>  UI provides no safeguards either to prevent a user from defining a cube with 
> Top_N metric that would blow up on the cube building stage with a perplexing 
> stack trace like the following:
> {quote}2018-10-22 16:15:50,388 ERROR [main] 
> org.apache.kylin.engine.mr.KylinMapper:
>  java.lang.ArrayIndexOutOfBoundsException
>  at java.lang.System.arraycopy(Native Method)
>  at 
> org.apache.kylin.engine.mr.common.NDCuboidBuilder.buildKeyInternal(NDCuboidBuilder.java:106)
>  at 
> org.apache.kylin.engine.mr.common.NDCuboidBuilder.buildKey(NDCuboidBuilder.java:71)
>  at 
> org.apache.kylin.engine.mr.steps.NDCuboidMapper.doMap(NDCuboidMapper.java:112)
>  at 
> org.apache.kylin.engine.mr.steps.NDCuboidMapper.doMap(NDCuboidMapper.java:47)
>  at org.apache.kylin.engine.mr.KylinMapper.map(KylinMapper.java:77)
>  at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:146)
> {quote}
> Please, either:
> – modify Top_N code to support all cube storage types (not only 
> ID_SHARDED_HBASE),
>  or 
>  – modify Top_N code to perform explicit check for cube storage type and 
> raise descriptive exception, when cube storage is not the one that is 
> expected. Plus update the UI to prevent the user from creating cube 
> definitions that are incompatible with the storage type compatible with Top_N 
> measure
> PS: NDCCuboidBuilder,java contains the following line:
> {quote}int offset = RowConstants.ROWKEY_SHARDID_LEN + 
> RowConstants.ROWKEY_CUBOIDID_LEN; // skip shard and cuboidId{quote}
> If cube storage type is not ID_SHARDED_HBASE, offset is calculated 
> incorrectly, which leads to ArrayIndexOutOfBounds exception.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (KYLIN-3690) New streaming backend implementation

2018-12-04 Thread Hubert STEFANI (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3690?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16708544#comment-16708544
 ] 

Hubert STEFANI commented on KYLIN-3690:
---

Hi Ma,

is there any rescheduling of the PR ?

We feel so impatient to help you on making it production ready !

Hu{color:#FF}bert{color}

> New streaming backend implementation
> 
>
> Key: KYLIN-3690
> URL: https://issues.apache.org/jira/browse/KYLIN-3690
> Project: Kylin
>  Issue Type: Sub-task
>Reporter: Ma Gang
>Assignee: Ma Gang
>Priority: Major
>
> New streaming backend implementation



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (KYLIN-3406) When the /execute_output/ metadata file sinking to HDFS is deleted, the monitoring page does not display any tasks.

2018-12-04 Thread XiaoXiang Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/KYLIN-3406?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

XiaoXiang Yu reassigned KYLIN-3406:
---

Assignee: XiaoXiang Yu

> When the /execute_output/ metadata file sinking to HDFS is deleted, the 
> monitoring page does not display any tasks.
> ---
>
> Key: KYLIN-3406
> URL: https://issues.apache.org/jira/browse/KYLIN-3406
> Project: Kylin
>  Issue Type: Bug
>  Components: Metadata
>Affects Versions: v2.3.1
>Reporter: yujie.sun
>Assignee: XiaoXiang Yu
>Priority: Major
> Fix For: v2.6.0
>
>
> *{color:#24292e}Issue description:{color}*
> When a single file of Hbase metadata is greater than a certain value, the 
> file is sunk into HDFS for storage. When files in execute_output/ are deleted 
> in HDFS, the monitoring page will not display any tasks.
> *{color:#24292e}Reproduction step:{color}*
> {color:#24292e}1.Use Hbase as metadata storage{color}
> {color:#24292e}2.Modify the hbase configuration ,configure 
> "hbase.client.keyvalue.maxsize" into a small value, such as 524288.{color}
> {color:#24292e}3.Create a new build task, after the build task finished,find 
> the /execute_output hbase metadata path on HDFS, 
> /\{working_dir}/\{metadata_url}/resources/execute_output/..., and delete one 
> of the files.{color}
> {color:#24292e}4.At this time,{color:#33}the monitoring page will not 
> display any tasks.{color}{color}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (KYLIN-3695) Error while creating hive table through Kylin build cube with mysql imported tables

2018-12-04 Thread XiaoXiang Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3695?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16708395#comment-16708395
 ] 

XiaoXiang Yu edited comment on KYLIN-3695 at 12/4/18 9:04 AM:
--

[~gauravbrills] I debug in my test env.

I found when using decimal data type with empty precision to build a cube. The 
front end(JavaScript) will use 'undefined'  for null value and send it to 
back-end to validate cube metadata. It clear that  decimal(19, undefined) do 
not match a valid regex pattern, so back end return error message to front end 
as my above message.

Source code at 
org.apache.kylin.metadata.datatype.DataType.[https://github.com/apache/kylin/blob/master/core-metadata/src/main/java/org/apache/kylin/metadata/datatype/DataType.java]

I think it has no need to fix it because the root cause is clear.

 

The following is part of preload send by front end.       

{
 "name":"SUM_INTEREST_SCORE2",
 "function":{
 "expression":"SUM",
 "returntype":"decimal(19,undefined)",
 "parameter":

{ "type":"column", "value":"USERACTION.INTEREST_SCORE2" 
}

}
 }


was (Author: hit_lacus):
[~gauravbrills] I debug in my test env.

I found when using decimal data type with empty precision to build a cube. The 
front end(JavaScript) will use 'undefined'  for null value and send it to 
back-end to validate cube metadata. It clear that  decimal(19, undefined) do 
not match a valid regex pattern, so back end return error message to front end 
as my above message.

Source code at 
org.apache.kylin.metadata.datatype.DataType.[https://github.com/apache/kylin/blob/master/core-metadata/src/main/java/org/apache/kylin/metadata/datatype/DataType.java]

I think it has no need to fix it because the root cause is clear.

 

The following is part of preload send by backend.       

{
 "name":"SUM_INTEREST_SCORE2",
 "function":{
 "expression":"SUM",
 "returntype":"decimal(19,undefined)",
 "parameter":

{ "type":"column", "value":"USERACTION.INTEREST_SCORE2" 
}

}
 }

> Error while creating hive table through Kylin build cube with mysql imported 
> tables
> ---
>
> Key: KYLIN-3695
> URL: https://issues.apache.org/jira/browse/KYLIN-3695
> Project: Kylin
>  Issue Type: Bug
>  Components: Integration
>Affects Versions: v2.5.0
>Reporter: Gaurav Rawat
>Assignee: XiaoXiang Yu
>Priority: Minor
> Attachments: image-2018-12-04-15-53-49-983.png
>
>
> HI I am trying to build a cube with Kylin, the data gets souced fine from 
> sqoop but the next step for creating hive tables fails . Looking at the 
> command being fired it looks weird as the create statement looks good to me .
> I think the issue is with DOUBLE types as when I remove the same the create 
> statement works fine . Can someone please help .
> I am using the stack in AWS EMR, kylin 2.5 hive 2.3.0
> The errors logs with commands as as below, the table is a msql table which 
> had columns with DOUBLE type
> Command
> {code:java}
> hive -e "USE default; DROP TABLE IF EXISTS 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368; 
> CREATE EXTERNAL TABLE IF NOT EXISTS 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368 ( 
> HOLDINGS_STOCK_INVESTOR_ID string ,STOCK_INVESTORS_CHANNEL string 
> ,STOCK_STOCK_ID string ,STOCK_DOMICILE string ,STOCK_STOCK_NM string 
> ,STOCK_APPROACH string ,STOCK_STOCK_TYP string ,INVESTOR_ID string 
> ,INVESTOR_NM string ,INVESTOR_DOMICILE_CNTRY string ,CLIENT_NM string 
> ,INVESTOR_HOLDINGS_GROSS_ASSETS_USD double(22) 
> ,INVESTOR_HOLDINGS_NET_ASSETS_USD double(22) ) ROW FORMAT DELIMITED FIELDS 
> TERMINATED BY '|' STORED AS TEXTFILE LOCATION 
> 's3://wfg1tst-models/kylin/kylin_metadata/kylin-4ae3b18b-831b-da66-eb8c-7318245c4448/kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368';
>  ALTER TABLE 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368 SET 
> TBLPROPERTIES('auto.purge'='true'); " --hiveconf hive.merge.mapredfiles=false 
> --hiveconf hive.auto.convert.join=true --hiveconf dfs.replication=2 
> --hiveconf hive.exec.compress.output=true --hiveconf 
> hive.auto.convert.join.noconditionaltask=true --hiveconf 
> mapreduce.job.split.metainfo.maxsize=-1 --hiveconf hive.merge.mapfiles=false 
> --hiveconf hive.auto.convert.join.noconditionaltask.size=1 --hiveconf 
> hive.stats.autogather=true{code}
> Error is as below
> {code:java}
> OK Time taken: 1.315 seconds OK Time taken: 0.09 seconds 
> MismatchedTokenException(334!=347) at 
> org.antlr.runtime.BaseRecognizer.recoverFromMismatchedToken(BaseRecognizer.java:617)
>  at org.antlr.runtime

[jira] [Commented] (KYLIN-3617) Reduce number of visiting metastore for job scheduler

2018-12-04 Thread XiaoXiang Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3617?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16708421#comment-16708421
 ] 

XiaoXiang Yu commented on KYLIN-3617:
-

I will try to verfiy these patch and merge to master branch.

> Reduce number of visiting metastore for job scheduler
> -
>
> Key: KYLIN-3617
> URL: https://issues.apache.org/jira/browse/KYLIN-3617
> Project: Kylin
>  Issue Type: Improvement
>  Components: Job Engine
>Affects Versions: v2.4.1
>Reporter: nichunen
>Assignee: nichunen
>Priority: Major
> Fix For: v2.6.0
>
>
> For KYLIN-3470 introduced cache for jobs' metadata, it's also can be used in 
> job scheduler to reduce the pressure on metastore



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (KYLIN-3695) Error while creating hive table through Kylin build cube with mysql imported tables

2018-12-04 Thread XiaoXiang Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3695?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16708395#comment-16708395
 ] 

XiaoXiang Yu edited comment on KYLIN-3695 at 12/4/18 8:55 AM:
--

[~gauravbrills] I debug in my test env.

I found when using decimal data type with empty precision to build a cube. The 
front end(JavaScript) will use 'undefined'  for null value and send it to 
back-end to validate cube metadata. It clear that  decimal(19, undefined) do 
not match a valid regex pattern, so back end return error message to front end 
as my above message.

Source code at 
org.apache.kylin.metadata.datatype.DataType.[https://github.com/apache/kylin/blob/master/core-metadata/src/main/java/org/apache/kylin/metadata/datatype/DataType.java]

I think it has no need to fix it because the root cause is clear.

 

The following is part of preload send by backend.       

{
 "name":"SUM_INTEREST_SCORE2",
 "function":{
 "expression":"SUM",
 "returntype":"decimal(19,undefined)",
 "parameter":

{ "type":"column", "value":"USERACTION.INTEREST_SCORE2" 
}

}
 }


was (Author: hit_lacus):
[~gauravbrills] I debug in my test env.

I found when using decimal data type with empty precision to build a cube. The 
front end(JavaScript) will use 'undefined'  for null value and send it to 
back-end to validate cube metadata. It clear that  decimal(19, undefined) do 
not match a valid regex pattern, so back end return error message to front end 
as my above message.

Source code at 
org.apache.kylin.metadata.datatype.DataType.[https://github.com/apache/kylin/blob/master/core-metadata/src/main/java/org/apache/kylin/metadata/datatype/DataType.java]

I think it has no need to fix it because the root cause is clear.

 
       

{
"name":"SUM_INTEREST_SCORE2",
"function":{
"expression":"SUM",
"returntype":"decimal(19,undefined)",
"parameter":{
"type":"column",
"value":"USERACTION.INTEREST_SCORE2"
}
}
}

> Error while creating hive table through Kylin build cube with mysql imported 
> tables
> ---
>
> Key: KYLIN-3695
> URL: https://issues.apache.org/jira/browse/KYLIN-3695
> Project: Kylin
>  Issue Type: Bug
>  Components: Integration
>Affects Versions: v2.5.0
>Reporter: Gaurav Rawat
>Assignee: XiaoXiang Yu
>Priority: Minor
> Attachments: image-2018-12-04-15-53-49-983.png
>
>
> HI I am trying to build a cube with Kylin, the data gets souced fine from 
> sqoop but the next step for creating hive tables fails . Looking at the 
> command being fired it looks weird as the create statement looks good to me .
> I think the issue is with DOUBLE types as when I remove the same the create 
> statement works fine . Can someone please help .
> I am using the stack in AWS EMR, kylin 2.5 hive 2.3.0
> The errors logs with commands as as below, the table is a msql table which 
> had columns with DOUBLE type
> Command
> {code:java}
> hive -e "USE default; DROP TABLE IF EXISTS 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368; 
> CREATE EXTERNAL TABLE IF NOT EXISTS 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368 ( 
> HOLDINGS_STOCK_INVESTOR_ID string ,STOCK_INVESTORS_CHANNEL string 
> ,STOCK_STOCK_ID string ,STOCK_DOMICILE string ,STOCK_STOCK_NM string 
> ,STOCK_APPROACH string ,STOCK_STOCK_TYP string ,INVESTOR_ID string 
> ,INVESTOR_NM string ,INVESTOR_DOMICILE_CNTRY string ,CLIENT_NM string 
> ,INVESTOR_HOLDINGS_GROSS_ASSETS_USD double(22) 
> ,INVESTOR_HOLDINGS_NET_ASSETS_USD double(22) ) ROW FORMAT DELIMITED FIELDS 
> TERMINATED BY '|' STORED AS TEXTFILE LOCATION 
> 's3://wfg1tst-models/kylin/kylin_metadata/kylin-4ae3b18b-831b-da66-eb8c-7318245c4448/kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368';
>  ALTER TABLE 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368 SET 
> TBLPROPERTIES('auto.purge'='true'); " --hiveconf hive.merge.mapredfiles=false 
> --hiveconf hive.auto.convert.join=true --hiveconf dfs.replication=2 
> --hiveconf hive.exec.compress.output=true --hiveconf 
> hive.auto.convert.join.noconditionaltask=true --hiveconf 
> mapreduce.job.split.metainfo.maxsize=-1 --hiveconf hive.merge.mapfiles=false 
> --hiveconf hive.auto.convert.join.noconditionaltask.size=1 --hiveconf 
> hive.stats.autogather=true{code}
> Error is as below
> {code:java}
> OK Time taken: 1.315 seconds OK Time taken: 0.09 seconds 
> MismatchedTokenException(334!=347) at 
> org.antlr.runtime.BaseRecognizer.recoverFromMismatchedToken(BaseRecognizer.java:617)
>  at org.antlr.runtime.BaseRecognizer.match(BaseRecognizer.java:115) at 
> org.apach

[jira] [Updated] (KYLIN-3709) Pass the wrong parameter to addResource function, kylin can not load the configuration file hbase.hdfs.xml .

2018-12-04 Thread Hongtao He (JIRA)


 [ 
https://issues.apache.org/jira/browse/KYLIN-3709?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hongtao He updated KYLIN-3709:
--
Description: 
When calling Configuration.addResource, KYLIN-3648 use a Path as the parameter 
instead of a string in Kylin's HBaseConnection.
{code:java}
String hdfsConfigFile = 
KylinConfig.getInstanceFromEnv().getHBaseClusterHDFSConfigFile();
if (hdfsConfigFile == null || hdfsConfigFile.isEmpty()) {
return;
}
Configuration hdfsConf = new Configuration(false);
hdfsConf.addResource(new Path(hdfsConfigFile));
{code}
Use a Path as the parameter of Configuration.addResource is better. 
Unfortunately,  the parameter which passed to the addResource function is 
wrong. The addResource function only accepts absolute paths, but the parameter 
is just a filename. For example, the value of 

hdfsConfigFile is "hbase.hdfs.xml", so addResource function will not work . The 
end result is that kylin can not load the hbase configuration file 
hbase.hdfs.xml .

There are two ways to fix this bug, and I think method 1 is better.

Method-1.revert the code
{code:java}
String hdfsConfigFile = 
KylinConfig.getInstanceFromEnv().getHBaseClusterHDFSConfigFile();
if (hdfsConfigFile == null || hdfsConfigFile.isEmpty()) {
return;
}
Configuration hdfsConf = new Configuration(false);
hdfsConf.addResource(hdfsConfigFile);{code}
Method-2.Get the absolute path of the configuration file
{code:java}
String hdfsConfigFile = 
KylinConfig.getInstanceFromEnv().getHBaseClusterHDFSConfigFile();
if (hdfsConfigFile == null || hdfsConfigFile.isEmpty()) {
return;
}
Configuration hdfsConf = new Configuration(false);
String hbaseHdfsConfigPath = System.getProperty("user.dir") + "/../conf/" + 
hdfsConfigFile;
hdfsConf.addResource(new Path(hbaseHdfsConfigPath));
{code}
 

  was:
When calling Configuration.addResource, 
[KYLIN-3648|https://jira.apache.org/jira/browse/KYLIN-3648] use a Path as the 
parameter instead of a string in Kylin's HBaseConnection.

 
{code:java}
String hdfsConfigFile = 
KylinConfig.getInstanceFromEnv().getHBaseClusterHDFSConfigFile();
if (hdfsConfigFile == null || hdfsConfigFile.isEmpty()) {
return;
}
Configuration hdfsConf = new Configuration(false);
hdfsConf.addResource(new Path(hdfsConfigFile));
{code}
Use a Path as the parameter of Configuration.addResource is better. 
Unfortunately,  the parameter which passed to the addResource function is 
wrong. The addResource function only accepts absolute paths, but the parameter 
is just a filename. For example, the value of 

hdfsConfigFile is "hbase.hdfs.xml", so addResource function will not work . The 
end result is that kylin can not load the hbase configuration file 
hbase.hdfs.xml .

There are two ways to fix this bug, and I think method 1 is better.

Method-1.revert the code
{code:java}
String hdfsConfigFile = 
KylinConfig.getInstanceFromEnv().getHBaseClusterHDFSConfigFile();
if (hdfsConfigFile == null || hdfsConfigFile.isEmpty()) {
return;
}
Configuration hdfsConf = new Configuration(false);
hdfsConf.addResource(hdfsConfigFile);{code}
Method-2.Get the absolute path of the configuration file
{code:java}
String hdfsConfigFile = 
KylinConfig.getInstanceFromEnv().getHBaseClusterHDFSConfigFile();
if (hdfsConfigFile == null || hdfsConfigFile.isEmpty()) {
return;
}
Configuration hdfsConf = new Configuration(false);
String hbaseHdfsConfigPath = System.getProperty("user.dir") + "/../conf/" + 
hdfsConfigFile;
hdfsConf.addResource(new Path(hbaseHdfsConfigPath));
{code}
 


> Pass the wrong parameter to addResource function, kylin can not load the 
> configuration file hbase.hdfs.xml .
> 
>
> Key: KYLIN-3709
> URL: https://issues.apache.org/jira/browse/KYLIN-3709
> Project: Kylin
>  Issue Type: Bug
>  Components: Job Engine
>Affects Versions: v2.5.1, v2.5.2
>Reporter: Hongtao He
>Assignee: Shaofeng SHI
>Priority: Critical
> Attachments: 
> method-1-Pass-the-wrong-parameter-to-addResource-function-kyl.patch, 
> method-2-Pass-the-wrong-parameter-to-addResource-function-kyl.patch
>
>
> When calling Configuration.addResource, KYLIN-3648 use a Path as the 
> parameter instead of a string in Kylin's HBaseConnection.
> {code:java}
> String hdfsConfigFile = 
> KylinConfig.getInstanceFromEnv().getHBaseClusterHDFSConfigFile();
> if (hdfsConfigFile == null || hdfsConfigFile.isEmpty()) {
> return;
> }
> Configuration hdfsConf = new Configuration(false);
> hdfsConf.addResource(new Path(hdfsConfigFile));
> {code}
> Use a Path as the parameter of Configuration.addResource is better. 
> Unfortunately,  the parameter which passed to the addResource function is 
> wrong. The addResource function only accepts absolute paths, but the 
> parameter is just a filename. For example

[jira] [Comment Edited] (KYLIN-3695) Error while creating hive table through Kylin build cube with mysql imported tables

2018-12-04 Thread XiaoXiang Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3695?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16708395#comment-16708395
 ] 

XiaoXiang Yu edited comment on KYLIN-3695 at 12/4/18 8:46 AM:
--

[~gauravbrills] I debug in my test env.

I found when using decimal data type with empty precision to build a cube. The 
front end(JavaScript) will use 'undefined'  for null value and send it to 
back-end to validate cube metadata. It clear that  decimal(19, undefined) do 
not match a valid regex pattern, so back end return error message to front end 
as my above message.

Source code at 
org.apache.kylin.metadata.datatype.DataType.[https://github.com/apache/kylin/blob/master/core-metadata/src/main/java/org/apache/kylin/metadata/datatype/DataType.java]

I think it has no need to fix it because the root cause is clear.

 
       

{
"name":"SUM_INTEREST_SCORE2",
"function":{
"expression":"SUM",
"returntype":"decimal(19,undefined)",
"parameter":{
"type":"column",
"value":"USERACTION.INTEREST_SCORE2"
}
}
}


was (Author: hit_lacus):
[~gauravbrills] I debug in my test env.

I found when using decimal data type with empty precision to build a cube. The 
front end(JavaScript) will use 'undefined'  for null value and send it to 
back-end to validate cube metadata. It clear that  decimal(19, undefined) do 
not match a valid regex pattern, so back end return error message to front end 
as my above message.

Source code at 
org.apache.kylin.metadata.datatype.DataType.[https://github.com/apache/kylin/blob/master/core-metadata/src/main/java/org/apache/kylin/metadata/datatype/DataType.java]

I think it has no need to fix it because the root cause is clear.

 
   
{quote}{
 "name":"SUM_INTEREST_SCORE2",
 "function":{
 "expression":"SUM",
 "returntype":"decimal(19,undefined)",
 "parameter":
Unknown macro: \{ "type"}
}
 }
{quote}

> Error while creating hive table through Kylin build cube with mysql imported 
> tables
> ---
>
> Key: KYLIN-3695
> URL: https://issues.apache.org/jira/browse/KYLIN-3695
> Project: Kylin
>  Issue Type: Bug
>  Components: Integration
>Affects Versions: v2.5.0
>Reporter: Gaurav Rawat
>Assignee: XiaoXiang Yu
>Priority: Minor
> Attachments: image-2018-12-04-15-53-49-983.png
>
>
> HI I am trying to build a cube with Kylin, the data gets souced fine from 
> sqoop but the next step for creating hive tables fails . Looking at the 
> command being fired it looks weird as the create statement looks good to me .
> I think the issue is with DOUBLE types as when I remove the same the create 
> statement works fine . Can someone please help .
> I am using the stack in AWS EMR, kylin 2.5 hive 2.3.0
> The errors logs with commands as as below, the table is a msql table which 
> had columns with DOUBLE type
> Command
> {code:java}
> hive -e "USE default; DROP TABLE IF EXISTS 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368; 
> CREATE EXTERNAL TABLE IF NOT EXISTS 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368 ( 
> HOLDINGS_STOCK_INVESTOR_ID string ,STOCK_INVESTORS_CHANNEL string 
> ,STOCK_STOCK_ID string ,STOCK_DOMICILE string ,STOCK_STOCK_NM string 
> ,STOCK_APPROACH string ,STOCK_STOCK_TYP string ,INVESTOR_ID string 
> ,INVESTOR_NM string ,INVESTOR_DOMICILE_CNTRY string ,CLIENT_NM string 
> ,INVESTOR_HOLDINGS_GROSS_ASSETS_USD double(22) 
> ,INVESTOR_HOLDINGS_NET_ASSETS_USD double(22) ) ROW FORMAT DELIMITED FIELDS 
> TERMINATED BY '|' STORED AS TEXTFILE LOCATION 
> 's3://wfg1tst-models/kylin/kylin_metadata/kylin-4ae3b18b-831b-da66-eb8c-7318245c4448/kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368';
>  ALTER TABLE 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368 SET 
> TBLPROPERTIES('auto.purge'='true'); " --hiveconf hive.merge.mapredfiles=false 
> --hiveconf hive.auto.convert.join=true --hiveconf dfs.replication=2 
> --hiveconf hive.exec.compress.output=true --hiveconf 
> hive.auto.convert.join.noconditionaltask=true --hiveconf 
> mapreduce.job.split.metainfo.maxsize=-1 --hiveconf hive.merge.mapfiles=false 
> --hiveconf hive.auto.convert.join.noconditionaltask.size=1 --hiveconf 
> hive.stats.autogather=true{code}
> Error is as below
> {code:java}
> OK Time taken: 1.315 seconds OK Time taken: 0.09 seconds 
> MismatchedTokenException(334!=347) at 
> org.antlr.runtime.BaseRecognizer.recoverFromMismatchedToken(BaseRecognizer.java:617)
>  at org.antlr.runtime.BaseRecognizer.match(BaseRecognizer.java:115) at 
> org.apache.hadoop.hive.ql.parse.HiveParser.createTableStatement(HiveParser.java:6179)
>  at 
> org.a

[jira] [Updated] (KYLIN-3709) Pass the wrong parameter to addResource function, kylin can not load the configuration file hbase.hdfs.xml .

2018-12-04 Thread Hongtao He (JIRA)


 [ 
https://issues.apache.org/jira/browse/KYLIN-3709?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hongtao He updated KYLIN-3709:
--
Description: 
When calling Configuration.addResource, KYLIN-3648 use a Path as the parameter 
instead of a string in Kylin's HBaseConnection.
{code:java}
String hdfsConfigFile = 
KylinConfig.getInstanceFromEnv().getHBaseClusterHDFSConfigFile();
if (hdfsConfigFile == null || hdfsConfigFile.isEmpty()) {
return;
}
Configuration hdfsConf = new Configuration(false);
hdfsConf.addResource(new Path(hdfsConfigFile));
{code}
Use a Path as the parameter of Configuration.addResource is better. 
Unfortunately,  the parameter which passed to the addResource function is 
wrong. The addResource function only accepts absolute paths, but the parameter 
is just a filename. For example, the value of hdfsConfigFile is 
"hbase.hdfs.xml", so addResource function will not work . The end result is 
that kylin can not load the hbase configuration file hbase.hdfs.xml .

There are two ways to fix this bug, and I think method 1 is better.

Method-1.revert the code
{code:java}
String hdfsConfigFile = 
KylinConfig.getInstanceFromEnv().getHBaseClusterHDFSConfigFile();
if (hdfsConfigFile == null || hdfsConfigFile.isEmpty()) {
return;
}
Configuration hdfsConf = new Configuration(false);
hdfsConf.addResource(hdfsConfigFile);{code}
Method-2.Get the absolute path of the configuration file
{code:java}
String hdfsConfigFile = 
KylinConfig.getInstanceFromEnv().getHBaseClusterHDFSConfigFile();
if (hdfsConfigFile == null || hdfsConfigFile.isEmpty()) {
return;
}
Configuration hdfsConf = new Configuration(false);
String hbaseHdfsConfigPath = System.getProperty("user.dir") + "/../conf/" + 
hdfsConfigFile;
hdfsConf.addResource(new Path(hbaseHdfsConfigPath));
{code}
 

  was:
When calling Configuration.addResource, KYLIN-3648 use a Path as the parameter 
instead of a string in Kylin's HBaseConnection.
{code:java}
String hdfsConfigFile = 
KylinConfig.getInstanceFromEnv().getHBaseClusterHDFSConfigFile();
if (hdfsConfigFile == null || hdfsConfigFile.isEmpty()) {
return;
}
Configuration hdfsConf = new Configuration(false);
hdfsConf.addResource(new Path(hdfsConfigFile));
{code}
Use a Path as the parameter of Configuration.addResource is better. 
Unfortunately,  the parameter which passed to the addResource function is 
wrong. The addResource function only accepts absolute paths, but the parameter 
is just a filename. For example, the value of 

hdfsConfigFile is "hbase.hdfs.xml", so addResource function will not work . The 
end result is that kylin can not load the hbase configuration file 
hbase.hdfs.xml .

There are two ways to fix this bug, and I think method 1 is better.

Method-1.revert the code
{code:java}
String hdfsConfigFile = 
KylinConfig.getInstanceFromEnv().getHBaseClusterHDFSConfigFile();
if (hdfsConfigFile == null || hdfsConfigFile.isEmpty()) {
return;
}
Configuration hdfsConf = new Configuration(false);
hdfsConf.addResource(hdfsConfigFile);{code}
Method-2.Get the absolute path of the configuration file
{code:java}
String hdfsConfigFile = 
KylinConfig.getInstanceFromEnv().getHBaseClusterHDFSConfigFile();
if (hdfsConfigFile == null || hdfsConfigFile.isEmpty()) {
return;
}
Configuration hdfsConf = new Configuration(false);
String hbaseHdfsConfigPath = System.getProperty("user.dir") + "/../conf/" + 
hdfsConfigFile;
hdfsConf.addResource(new Path(hbaseHdfsConfigPath));
{code}
 


> Pass the wrong parameter to addResource function, kylin can not load the 
> configuration file hbase.hdfs.xml .
> 
>
> Key: KYLIN-3709
> URL: https://issues.apache.org/jira/browse/KYLIN-3709
> Project: Kylin
>  Issue Type: Bug
>  Components: Job Engine
>Affects Versions: v2.5.1, v2.5.2
>Reporter: Hongtao He
>Assignee: Shaofeng SHI
>Priority: Critical
> Attachments: 
> method-1-Pass-the-wrong-parameter-to-addResource-function-kyl.patch, 
> method-2-Pass-the-wrong-parameter-to-addResource-function-kyl.patch
>
>
> When calling Configuration.addResource, KYLIN-3648 use a Path as the 
> parameter instead of a string in Kylin's HBaseConnection.
> {code:java}
> String hdfsConfigFile = 
> KylinConfig.getInstanceFromEnv().getHBaseClusterHDFSConfigFile();
> if (hdfsConfigFile == null || hdfsConfigFile.isEmpty()) {
> return;
> }
> Configuration hdfsConf = new Configuration(false);
> hdfsConf.addResource(new Path(hdfsConfigFile));
> {code}
> Use a Path as the parameter of Configuration.addResource is better. 
> Unfortunately,  the parameter which passed to the addResource function is 
> wrong. The addResource function only accepts absolute paths, but the 
> parameter is just a filename. For example, the value of hdfsConfigFile is 
> "hbase.hdfs.xml", s

[jira] [Assigned] (KYLIN-3680) Spark cubing failed with JDBC resource

2018-12-04 Thread Chao Long (JIRA)


 [ 
https://issues.apache.org/jira/browse/KYLIN-3680?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chao Long reassigned KYLIN-3680:


Assignee: Chao Long  (was: Zhixiong Chen)

> Spark cubing failed with JDBC resource
> --
>
> Key: KYLIN-3680
> URL: https://issues.apache.org/jira/browse/KYLIN-3680
> Project: Kylin
>  Issue Type: Bug
>  Components: Spark Engine
>Affects Versions: v2.4.0, v2.4.1, v2.5.0, v2.5.1
>Reporter: Zhixiong Chen
>Assignee: Chao Long
>Priority: Major
> Fix For: v2.6.0
>
>
> I find in JdbcHiveMRInput.java , the  system save flat table with TextFile.
> However in SparkCubingByLayer.java, the system getFlatTableStorageFormat is 
> SequenceFile .It read the flat table as SequenceFile and failed.
> I workaround with add property 
> "kylin.source.hive.flat-table-storage-format=TEXTFILE" in cube.
> But I think maybe we could do it better.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (KYLIN-3695) Error while creating hive table through Kylin build cube with mysql imported tables

2018-12-04 Thread XiaoXiang Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3695?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16708395#comment-16708395
 ] 

XiaoXiang Yu edited comment on KYLIN-3695 at 12/4/18 8:45 AM:
--

[~gauravbrills] I debug in my test env.

I found when using decimal data type with empty precision to build a cube. The 
front end(JavaScript) will use 'undefined'  for null value and send it to 
back-end to validate cube metadata. It clear that  decimal(19, undefined) do 
not match a valid regex pattern, so back end return error message to front end 
as my above message.

Source code at 
org.apache.kylin.metadata.datatype.DataType.[https://github.com/apache/kylin/blob/master/core-metadata/src/main/java/org/apache/kylin/metadata/datatype/DataType.java]

I think it has no need to fix it because the root cause is clear.

 
   
{quote}{
 "name":"SUM_INTEREST_SCORE2",
 "function":{
 "expression":"SUM",
 "returntype":"decimal(19,undefined)",
 "parameter":
Unknown macro: \{ "type"}
}
 }
{quote}


was (Author: hit_lacus):
[~gauravbrills] I debug in my test env.

I found when using decimal data type with empty precision to build a cube. The 
front end(JavaScript) will use 'undefined'  for null value and send it to 
back-end to validate cube metadata. It clear that  decimal(19, undefined) do 
not match a valid regex pattern, so back end return error message to front end 
as my above message.

I think it has no need to fix it because the root cause is clear.

 
  
{quote}{
"name":"SUM_INTEREST_SCORE2",
"function":{
"expression":"SUM",
"returntype":"decimal(19,undefined)",
"parameter":{
"type":"column",
"value":"USERACTION.INTEREST_SCORE2"
}
}
}{quote}

> Error while creating hive table through Kylin build cube with mysql imported 
> tables
> ---
>
> Key: KYLIN-3695
> URL: https://issues.apache.org/jira/browse/KYLIN-3695
> Project: Kylin
>  Issue Type: Bug
>  Components: Integration
>Affects Versions: v2.5.0
>Reporter: Gaurav Rawat
>Assignee: XiaoXiang Yu
>Priority: Minor
> Attachments: image-2018-12-04-15-53-49-983.png
>
>
> HI I am trying to build a cube with Kylin, the data gets souced fine from 
> sqoop but the next step for creating hive tables fails . Looking at the 
> command being fired it looks weird as the create statement looks good to me .
> I think the issue is with DOUBLE types as when I remove the same the create 
> statement works fine . Can someone please help .
> I am using the stack in AWS EMR, kylin 2.5 hive 2.3.0
> The errors logs with commands as as below, the table is a msql table which 
> had columns with DOUBLE type
> Command
> {code:java}
> hive -e "USE default; DROP TABLE IF EXISTS 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368; 
> CREATE EXTERNAL TABLE IF NOT EXISTS 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368 ( 
> HOLDINGS_STOCK_INVESTOR_ID string ,STOCK_INVESTORS_CHANNEL string 
> ,STOCK_STOCK_ID string ,STOCK_DOMICILE string ,STOCK_STOCK_NM string 
> ,STOCK_APPROACH string ,STOCK_STOCK_TYP string ,INVESTOR_ID string 
> ,INVESTOR_NM string ,INVESTOR_DOMICILE_CNTRY string ,CLIENT_NM string 
> ,INVESTOR_HOLDINGS_GROSS_ASSETS_USD double(22) 
> ,INVESTOR_HOLDINGS_NET_ASSETS_USD double(22) ) ROW FORMAT DELIMITED FIELDS 
> TERMINATED BY '|' STORED AS TEXTFILE LOCATION 
> 's3://wfg1tst-models/kylin/kylin_metadata/kylin-4ae3b18b-831b-da66-eb8c-7318245c4448/kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368';
>  ALTER TABLE 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368 SET 
> TBLPROPERTIES('auto.purge'='true'); " --hiveconf hive.merge.mapredfiles=false 
> --hiveconf hive.auto.convert.join=true --hiveconf dfs.replication=2 
> --hiveconf hive.exec.compress.output=true --hiveconf 
> hive.auto.convert.join.noconditionaltask=true --hiveconf 
> mapreduce.job.split.metainfo.maxsize=-1 --hiveconf hive.merge.mapfiles=false 
> --hiveconf hive.auto.convert.join.noconditionaltask.size=1 --hiveconf 
> hive.stats.autogather=true{code}
> Error is as below
> {code:java}
> OK Time taken: 1.315 seconds OK Time taken: 0.09 seconds 
> MismatchedTokenException(334!=347) at 
> org.antlr.runtime.BaseRecognizer.recoverFromMismatchedToken(BaseRecognizer.java:617)
>  at org.antlr.runtime.BaseRecognizer.match(BaseRecognizer.java:115) at 
> org.apache.hadoop.hive.ql.parse.HiveParser.createTableStatement(HiveParser.java:6179)
>  at 
> org.apache.hadoop.hive.ql.parse.HiveParser.ddlStatement(HiveParser.java:3808) 
> at 
> org.apache.hadoop.hive.ql.parse.HiveParser.execStatement(HiveParser.java:2382)
>  at 
> or

[jira] [Commented] (KYLIN-3695) Error while creating hive table through Kylin build cube with mysql imported tables

2018-12-04 Thread XiaoXiang Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/KYLIN-3695?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16708395#comment-16708395
 ] 

XiaoXiang Yu commented on KYLIN-3695:
-

[~gauravbrills] I debug in my test env.

I found when using decimal data type with empty precision to build a cube. The 
front end(JavaScript) will use 'undefined'  for null value and send it to 
back-end to validate cube metadata. It clear that  decimal(19, undefined) do 
not match a valid regex pattern, so back end return error message to front end 
as my above message.

I think it has no need to fix it because the root cause is clear.

 
  
{quote}{
"name":"SUM_INTEREST_SCORE2",
"function":{
"expression":"SUM",
"returntype":"decimal(19,undefined)",
"parameter":{
"type":"column",
"value":"USERACTION.INTEREST_SCORE2"
}
}
}{quote}

> Error while creating hive table through Kylin build cube with mysql imported 
> tables
> ---
>
> Key: KYLIN-3695
> URL: https://issues.apache.org/jira/browse/KYLIN-3695
> Project: Kylin
>  Issue Type: Bug
>  Components: Integration
>Affects Versions: v2.5.0
>Reporter: Gaurav Rawat
>Assignee: XiaoXiang Yu
>Priority: Minor
> Attachments: image-2018-12-04-15-53-49-983.png
>
>
> HI I am trying to build a cube with Kylin, the data gets souced fine from 
> sqoop but the next step for creating hive tables fails . Looking at the 
> command being fired it looks weird as the create statement looks good to me .
> I think the issue is with DOUBLE types as when I remove the same the create 
> statement works fine . Can someone please help .
> I am using the stack in AWS EMR, kylin 2.5 hive 2.3.0
> The errors logs with commands as as below, the table is a msql table which 
> had columns with DOUBLE type
> Command
> {code:java}
> hive -e "USE default; DROP TABLE IF EXISTS 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368; 
> CREATE EXTERNAL TABLE IF NOT EXISTS 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368 ( 
> HOLDINGS_STOCK_INVESTOR_ID string ,STOCK_INVESTORS_CHANNEL string 
> ,STOCK_STOCK_ID string ,STOCK_DOMICILE string ,STOCK_STOCK_NM string 
> ,STOCK_APPROACH string ,STOCK_STOCK_TYP string ,INVESTOR_ID string 
> ,INVESTOR_NM string ,INVESTOR_DOMICILE_CNTRY string ,CLIENT_NM string 
> ,INVESTOR_HOLDINGS_GROSS_ASSETS_USD double(22) 
> ,INVESTOR_HOLDINGS_NET_ASSETS_USD double(22) ) ROW FORMAT DELIMITED FIELDS 
> TERMINATED BY '|' STORED AS TEXTFILE LOCATION 
> 's3://wfg1tst-models/kylin/kylin_metadata/kylin-4ae3b18b-831b-da66-eb8c-7318245c4448/kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368';
>  ALTER TABLE 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368 SET 
> TBLPROPERTIES('auto.purge'='true'); " --hiveconf hive.merge.mapredfiles=false 
> --hiveconf hive.auto.convert.join=true --hiveconf dfs.replication=2 
> --hiveconf hive.exec.compress.output=true --hiveconf 
> hive.auto.convert.join.noconditionaltask=true --hiveconf 
> mapreduce.job.split.metainfo.maxsize=-1 --hiveconf hive.merge.mapfiles=false 
> --hiveconf hive.auto.convert.join.noconditionaltask.size=1 --hiveconf 
> hive.stats.autogather=true{code}
> Error is as below
> {code:java}
> OK Time taken: 1.315 seconds OK Time taken: 0.09 seconds 
> MismatchedTokenException(334!=347) at 
> org.antlr.runtime.BaseRecognizer.recoverFromMismatchedToken(BaseRecognizer.java:617)
>  at org.antlr.runtime.BaseRecognizer.match(BaseRecognizer.java:115) at 
> org.apache.hadoop.hive.ql.parse.HiveParser.createTableStatement(HiveParser.java:6179)
>  at 
> org.apache.hadoop.hive.ql.parse.HiveParser.ddlStatement(HiveParser.java:3808) 
> at 
> org.apache.hadoop.hive.ql.parse.HiveParser.execStatement(HiveParser.java:2382)
>  at 
> org.apache.hadoop.hive.ql.parse.HiveParser.statement(HiveParser.java:1333) at 
> org.apache.hadoop.hive.ql.parse.ParseDriver.parse(ParseDriver.java:204) at 
> org.apache.hadoop.hive.ql.parse.ParseUtils.parse(ParseUtils.java:77) at 
> org.apache.hadoop.hive.ql.parse.ParseUtils.parse(ParseUtils.java:70) at 
> org.apache.hadoop.hive.ql.Driver.compile(Driver.java:468) at 
> org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1316) at 
> org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1456) at 
> org.apache.hadoop.hive.ql.Driver.run(Driver.java:1236) at 
> org.apache.hadoop.hive.ql.Driver.run(Driver.java:1226) at 
> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:233) at 
> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:184) at 
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:403) at 
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:3

[jira] [Created] (KYLIN-3709) Pass the wrong parameter to addResource function, kylin can not load the configuration file hbase.hdfs.xml .

2018-12-04 Thread Hongtao He (JIRA)
Hongtao He created KYLIN-3709:
-

 Summary: Pass the wrong parameter to addResource function, kylin 
can not load the configuration file hbase.hdfs.xml .
 Key: KYLIN-3709
 URL: https://issues.apache.org/jira/browse/KYLIN-3709
 Project: Kylin
  Issue Type: Bug
  Components: Job Engine
Affects Versions: v2.5.2, v2.5.1
Reporter: Hongtao He
Assignee: Shaofeng SHI
 Attachments: 
method-1-Pass-the-wrong-parameter-to-addResource-function-kyl.patch, 
method-2-Pass-the-wrong-parameter-to-addResource-function-kyl.patch

When calling Configuration.addResource, 
[KYLIN-3648|https://jira.apache.org/jira/browse/KYLIN-3648] use a Path as the 
parameter instead of a string in Kylin's HBaseConnection.

 
{code:java}
String hdfsConfigFile = 
KylinConfig.getInstanceFromEnv().getHBaseClusterHDFSConfigFile();
if (hdfsConfigFile == null || hdfsConfigFile.isEmpty()) {
return;
}
Configuration hdfsConf = new Configuration(false);
hdfsConf.addResource(new Path(hdfsConfigFile));
{code}
Use a Path as the parameter of Configuration.addResource is better. 
Unfortunately,  the parameter which passed to the addResource function is 
wrong. The addResource function only accepts absolute paths, but the parameter 
is just a filename. For example, the value of 

hdfsConfigFile is "hbase.hdfs.xml", so addResource function will not work . The 
end result is that kylin can not load the hbase configuration file 
hbase.hdfs.xml .

There are two ways to fix this bug, and I think method 1 is better.

Method-1.revert the code
{code:java}
String hdfsConfigFile = 
KylinConfig.getInstanceFromEnv().getHBaseClusterHDFSConfigFile();
if (hdfsConfigFile == null || hdfsConfigFile.isEmpty()) {
return;
}
Configuration hdfsConf = new Configuration(false);
hdfsConf.addResource(hdfsConfigFile);{code}
Method-2.Get the absolute path of the configuration file
{code:java}
String hdfsConfigFile = 
KylinConfig.getInstanceFromEnv().getHBaseClusterHDFSConfigFile();
if (hdfsConfigFile == null || hdfsConfigFile.isEmpty()) {
return;
}
Configuration hdfsConf = new Configuration(false);
String hbaseHdfsConfigPath = System.getProperty("user.dir") + "/../conf/" + 
hdfsConfigFile;
hdfsConf.addResource(new Path(hbaseHdfsConfigPath));
{code}
 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (KYLIN-3126) Query fails with "Error while compiling generated Java code" when equality condition is used, and works when equivalent IN clause is specified

2018-12-04 Thread Shaofeng SHI (JIRA)


 [ 
https://issues.apache.org/jira/browse/KYLIN-3126?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shaofeng SHI updated KYLIN-3126:

Fix Version/s: (was: v2.6.0)

> Query fails with "Error while compiling generated Java code" when equality 
> condition is used, and works when equivalent IN clause is specified
> --
>
> Key: KYLIN-3126
> URL: https://issues.apache.org/jira/browse/KYLIN-3126
> Project: Kylin
>  Issue Type: Bug
>  Components: Query Engine
>Affects Versions: v2.2.0
> Environment: HDP 2.5.6, Kylin 2.2.0, sample cube
>Reporter: Vsevolod Ostapenko
>Priority: Major
>
> The following query fails with "Error while compiling generated Java code", 
> when equality condition is used {{(d0.year_beg_dt = '2012-01-01')}} and works 
> when IN clause is used {{(d0.year_beg_dt in ('2012-01-01'))}}
> {code:sql}
>  select
> d2.country,
> count(f.item_count) items_ttl
>  from
> kylin_sales f
>  join
> kylin_cal_dt d0
>  on
> f.part_dt = d0.cal_dt
>  join 
> kylin_account d1
>  on
> f.buyer_id = d1.account_id
>  join
> kylin_country d2
>  on
> d1.account_country = d2.country
>  where
> d0.year_beg_dt = '2012-01-01'  -- blows up
> -- d0.year_beg_dt in ('2012-01-01') -- works
> and
> d2.country in ('US', 'JP')
>  group by
> d2.country
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (KYLIN-3290) Avoid calling Class#newInstance

2018-12-04 Thread Shaofeng SHI (JIRA)


 [ 
https://issues.apache.org/jira/browse/KYLIN-3290?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shaofeng SHI resolved KYLIN-3290.
-
Resolution: Fixed

> Avoid calling Class#newInstance
> ---
>
> Key: KYLIN-3290
> URL: https://issues.apache.org/jira/browse/KYLIN-3290
> Project: Kylin
>  Issue Type: Task
>Reporter: Ted Yu
>Assignee: Wu Bin
>Priority: Minor
>  Labels: jdk
> Fix For: v2.6.0
>
>
> Class#newInstance is deprecated starting in Java 9 - 
> https://bugs.openjdk.java.net/browse/JDK-6850612 - because it may throw 
> undeclared checked exceptions.
> The suggested replacement is getDeclaredConstructor().newInstance(), which 
> wraps the checked exceptions in InvocationException.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (KYLIN-3695) Error while creating hive table through Kylin build cube with mysql imported tables

2018-12-04 Thread XiaoXiang Yu (JIRA)


 [ 
https://issues.apache.org/jira/browse/KYLIN-3695?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

XiaoXiang Yu updated KYLIN-3695:

Attachment: image-2018-12-04-15-53-49-983.png

> Error while creating hive table through Kylin build cube with mysql imported 
> tables
> ---
>
> Key: KYLIN-3695
> URL: https://issues.apache.org/jira/browse/KYLIN-3695
> Project: Kylin
>  Issue Type: Bug
>  Components: Integration
>Affects Versions: v2.5.0
>Reporter: Gaurav Rawat
>Assignee: XiaoXiang Yu
>Priority: Minor
> Attachments: image-2018-12-04-15-53-49-983.png
>
>
> HI I am trying to build a cube with Kylin, the data gets souced fine from 
> sqoop but the next step for creating hive tables fails . Looking at the 
> command being fired it looks weird as the create statement looks good to me .
> I think the issue is with DOUBLE types as when I remove the same the create 
> statement works fine . Can someone please help .
> I am using the stack in AWS EMR, kylin 2.5 hive 2.3.0
> The errors logs with commands as as below, the table is a msql table which 
> had columns with DOUBLE type
> Command
> {code:java}
> hive -e "USE default; DROP TABLE IF EXISTS 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368; 
> CREATE EXTERNAL TABLE IF NOT EXISTS 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368 ( 
> HOLDINGS_STOCK_INVESTOR_ID string ,STOCK_INVESTORS_CHANNEL string 
> ,STOCK_STOCK_ID string ,STOCK_DOMICILE string ,STOCK_STOCK_NM string 
> ,STOCK_APPROACH string ,STOCK_STOCK_TYP string ,INVESTOR_ID string 
> ,INVESTOR_NM string ,INVESTOR_DOMICILE_CNTRY string ,CLIENT_NM string 
> ,INVESTOR_HOLDINGS_GROSS_ASSETS_USD double(22) 
> ,INVESTOR_HOLDINGS_NET_ASSETS_USD double(22) ) ROW FORMAT DELIMITED FIELDS 
> TERMINATED BY '|' STORED AS TEXTFILE LOCATION 
> 's3://wfg1tst-models/kylin/kylin_metadata/kylin-4ae3b18b-831b-da66-eb8c-7318245c4448/kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368';
>  ALTER TABLE 
> kylin_intermediate_fm_inv_holdings_8a1c33df_d12b_3609_13ee_39e169169368 SET 
> TBLPROPERTIES('auto.purge'='true'); " --hiveconf hive.merge.mapredfiles=false 
> --hiveconf hive.auto.convert.join=true --hiveconf dfs.replication=2 
> --hiveconf hive.exec.compress.output=true --hiveconf 
> hive.auto.convert.join.noconditionaltask=true --hiveconf 
> mapreduce.job.split.metainfo.maxsize=-1 --hiveconf hive.merge.mapfiles=false 
> --hiveconf hive.auto.convert.join.noconditionaltask.size=1 --hiveconf 
> hive.stats.autogather=true{code}
> Error is as below
> {code:java}
> OK Time taken: 1.315 seconds OK Time taken: 0.09 seconds 
> MismatchedTokenException(334!=347) at 
> org.antlr.runtime.BaseRecognizer.recoverFromMismatchedToken(BaseRecognizer.java:617)
>  at org.antlr.runtime.BaseRecognizer.match(BaseRecognizer.java:115) at 
> org.apache.hadoop.hive.ql.parse.HiveParser.createTableStatement(HiveParser.java:6179)
>  at 
> org.apache.hadoop.hive.ql.parse.HiveParser.ddlStatement(HiveParser.java:3808) 
> at 
> org.apache.hadoop.hive.ql.parse.HiveParser.execStatement(HiveParser.java:2382)
>  at 
> org.apache.hadoop.hive.ql.parse.HiveParser.statement(HiveParser.java:1333) at 
> org.apache.hadoop.hive.ql.parse.ParseDriver.parse(ParseDriver.java:204) at 
> org.apache.hadoop.hive.ql.parse.ParseUtils.parse(ParseUtils.java:77) at 
> org.apache.hadoop.hive.ql.parse.ParseUtils.parse(ParseUtils.java:70) at 
> org.apache.hadoop.hive.ql.Driver.compile(Driver.java:468) at 
> org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1316) at 
> org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1456) at 
> org.apache.hadoop.hive.ql.Driver.run(Driver.java:1236) at 
> org.apache.hadoop.hive.ql.Driver.run(Driver.java:1226) at 
> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:233) at 
> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:184) at 
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:403) at 
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:336) at 
> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:787) at 
> org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759) at 
> org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:686) at 
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>  at java.lang.reflect.Method.invoke(Method.java:498) at 
> org.apache.hadoop.util.RunJar.run(RunJar.java:221) at 
> org.apache.hadoop.util.RunJar.main(RunJar.java:136) FAILED: ParseException 
> line 15:42 mismatched input '(' expecting ) near 'double' in c