Hi Santosh, 0.7.1 hasn’t been formally released; from 0.6.x to 0.7.x we have metadata structure change; While, the web UI (cube wizard) for this change hasn’t been stabilized; So it is not strange that you got trouble when saving the cube;
@Jason, any idea about the JS error? On 2/26/15, 5:08 PM, "Santosh Akhilesh" <[email protected]> wrote: >Hi Shaofeng, > >I am using the binary distribution 0.7.1. I have not been able to save >cube >even once. I have tried creating new project and from local machine and >server machine. But I am always stuck with this error. I am never allowed >to add measures and never been able to save the cube. I also see the >kylin.log and it always tries to save cube with append mode. One thing I >need to tell that at partition stage since I don't have a big fact table >now I have not partititioned the fact table and I skip this step. Does >this >have affect in saving the cube. Is this because some metadata is available >and it tries to modify the cube? I am using latest Hadoop 2.6.6. Yes kylin >propert I have not added the jar. I will add them and check. But cube >creation failure is really puzzling me. I could see no error logs in >kylin.log. >Regards >Santosh >On Thu, 26 Feb 2015 at 1:40 pm, Shi, Shaofeng <[email protected]> wrote: > >> Which version or code branch are you using? I assume you’re using the >> stable version from master; Seems you’re trying to edit an existing cube >> to add new measurement, try refresh your browser's cache; if it still >> couldn’t be saved, try to create a new cube; >> >> The two error traces in tomcat need be taken care: >> >> 1) java.lang.NoClassDefFoundError: >>org/apache/kylin/common/mr/KylinMapper >> Please check kylin.properties file, making sure the >>“kylin.job.jar” >> points to a right jar file; It will be loaded in Map-reduce; >> >> 2) java.lang.IllegalArgumentException: No enum constant >> org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS >> This indicates your hadoop version might be old; Please check >>and >> ensure >> hadoop version is 2.2 or above. >> >> On 2/26/15, 3:21 PM, "Santoshakhilesh" <[email protected]> >> wrote: >> >> >Hi Shaofeng , >> > >> > I am using chrome , When I click on button to add measures , >>following >> >is error on chrome console. When I try to save the cube there is no >>error >> >in console. I just get a error dialog saying failed to take action and >> >gives me the JSON cube schema. >> > >> >Error on chrome debug console is as below; >> > >> > ReferenceError: CubeDescModel is not defined >> > at h.$scope.addNewMeasure (scripts.min.0.js:15984) >> > at scripts.min.0.js:180 >> > at scripts.min.0.js:197 >> > at h.$eval (scripts.min.0.js:119) >> > at h.$apply (scripts.min.0.js:119) >> > at HTMLButtonElement.<anonymous> (scripts.min.0.js:197) >> > at HTMLButtonElement.m.event.dispatch (scripts.min.0.js:3) >> > at HTMLButtonElement.r.handle >> >(scripts.min.0.js:3)scripts.min.0.js:100 (anonymous >> >function)scripts.min.0.js:77 (anonymous function)scripts.min.0.js:119 >> >h.$applyscripts.min.0.js:197 (anonymous function)scripts.min.0.js:3 >> >m.event.dispatchscripts.min.0.js:3 r.handle >> > >> > About the hive table import , I got pass the run shell command >> >exception but it still fails the hadoop log is; >> >2015-02-26 20:46:48,332 INFO [main] org.apache.hadoop.mapred.YarnChild: >> >mapreduce.cluster.local.dir for child: >> >/tmp/hadoop-root/nm-local-dir/usercache/root/appcache/appli >> cation_14249530 >> >91340_0002 >> >2015-02-26 20:46:48,776 INFO [main] >> >org.apache.hadoop.conf.Configuration.deprecation: session.id is >> >deprecated. Instead, use dfs.metrics.session-id >> >2015-02-26 20:46:49,310 INFO [main] org.apache.hadoop.mapred.Task: >>Using >> >ResourceCalculatorProcessTree : [ ] >> >2015-02-26 20:46:49,386 FATAL [main] >>org.apache.hadoop.mapred.YarnChild: >> >Error running child : java.lang.NoClassDefFoundError: >> >org/apache/kylin/common/mr/KylinMapper >> > at java.lang.ClassLoader.defineClass1(Native Method) >> > at java.lang.ClassLoader.defineClass(ClassLoader.java:800) >> > at >> >java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) >> > at java.net.URLClassLoader.defineClass(URLClassLoader.java:449) >> > at java.net.URLClassLoader.access$100(URLClassLoader.java:71) >> > at java.net.URLClassLoader$1.run(URLClassLoader.java:361) >> > at java.net.URLClassLoader$1.run(URLClassLoader.java:355) >> > at java.security.AccessController.doPrivileged(Native Method) >> > at java.net.URLClassLoader.findClass(URLClassLoader.java:354) >> > at java.lang.ClassLoader.loadClass(ClassLoader.java:425) >> > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) >> > >> >tomcat logs: >> >usage: HiveColumnCardinalityJob >> > -output <path> Output path >> > -table <table name> The hive table name >> >[pool-4-thread-2]:[2015-02-26 >> >20:47:49,936][ERROR][org.apache.kylin.job.common.HadoopShel >> lExecutable.doW >> >ork(HadoopShellExecutable.java:64)] - error execute >> >HadoopShellExecutable{id=d4730d26-7fe6-412e-9841-3288ab362c5b-00, >> >name=null, state=RUNNING} >> >java.lang.IllegalArgumentException: No enum constant >> >org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS >> > at java.lang.Enum.valueOf(Enum.java:236) >> > at >> >org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.valueOf( >> Framewo >> >rkCounterGroup.java:148) >> > at >> >org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup. >> findCounter(Fra >> >meworkCounterGroup.java:182) >> > at >> >org.apache.hadoop.mapreduce.counters.AbstractCounters.findC >> ounter(Abstract >> >Counters.java:154) >> > at >> >org.apache.hadoop.mapreduce.TypeConverter.fromYarn(TypeConv >> erter.java:240) >> > at >> >org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounte >> rs(ClientServic >> >eDelegate.java:370) >> > at >> >org.apache.hadoop.mapred.YARNRunner.getJobCounters(YARNRunner.java:511) >> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756) >> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753) >> > at java.security.AccessController.doPrivileged(Native Method) >> > at javax.security.auth.Subject.doAs(Subject.java:415) >> > at >> >org.apache.hadoop.security.UserGroupInformation.doAs(UserGr >> oupInformation. >> >java:1491) >> > at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753) >> > at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1361) >> > at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1289) >> > at >> >org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForComple >> tion(AbstractHa >> >doopJob.java:134) >> > at >> >org.apache.kylin.job.hadoop.cardinality.HiveColumnCardinalityJob.run( >> HiveC >> >olumnCardinalityJob.java:114) >> > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) >> > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84) >> > at >> >org.apache.kylin.job.common.HadoopShellExecutable.doWork(Ha >> doopShellExecut >> >able.java:62) >> > at >> >org.apache.kylin.job.execution.AbstractExecutable.execute(A >> bstractExecutab >> >le.java:99) >> > at >> >org.apache.kylin.job.execution.DefaultChainedExecutable.doW >> ork(DefaultChai >> >nedExecutable.java:50) >> > at >> >org.apache.kylin.job.execution.AbstractExecutable.execute(A >> bstractExecutab >> >le.java:99) >> > at >> >org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRu >> nner.run(Defaul >> >tScheduler.java:132) >> > at >> >java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoo >> lExecutor.java: >> >1145) >> > at >> >java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPo >> olExecutor.java >> >:615) >> > at java.lang.Thread.run(Thread.java:745) >> > >> >Regards, >> >Santosh Akhilesh >> >Bangalore R&D >> >HUAWEI TECHNOLOGIES CO.,LTD. >> > >> >www.huawei.com >> >----------------------------------------------------------- >> --------------- >> >----------------------------------------------------------- >> >This e-mail and its attachments contain confidential information from >> >HUAWEI, which >> >is intended only for the person or entity whose address is listed >>above. >> >Any use of the >> >information contained herein in any way (including, but not limited to, >> >total or partial >> >disclosure, reproduction, or dissemination) by persons other than the >> >intended >> >recipient(s) is prohibited. If you receive this e-mail in error, please >> >notify the sender by >> >phone or email immediately and delete it! >> > >> >________________________________________ >> >From: Shi, Shaofeng [[email protected]] >> >Sent: Thursday, February 26, 2015 11:32 AM >> >To: [email protected] >> >Cc: Kulbhushan Rana >> >Subject: Re: Error while making cube & Measure option is not responding >> >on GUI >> > >> >Hi Santosh, hive table importing issue should not impact on cube >>saving. >> > >> >If you couldn’t save the cube, firstly please check whether there is >>error >> >in the tomcat’s log; If not please check your web browser; We suggest >>use >> >Firefox (with firebug add-on) or Chrome, open the JS console (press >>F12) >> >and then operate web UI, check whether there is any error reported in >> >browser. >> > >> > >> >On 2/26/15, 1:08 PM, "Santoshakhilesh" <[email protected]> >> >wrote: >> > >> >>Hi Shaofeng , >> >> Thanks for replying. >> >> Yes I am checking the yarn exception, But I find that this error >>comes >> >>while importing the hive table to kylin. >> >> Even if this error comes , hive tables is exported successfully in >> >>kylin. Is this the reason why cube saving has failed ? >> >> >> >> Next step when I go on creating the cube for following schema I >>get >> >>error at last step while saving and I am unable to add any measures , >> >>clicking on measure option just dont pop up any dialog, >> >> >> >>I am using star schema with fact_sales as fact table and dim_* as >> >>dimension tables. >> >> >> >> fact_sales: >> >> storeid int >> >> itemid int >> >> custid int >> >> qty int >> >> price double >> >> >> >>dim_customer >> >> custid int >> >> name string >> >> >> >> dim_item >> >> itemid int >> >> category string >> >> brand string >> >> color string >> >> >> >>dim_store >> >> storeid int >> >> city string >> >> state string >> >> >> >>JSON is as below. >> >> The JSON is as below. >> >> >> >> { >> >> "name": "Retail_Cube", >> >> "description": "", >> >> "dimensions": [ >> >> { >> >> "name": "RETAIL.FACT_SALES.STOREID", >> >> "table": "RETAIL.FACT_SALES", >> >> "hierarchy": false, >> >> "derived": null, >> >> "column": [ >> >> "STOREID" >> >> ], >> >> "id": 1 >> >> }, >> >> { >> >> "name": "RETAIL.FACT_SALES.ITEMID", >> >> "table": "RETAIL.FACT_SALES", >> >> "hierarchy": false, >> >> "derived": null, >> >> "column": [ >> >> "ITEMID" >> >> ], >> >> "id": 2 >> >> }, >> >> { >> >> "name": "RETAIL.FACT_SALES.CUSTID", >> >> "table": "RETAIL.FACT_SALES", >> >> "hierarchy": false, >> >> "derived": null, >> >> "column": [ >> >> "CUSTID" >> >> ], >> >> "id": 3 >> >> } >> >> ], >> >> "measures": [ >> >> { >> >> "id": 1, >> >> "name": "_COUNT_", >> >> "function": { >> >> "expression": "COUNT", >> >> "returntype": "bigint", >> >> "parameter": { >> >> "type": "constant", >> >> "value": 1 >> >> } >> >> } >> >> } >> >> ], >> >> "rowkey": { >> >> "rowkey_columns": [ >> >> { >> >> "column": "STOREID", >> >> "length": 0, >> >> "dictionary": "true", >> >> "mandatory": false >> >> }, >> >> { >> >> "column": "ITEMID", >> >> "length": 0, >> >> "dictionary": "true", >> >> "mandatory": false >> >> }, >> >> { >> >> "column": "CUSTID", >> >> "length": 0, >> >> "dictionary": "true", >> >> "mandatory": false >> >> } >> >> ], >> >> "aggregation_groups": [ >> >> [ >> >> "STOREID", >> >> "ITEMID", >> >> "CUSTID" >> >> ] >> >> ] >> >> }, >> >> "notify_list": [], >> >> "capacity": "", >> >> "hbase_mapping": { >> >> "column_family": [ >> >> { >> >> "name": "f1", >> >> "columns": [ >> >> { >> >> "qualifier": "m", >> >> "measure_refs": [ >> >> "_COUNT_" >> >> ] >> >> } >> >> ] >> >> } >> >> ] >> >> }, >> >> "project": "RetailProject", >> >> "model_name": "Retail_Cube" >> >> } >> >> >> >>Regards, >> >>Santosh Akhilesh >> >>Bangalore R&D >> >>HUAWEI TECHNOLOGIES CO.,LTD. >> >> >> >>www.huawei.com >> >>---------------------------------------------------------- >> --------------- >> >>- >> >>----------------------------------------------------------- >> >>This e-mail and its attachments contain confidential information from >> >>HUAWEI, which >> >>is intended only for the person or entity whose address is listed >>above. >> >>Any use of the >> >>information contained herein in any way (including, but not limited >>to, >> >>total or partial >> >>disclosure, reproduction, or dissemination) by persons other than the >> >>intended >> >>recipient(s) is prohibited. If you receive this e-mail in error, >>please >> >>notify the sender by >> >>phone or email immediately and delete it! >> >> >> >>________________________________________ >> >>From: Shi, Shaofeng [[email protected]] >> >>Sent: Thursday, February 26, 2015 7:01 AM >> >>To: [email protected] >> >>Subject: Re: Error while making cube & Measure option is not >>responding >> >>on GUI >> >> >> >>Hi Santosh, >> >> >> >>It looks like hadoop failed to execute some shell command in the >> >>container; You need dive into hadoop to see what¹s the concrete error. >> >>You >> >>can use yarn logs command to fetch all logs: >> >> >> >>yarn logs -applicationId <app_id> >> >> >> >> >> >>On 2/25/15, 7:39 PM, "Santosh Akhilesh" <[email protected]> >> >>wrote: >> >> >> >>>Hi Luke / Shaofeng , >> >>> Can you please help me to check this issue. >> >>>Regards, >> >>>Santosh Akhilesh >> >>> >> >>>On Tue, Feb 24, 2015 at 10:41 PM, Santosh Akhilesh < >> >>>[email protected]> wrote: >> >>> >> >>>> Hi All , >> >>>> is it because of following error in map reduce job ? what >> >>>>could >> >>>>be >> >>>> way to resolve this , a google search says that its issue of Yarn >> >>>>class >> >>>> path , but I am not sure what it is ? >> >>>> >> >>>> Kylin Hive Column Cardinality Job table=RETAIL.FACT_SALES >> >>>> output=/tmp/cardinality/RETAIL.FACT_SALES >> >>>> >> >>>> Application application_1424791969399_0008 failed 2 times due to AM >> >>>> Container for appattempt_1424791969399_0008_000002 exited with >> >>>>exitCode: 1 >> >>>> For more detailed output, check application tracking page: >> >>>> http://santosh:8088/proxy/application_1424791969399_0008/Then, >>click >> >>>>on >> >>>> links to logs of each attempt. >> >>>> Diagnostics: Exception from container-launch. >> >>>> Container id: container_1424791969399_0008_02_000001 >> >>>> Exit code: 1 >> >>>> Stack trace: ExitCodeException exitCode=1: >> >>>> at org.apache.hadoop.util.Shell.runCommand(Shell.java:538) >> >>>> at org.apache.hadoop.util.Shell.run(Shell.java:455) >> >>>> at >> >>>> >> >>>>org.apache.hadoop.util.Shell$ShellCommandExecutor.execut >> e(Shell.java:71 >> >>>>5 >> >>>>) >> >>>> at >> >>>> >> >>>>org.apache.hadoop.yarn.server.nodemanager.DefaultContain >> erExecutor.laun >> >>>>c >> >>>>h >> >>>>Container(DefaultContainerExecutor.java:211) >> >>>> at >> >>>> >> >>>>org.apache.hadoop.yarn.server.nodemanager.containermanag >> er.launcher.Con >> >>>>t >> >>>>a >> >>>>inerLaunch.call(ContainerLaunch.java:302) >> >>>> at >> >>>> >> >>>>org.apache.hadoop.yarn.server.nodemanager.containermanag >> er.launcher.Con >> >>>>t >> >>>>a >> >>>>inerLaunch.call(ContainerLaunch.java:82) >> >>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262) >> >>>> at >> >>>> >> >>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread >> PoolExecutor.ja >> >>>>v >> >>>>a >> >>>>:1145) >> >>>> at >> >>>> >> >>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa >> dPoolExecutor.j >> >>>>a >> >>>>v >> >>>>a:615) >> >>>> at java.lang.Thread.run(Thread.java:745) >> >>>> Container exited with a non-zero exit code 1 >> >>>> Failing this attempt. Failing the application. >> >>>> >> >>>> ---------- Forwarded message ---------- >> >>>> From: Santoshakhilesh <[email protected]> >> >>>> Date: Tue, Feb 24, 2015 at 7:41 PM >> >>>> Subject: FW: Error while making cube & Measure option is not >> >>>>responding >> >>>>on >> >>>> GUI >> >>>> To: "[email protected]" >><[email protected]> >> >>>> >> >>>> >> >>>> hi , >> >>>> please someone give me a hand to resolve this issue , thanks. >> >>>> >> >>>> Regards, >> >>>> Santosh Akhilesh >> >>>> Bangalore R&D >> >>>> HUAWEI TECHNOLOGIES CO.,LTD. >> >>>> >> >>>> www.huawei.com >> >>>> >> >>>> >> >>>>-------------------------------------------------------- >> --------------- >> >>>>- >> >>>>- >> >>>>------------------------------------------------------------ >> >>>> This e-mail and its attachments contain confidential information >>from >> >>>> HUAWEI, which >> >>>> is intended only for the person or entity whose address is listed >> >>>>above. >> >>>> Any use of the >> >>>> information contained herein in any way (including, but not limited >> >>>>to, >> >>>> total or partial >> >>>> disclosure, reproduction, or dissemination) by persons other than >>the >> >>>> intended >> >>>> recipient(s) is prohibited. If you receive this e-mail in error, >> >>>>please >> >>>> notify the sender by >> >>>> phone or email immediately and delete it! >> >>>> >> >>>> ________________________________________ >> >>>> From: Santoshakhilesh [[email protected]] >> >>>> Sent: Tuesday, February 24, 2015 12:55 PM >> >>>> To: [email protected] >> >>>> Cc: Kulbhushan Rana >> >>>> Subject: FW: Error while making cube & Measure option is not >> >>>>responding >> >>>>on >> >>>> GUI >> >>>> >> >>>> 2. If I ignore and continue and try to save the cube I get an >> >>>>exception >> >>>> in Kylin.log , I have checked the path is set correctly and >> >>>>HCatInputFormat >> >>>> this file is present in hive-hcatalog-core-0.14.0.jar . Please let >>me >> >>>>know >> >>>> what can I do to resolve this ? >> >>>> >> >>>> -- This was path issue , now no more exception in kylin.log >> >>>> >> >>>> But saveing cube still fails with error. And still can't add >>measures. >> >>>> >> >>>> Error Message >> >>>> Failed to take action. >> >>>> >> >>>> In log I can find no exception. Following is the last log in >>kylin.log >> >>>> >> >>>> [pool-3-thread-1]:[2015-02-24 >> >>>> >> >>>>20:47:15,613][INFO][org.apache.kylin.job.impl.threadpool >> .DefaultSchedul >> >>>>e >> >>>>r >> >>>>$FetcherRunner.run(DefaultScheduler.java:117)] >> >>>> - Job Fetcher: 0 running, 0 actual running, 0 ready, 6 others >> >>>> [http-bio-7070-exec-2]:[2015-02-24 >> >>>> >> >>>>20:47:51,610][DEBUG][org.apache.kylin.rest.controller.Cu >> beController.de >> >>>>s >> >>>>e >> >>>>rializeDataModelDesc(CubeController.java:459)] >> >>>> - Saving cube { >> >>>> "name": "", >> >>>> "fact_table": "RETAIL.FACT_SALES", >> >>>> "lookups": [], >> >>>> "filter_condition": "", >> >>>> "capacity": "SMALL", >> >>>> "partition_desc": { >> >>>> "partition_date_column": "", >> >>>> "partition_date_start": 0, >> >>>> "partition_type": "APPEND" >> >>>> }, >> >>>> "last_modified": 0 >> >>>> } >> >>>> >> >>>> >> >>>> local access logs all with 200 , so seems ok. >> >>>> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:46:56 +0800] "GET >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:07 +0800] "GET >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:27 +0800] "GET >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:28 +0800] "GET >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:34 +0800] "GET >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:48 +0800] "GET >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:51 +0800] "POST >>/kylin/api/cubes >> >>>> HTTP/1.1" 200 701 >> >>>> >> >>>> >> >>>> Regards, >> >>>> Santosh Akhilesh >> >>>> Bangalore R&D >> >>>> HUAWEI TECHNOLOGIES CO.,LTD. >> >>>> >> >>>> www.huawei.com >> >>>> >> >>>> >> >>>>-------------------------------------------------------- >> --------------- >> >>>>- >> >>>>- >> >>>>------------------------------------------------------------ >> >>>> This e-mail and its attachments contain confidential information >>from >> >>>> HUAWEI, which >> >>>> is intended only for the person or entity whose address is listed >> >>>>above. >> >>>> Any use of the >> >>>> information contained herein in any way (including, but not limited >> >>>>to, >> >>>> total or partial >> >>>> disclosure, reproduction, or dissemination) by persons other than >>the >> >>>> intended >> >>>> recipient(s) is prohibited. If you receive this e-mail in error, >> >>>>please >> >>>> notify the sender by >> >>>> phone or email immediately and delete it! >> >>>> >> >>>> ________________________________________ >> >>>> From: Santoshakhilesh [[email protected]] >> >>>> Sent: Tuesday, February 24, 2015 12:09 PM >> >>>> To: [email protected] >> >>>> Cc: Kulbhushan Rana >> >>>> Subject: Error while making cube & Measure option is not >>responding on >> >>>>GUI >> >>>> >> >>>> Hi All , >> >>>> >> >>>> I am building a simple cube for test and using the binary build >> >>>>0.7.1 >> >>>> . I have following hive tables with columns. >> >>>> >> >>>> >> >>>> >> >>>> fact_sales: >> >>>> >> >>>> storeid int >> >>>> itemid int >> >>>> custid int >> >>>> qty int >> >>>> price double >> >>>> >> >>>> dim_customer >> >>>> custid int >> >>>> name string >> >>>> >> >>>> dim_item >> >>>> >> >>>> itemid int >> >>>> category string >> >>>> brand string >> >>>> color string >> >>>> >> >>>> dim_store >> >>>> >> >>>> storeid int >> >>>> city string >> >>>> state string >> >>>> >> >>>> Please help me to answer following issues; >> >>>> >> >>>> >> >>>> >> >>>> 1. When I go to measure section and click on measure option , >>there is >> >>>>no >> >>>> response , I want add measure on qty and price with sum >> >>>> >> >>>> 2. If I ignore and continue and try to save the cube I get an >> >>>>exception >> >>>> in Kylin.log , I have checked the path is set correctly and >> >>>>HCatInputFormat >> >>>> this file is present in hive-hcatalog-core-0.14.0.jar . Please let >>me >> >>>>know >> >>>> what can I do to resolve this ? >> >>>> >> >>>> 3. Also I have another question since this is a test and data is >>small >> >>>>I >> >>>> have not partitioned the fact table , is it ok to skip partition >>stage >> >>>> while cube build ? >> >>>> >> >>>> >> >>>> >> >>>> Exception >> >>>> >> >>>> pool-4-thread-4]:[2015-02-24 >> >>>> >> >>>>19:26:32,577][ERROR][org.apache.kylin.job.impl.threadpoo >> l.DefaultSchedu >> >>>>l >> >>>>e >> >>>>r$JobRunner.run(DefaultScheduler.java:134)] >> >>>> - ExecuteException job:c3532a6f-97ea-474a-b36a-218dd517cedb >> >>>> org.apache.kylin.job.exception.ExecuteException: >> >>>> org.apache.kylin.job.exception.ExecuteException: >> >>>> java.lang.NoClassDefFoundError: >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat >> >>>> at >> >>>> >> >>>>org.apache.kylin.job.execution.AbstractExecutable.execut >> e(AbstractExecu >> >>>>t >> >>>>a >> >>>>ble.java:102) >> >>>> at >> >>>> >> >>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$Jo >> bRunner.run(Def >> >>>>a >> >>>>u >> >>>>ltScheduler.java:132) >> >>>> at >> >>>> >> >>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread >> PoolExecutor.ja >> >>>>v >> >>>>a >> >>>>:1145) >> >>>> at >> >>>> >> >>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa >> dPoolExecutor.j >> >>>>a >> >>>>v >> >>>>a:615) >> >>>> at java.lang.Thread.run(Thread.java:745) >> >>>> Caused by: org.apache.kylin.job.exception.ExecuteException: >> >>>> java.lang.NoClassDefFoundError: >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat >> >>>> >> >>>> >> >>>> >> >>>> The JSON is as below. >> >>>> >> >>>> >> >>>> >> >>>> { >> >>>> "name": "Retail_Cube", >> >>>> "description": "", >> >>>> "dimensions": [ >> >>>> { >> >>>> "name": "RETAIL.FACT_SALES.STOREID", >> >>>> "table": "RETAIL.FACT_SALES", >> >>>> "hierarchy": false, >> >>>> "derived": null, >> >>>> "column": [ >> >>>> "STOREID" >> >>>> ], >> >>>> "id": 1 >> >>>> }, >> >>>> { >> >>>> "name": "RETAIL.FACT_SALES.ITEMID", >> >>>> "table": "RETAIL.FACT_SALES", >> >>>> "hierarchy": false, >> >>>> "derived": null, >> >>>> "column": [ >> >>>> "ITEMID" >> >>>> ], >> >>>> "id": 2 >> >>>> }, >> >>>> { >> >>>> "name": "RETAIL.FACT_SALES.CUSTID", >> >>>> "table": "RETAIL.FACT_SALES", >> >>>> "hierarchy": false, >> >>>> "derived": null, >> >>>> "column": [ >> >>>> "CUSTID" >> >>>> ], >> >>>> "id": 3 >> >>>> } >> >>>> ], >> >>>> "measures": [ >> >>>> { >> >>>> "id": 1, >> >>>> "name": "_COUNT_", >> >>>> "function": { >> >>>> "expression": "COUNT", >> >>>> "returntype": "bigint", >> >>>> "parameter": { >> >>>> "type": "constant", >> >>>> "value": 1 >> >>>> } >> >>>> } >> >>>> } >> >>>> ], >> >>>> "rowkey": { >> >>>> "rowkey_columns": [ >> >>>> { >> >>>> "column": "STOREID", >> >>>> "length": 0, >> >>>> "dictionary": "true", >> >>>> "mandatory": false >> >>>> }, >> >>>> { >> >>>> "column": "ITEMID", >> >>>> "length": 0, >> >>>> "dictionary": "true", >> >>>> "mandatory": false >> >>>> }, >> >>>> { >> >>>> "column": "CUSTID", >> >>>> "length": 0, >> >>>> "dictionary": "true", >> >>>> "mandatory": false >> >>>> } >> >>>> ], >> >>>> "aggregation_groups": [ >> >>>> [ >> >>>> "STOREID", >> >>>> "ITEMID", >> >>>> "CUSTID" >> >>>> ] >> >>>> ] >> >>>> }, >> >>>> "notify_list": [], >> >>>> "capacity": "", >> >>>> "hbase_mapping": { >> >>>> "column_family": [ >> >>>> { >> >>>> "name": "f1", >> >>>> "columns": [ >> >>>> { >> >>>> "qualifier": "m", >> >>>> "measure_refs": [ >> >>>> "_COUNT_" >> >>>> ] >> >>>> } >> >>>> ] >> >>>> } >> >>>> ] >> >>>> }, >> >>>> "project": "RetailProject", >> >>>> "model_name": "Retail_Cube" >> >>>> } >> >>>> >> >>>> >> >>>> >> >>>> >> >>>> >> >>>> Regards, >> >>>> Santosh Akhilesh >> >>>> Bangalore R&D >> >>>> HUAWEI TECHNOLOGIES CO.,LTD. >> >>>> >> >>>> www.huawei.com >> >>>> >> >>>> >> >>>>-------------------------------------------------------- >> --------------- >> >>>>- >> >>>>- >> >>>>------------------------------------------------------------ >> >>>> This e-mail and its attachments contain confidential information >>from >> >>>> HUAWEI, which >> >>>> is intended only for the person or entity whose address is listed >> >>>>above. >> >>>> Any use of the >> >>>> information contained herein in any way (including, but not limited >> >>>>to, >> >>>> total or partial >> >>>> disclosure, reproduction, or dissemination) by persons other than >>the >> >>>> intended >> >>>> recipient(s) is prohibited. If you receive this e-mail in error, >> >>>>please >> >>>> notify the sender by >> >>>> phone or email immediately and delete it! >> >>>> >> >>>> >> >>>> >> >>>> -- >> >>>> Regards, >> >>>> Santosh Akhilesh >> >>>> +91-0-9845482201 >> >>>> >> >>> >> >>> >> >>> >> >>>-- >> >>>Regards, >> >>>Santosh Akhilesh >> >>>+91-0-9845482201 >> >>
