Hi Shaofeng,

   About hive table import exception , I have added the jars in kylin property 
file

# Path to the local(relative to job engine) job jar, job engine will use this 
jar
kylin.job.jar=kylin-job-0.7.1-SNAPSHOT.jar
# Path to the local(relative to job engine) coprocessor jar, job engine will 
upload this file to hbase
kylin.coprocessor.local.jar=kylin-coprocessor-0.7.1-SNAPSHOT.jar

After that when I import hive table , in the yarn resource manager I can see 
the Mapreduce job as success
application_1424957178195_0010 root Kylin Hive Column Cardinality Job 
table=RETAIL.FACT_SALES output=/tmp/cardinality/RETAIL.FACT_SALES MAPREDUCE 
default Thu, 26 Feb 2015 18:36:36 GMT Thu, 26 Feb 2015 18:36:54 GMT FINISHED 
SUCCEEDED

But Kylin.log has same exception: I am using hadoop version 2.6.0


[pool-4-thread-1]:[2015-02-27 
02:36:34,902][INFO][org.apache.kylin.job.manager.ExecutableManager.updateJobOutput(ExecutableManager.java:218)]
 - job id:ef5da34c-eee2-4dac-9972-a2c4ce7914c0-00 from READY to RUNNING
[pool-4-thread-1]:[2015-02-27 
02:36:34,908][INFO][org.apache.kylin.job.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:57)]
 - parameters of the HadoopShellExecutable:
[pool-4-thread-1]:[2015-02-27 
02:36:34,908][INFO][org.apache.kylin.job.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:58)]
 - -table RETAIL.FACT_SALES -output /tmp/cardinality/RETAIL.FACT_SALES
[pool-4-thread-1]:[2015-02-27 
02:36:35,782][INFO][org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForCompletion(AbstractHadoopJob.java:122)]
 - append kylin.hive.dependency: 
/opt/hive/apache-hive-0.14.0-bin/conf/:/opt/hive/apache-hive-0.14.0-bin/lib/*:/opt/hive/apache-hive-0.14.0-bin/hcatalog/share/hcatalog/*
 to mapreduce.application.classpath
[pool-4-thread-1]:[2015-02-27 
02:36:55,928][ERROR][org.apache.kylin.job.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:64)]
 - error execute 
HadoopShellExecutable{id=ef5da34c-eee2-4dac-9972-a2c4ce7914c0-00, name=null, 
state=RUNNING}
java.lang.IllegalArgumentException: No enum constant 
org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_REDUCES
 at java.lang.Enum.valueOf(Enum.java:236)
 at 
org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.valueOf(FrameworkCounterGroup.java:148)
 at 
org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.findCounter(FrameworkCounterGroup.java:182)
 at 
org.apache.hadoop.mapreduce.counters.AbstractCounters.findCounter(AbstractCounters.java:154)
 at org.apache.hadoop.mapreduce.TypeConverter.fromYarn(TypeConverter.java:240)
 at 
org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounters(ClientServiceDelegate.java:370)
 at org.apache.hadoop.mapred.YARNRunner.getJobCounters(YARNRunner.java:511)
 at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756)
 at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:415)
 at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
 at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753)
 at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1361)
 at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1289)
 at 
org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForCompletion(AbstractHadoopJob.java:134)
 at 
org.apache.kylin.job.hadoop.cardinality.HiveColumnCardinalityJob.run(HiveColumnCardinalityJob.java:114)
 at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
 at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
 at 
org.apache.kylin.job.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:62)
 at 
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:99)
 at 
org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:50)
 at 
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:99)
 at 
org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:132)
 at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
 at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
 at java.lang.Thread.run(Thread.java:745)


Regards,
Santosh Akhilesh
Bangalore R&D
HUAWEI TECHNOLOGIES CO.,LTD.

www.huawei.com
-------------------------------------------------------------------------------------------------------------------------------------
This e-mail and its attachments contain confidential information from HUAWEI, 
which
is intended only for the person or entity whose address is listed above. Any 
use of the
information contained herein in any way (including, but not limited to, total 
or partial
disclosure, reproduction, or dissemination) by persons other than the intended
recipient(s) is prohibited. If you receive this e-mail in error, please notify 
the sender by
phone or email immediately and delete it!

________________________________________
From: Shi, Shaofeng [[email protected]]
Sent: Thursday, February 26, 2015 1:39 PM
To: [email protected]
Cc: Kulbhushan Rana
Subject: Re: Error while making cube & Measure option is not responding on GUI

Which version or code branch are you using? I assume you’re using the
stable version from master; Seems you’re trying to edit an existing cube
to add new measurement, try refresh your browser's cache; if it still
couldn’t be saved, try to create a new cube;

The two error traces in tomcat need be taken care:

1) java.lang.NoClassDefFoundError: org/apache/kylin/common/mr/KylinMapper
        Please check kylin.properties file, making sure the “kylin.job.jar”
points to a right jar file; It will be loaded in Map-reduce;

2) java.lang.IllegalArgumentException: No enum constant
org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS
        This indicates your hadoop version might be old; Please check and ensure
hadoop version is 2.2 or above.

On 2/26/15, 3:21 PM, "Santoshakhilesh" <[email protected]> wrote:

>Hi Shaofeng ,
>
>   I am using chrome , When I click on button to add measures , following
>is error on chrome console. When I try to save the cube there is no error
>in console. I just get a error dialog saying failed to take action and
>gives me the JSON cube schema.
>
>Error on chrome debug console is as below;
>
> ReferenceError: CubeDescModel is not defined
>    at h.$scope.addNewMeasure (scripts.min.0.js:15984)
>    at scripts.min.0.js:180
>    at scripts.min.0.js:197
>    at h.$eval (scripts.min.0.js:119)
>    at h.$apply (scripts.min.0.js:119)
>    at HTMLButtonElement.<anonymous> (scripts.min.0.js:197)
>    at HTMLButtonElement.m.event.dispatch (scripts.min.0.js:3)
>    at HTMLButtonElement.r.handle
>(scripts.min.0.js:3)scripts.min.0.js:100 (anonymous
>function)scripts.min.0.js:77 (anonymous function)scripts.min.0.js:119
>h.$applyscripts.min.0.js:197 (anonymous function)scripts.min.0.js:3
>m.event.dispatchscripts.min.0.js:3 r.handle
>
>   About the hive table import , I got pass the run shell command
>exception but it still fails the hadoop log is;
>2015-02-26 20:46:48,332 INFO [main] org.apache.hadoop.mapred.YarnChild:
>mapreduce.cluster.local.dir for child:
>/tmp/hadoop-root/nm-local-dir/usercache/root/appcache/application_14249530
>91340_0002
>2015-02-26 20:46:48,776 INFO [main]
>org.apache.hadoop.conf.Configuration.deprecation: session.id is
>deprecated. Instead, use dfs.metrics.session-id
>2015-02-26 20:46:49,310 INFO [main] org.apache.hadoop.mapred.Task:  Using
>ResourceCalculatorProcessTree : [ ]
>2015-02-26 20:46:49,386 FATAL [main] org.apache.hadoop.mapred.YarnChild:
>Error running child : java.lang.NoClassDefFoundError:
>org/apache/kylin/common/mr/KylinMapper
> at java.lang.ClassLoader.defineClass1(Native Method)
> at java.lang.ClassLoader.defineClass(ClassLoader.java:800)
> at
>java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
> at java.net.URLClassLoader.defineClass(URLClassLoader.java:449)
> at java.net.URLClassLoader.access$100(URLClassLoader.java:71)
> at java.net.URLClassLoader$1.run(URLClassLoader.java:361)
> at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
> at java.security.AccessController.doPrivileged(Native Method)
> at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
>
>tomcat logs:
>usage: HiveColumnCardinalityJob
> -output <path>        Output path
> -table <table name>   The hive table name
>[pool-4-thread-2]:[2015-02-26
>20:47:49,936][ERROR][org.apache.kylin.job.common.HadoopShellExecutable.doW
>ork(HadoopShellExecutable.java:64)] - error execute
>HadoopShellExecutable{id=d4730d26-7fe6-412e-9841-3288ab362c5b-00,
>name=null, state=RUNNING}
>java.lang.IllegalArgumentException: No enum constant
>org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS
> at java.lang.Enum.valueOf(Enum.java:236)
> at
>org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.valueOf(Framewo
>rkCounterGroup.java:148)
> at
>org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.findCounter(Fra
>meworkCounterGroup.java:182)
> at
>org.apache.hadoop.mapreduce.counters.AbstractCounters.findCounter(Abstract
>Counters.java:154)
> at
>org.apache.hadoop.mapreduce.TypeConverter.fromYarn(TypeConverter.java:240)
> at
>org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounters(ClientServic
>eDelegate.java:370)
> at
>org.apache.hadoop.mapred.YARNRunner.getJobCounters(YARNRunner.java:511)
> at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756)
> at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:415)
> at
>org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.
>java:1491)
> at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753)
> at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1361)
> at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1289)
> at
>org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForCompletion(AbstractHa
>doopJob.java:134)
> at
>org.apache.kylin.job.hadoop.cardinality.HiveColumnCardinalityJob.run(HiveC
>olumnCardinalityJob.java:114)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
> at
>org.apache.kylin.job.common.HadoopShellExecutable.doWork(HadoopShellExecut
>able.java:62)
> at
>org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutab
>le.java:99)
> at
>org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChai
>nedExecutable.java:50)
> at
>org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutab
>le.java:99)
> at
>org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(Defaul
>tScheduler.java:132)
> at
>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:
>1145)
> at
>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java
>:615)
> at java.lang.Thread.run(Thread.java:745)
>
>Regards,
>Santosh Akhilesh
>Bangalore R&D
>HUAWEI TECHNOLOGIES CO.,LTD.
>
>www.huawei.com
>--------------------------------------------------------------------------
>-----------------------------------------------------------
>This e-mail and its attachments contain confidential information from
>HUAWEI, which
>is intended only for the person or entity whose address is listed above.
>Any use of the
>information contained herein in any way (including, but not limited to,
>total or partial
>disclosure, reproduction, or dissemination) by persons other than the
>intended
>recipient(s) is prohibited. If you receive this e-mail in error, please
>notify the sender by
>phone or email immediately and delete it!
>
>________________________________________
>From: Shi, Shaofeng [[email protected]]
>Sent: Thursday, February 26, 2015 11:32 AM
>To: [email protected]
>Cc: Kulbhushan Rana
>Subject: Re: Error while making cube & Measure option is not responding
>on GUI
>
>Hi Santosh, hive table importing issue should not impact on cube saving.
>
>If you couldn’t save the cube, firstly please check whether there is error
>in the tomcat’s log; If not please check your web browser; We suggest use
>Firefox (with firebug add-on) or Chrome, open the JS console (press F12)
>and then operate web UI, check whether there is any error reported in
>browser.
>
>
>On 2/26/15, 1:08 PM, "Santoshakhilesh" <[email protected]>
>wrote:
>
>>Hi Shaofeng ,
>>   Thanks for replying.
>>   Yes I am checking the yarn exception, But I find that this error comes
>>while importing the hive table to kylin.
>>   Even if this error comes , hive tables is exported successfully in
>>kylin. Is this the reason why cube saving has failed ?
>>
>>   Next step when I go on creating the cube for following schema  I get
>>error at last step while saving and  I am unable to add any measures ,
>>clicking on measure option just dont pop up any dialog,
>>
>>I am using star schema with fact_sales as fact table and dim_* as
>>dimension tables.
>>
>> fact_sales:
>> storeid                 int
>> itemid                  int
>> custid                  int
>> qty                     int
>> price                   double
>>
>>dim_customer
>> custid                  int
>> name                    string
>>
>> dim_item
>> itemid                  int
>> category                string
>> brand                   string
>> color                   string
>>
>>dim_store
>> storeid                 int
>> city                    string
>> state                   string
>>
>>JSON is as below.
>> The JSON is as below.
>>
>> {
>>   "name": "Retail_Cube",
>>   "description": "",
>>   "dimensions": [
>>     {
>>       "name": "RETAIL.FACT_SALES.STOREID",
>>       "table": "RETAIL.FACT_SALES",
>>       "hierarchy": false,
>>       "derived": null,
>>       "column": [
>>         "STOREID"
>>       ],
>>       "id": 1
>>     },
>>     {
>>       "name": "RETAIL.FACT_SALES.ITEMID",
>>       "table": "RETAIL.FACT_SALES",
>>       "hierarchy": false,
>>       "derived": null,
>>       "column": [
>>         "ITEMID"
>>       ],
>>       "id": 2
>>     },
>>     {
>>       "name": "RETAIL.FACT_SALES.CUSTID",
>>       "table": "RETAIL.FACT_SALES",
>>       "hierarchy": false,
>>       "derived": null,
>>       "column": [
>>         "CUSTID"
>>       ],
>>       "id": 3
>>     }
>>   ],
>>   "measures": [
>>     {
>>       "id": 1,
>>       "name": "_COUNT_",
>>       "function": {
>>         "expression": "COUNT",
>>         "returntype": "bigint",
>>         "parameter": {
>>           "type": "constant",
>>           "value": 1
>>         }
>>       }
>>     }
>>   ],
>>   "rowkey": {
>>     "rowkey_columns": [
>>       {
>>         "column": "STOREID",
>>         "length": 0,
>>         "dictionary": "true",
>>         "mandatory": false
>>       },
>>       {
>>         "column": "ITEMID",
>>         "length": 0,
>>         "dictionary": "true",
>>         "mandatory": false
>>       },
>>       {
>>         "column": "CUSTID",
>>         "length": 0,
>>         "dictionary": "true",
>>         "mandatory": false
>>       }
>>     ],
>>     "aggregation_groups": [
>>       [
>>         "STOREID",
>>         "ITEMID",
>>         "CUSTID"
>>       ]
>>     ]
>>   },
>>   "notify_list": [],
>>   "capacity": "",
>>   "hbase_mapping": {
>>     "column_family": [
>>       {
>>         "name": "f1",
>>         "columns": [
>>           {
>>             "qualifier": "m",
>>             "measure_refs": [
>>               "_COUNT_"
>>             ]
>>           }
>>         ]
>>       }
>>     ]
>>   },
>>   "project": "RetailProject",
>>   "model_name": "Retail_Cube"
>> }
>>
>>Regards,
>>Santosh Akhilesh
>>Bangalore R&D
>>HUAWEI TECHNOLOGIES CO.,LTD.
>>
>>www.huawei.com
>>-------------------------------------------------------------------------
>>-
>>-----------------------------------------------------------
>>This e-mail and its attachments contain confidential information from
>>HUAWEI, which
>>is intended only for the person or entity whose address is listed above.
>>Any use of the
>>information contained herein in any way (including, but not limited to,
>>total or partial
>>disclosure, reproduction, or dissemination) by persons other than the
>>intended
>>recipient(s) is prohibited. If you receive this e-mail in error, please
>>notify the sender by
>>phone or email immediately and delete it!
>>
>>________________________________________
>>From: Shi, Shaofeng [[email protected]]
>>Sent: Thursday, February 26, 2015 7:01 AM
>>To: [email protected]
>>Subject: Re: Error while making cube & Measure option is not responding
>>on GUI
>>
>>Hi Santosh,
>>
>>It looks like hadoop failed to execute some shell command in the
>>container; You need dive into hadoop to see what¹s the concrete error.
>>You
>>can use yarn logs command to fetch all logs:
>>
>>yarn logs -applicationId <app_id>
>>
>>
>>On 2/25/15, 7:39 PM, "Santosh Akhilesh" <[email protected]>
>>wrote:
>>
>>>Hi Luke / Shaofeng ,
>>>           Can you please help me to check this issue.
>>>Regards,
>>>Santosh Akhilesh
>>>
>>>On Tue, Feb 24, 2015 at 10:41 PM, Santosh Akhilesh <
>>>[email protected]> wrote:
>>>
>>>> Hi All ,
>>>>         is it because of following error in map reduce job ? what
>>>>could
>>>>be
>>>> way to resolve this , a google search says that its issue of Yarn
>>>>class
>>>> path , but I am not sure what it is ?
>>>>
>>>> Kylin Hive Column Cardinality Job table=RETAIL.FACT_SALES
>>>> output=/tmp/cardinality/RETAIL.FACT_SALES
>>>>
>>>> Application application_1424791969399_0008 failed 2 times due to AM
>>>> Container for appattempt_1424791969399_0008_000002 exited with
>>>>exitCode: 1
>>>> For more detailed output, check application tracking page:
>>>> http://santosh:8088/proxy/application_1424791969399_0008/Then, click
>>>>on
>>>> links to logs of each attempt.
>>>> Diagnostics: Exception from container-launch.
>>>> Container id: container_1424791969399_0008_02_000001
>>>> Exit code: 1
>>>> Stack trace: ExitCodeException exitCode=1:
>>>> at org.apache.hadoop.util.Shell.runCommand(Shell.java:538)
>>>> at org.apache.hadoop.util.Shell.run(Shell.java:455)
>>>> at
>>>>
>>>>org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:71
>>>>5
>>>>)
>>>> at
>>>>
>>>>org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.laun
>>>>c
>>>>h
>>>>Container(DefaultContainerExecutor.java:211)
>>>> at
>>>>
>>>>org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.Con
>>>>t
>>>>a
>>>>inerLaunch.call(ContainerLaunch.java:302)
>>>> at
>>>>
>>>>org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.Con
>>>>t
>>>>a
>>>>inerLaunch.call(ContainerLaunch.java:82)
>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>>>> at
>>>>
>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.ja
>>>>v
>>>>a
>>>>:1145)
>>>> at
>>>>
>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.j
>>>>a
>>>>v
>>>>a:615)
>>>> at java.lang.Thread.run(Thread.java:745)
>>>> Container exited with a non-zero exit code 1
>>>> Failing this attempt. Failing the application.
>>>>
>>>> ---------- Forwarded message ----------
>>>> From: Santoshakhilesh <[email protected]>
>>>> Date: Tue, Feb 24, 2015 at 7:41 PM
>>>> Subject: FW: Error while making cube & Measure option is not
>>>>responding
>>>>on
>>>> GUI
>>>> To: "[email protected]" <[email protected]>
>>>>
>>>>
>>>> hi ,
>>>>    please someone give me a hand to resolve this issue , thanks.
>>>>
>>>> Regards,
>>>> Santosh Akhilesh
>>>> Bangalore R&D
>>>> HUAWEI TECHNOLOGIES CO.,LTD.
>>>>
>>>> www.huawei.com
>>>>
>>>>
>>>>-----------------------------------------------------------------------
>>>>-
>>>>-
>>>>------------------------------------------------------------
>>>> This e-mail and its attachments contain confidential information from
>>>> HUAWEI, which
>>>> is intended only for the person or entity whose address is listed
>>>>above.
>>>> Any use of the
>>>> information contained herein in any way (including, but not limited
>>>>to,
>>>> total or partial
>>>> disclosure, reproduction, or dissemination) by persons other than the
>>>> intended
>>>> recipient(s) is prohibited. If you receive this e-mail in error,
>>>>please
>>>> notify the sender by
>>>> phone or email immediately and delete it!
>>>>
>>>> ________________________________________
>>>> From: Santoshakhilesh [[email protected]]
>>>> Sent: Tuesday, February 24, 2015 12:55 PM
>>>> To: [email protected]
>>>> Cc: Kulbhushan Rana
>>>> Subject: FW: Error while making cube & Measure option is not
>>>>responding
>>>>on
>>>> GUI
>>>>
>>>> 2. If  I ignore and continue and try to save the cube I get an
>>>>exception
>>>> in Kylin.log , I have checked the path is set correctly and
>>>>HCatInputFormat
>>>> this file is present in hive-hcatalog-core-0.14.0.jar . Please let me
>>>>know
>>>> what can I do to resolve this ?
>>>>
>>>>  -- This was path issue , now no more exception in kylin.log
>>>>
>>>> But saveing cube still fails with error. And still can't add measures.
>>>>
>>>> Error Message
>>>> Failed to take action.
>>>>
>>>> In log I can find no exception. Following is the last log in kylin.log
>>>>
>>>> [pool-3-thread-1]:[2015-02-24
>>>>
>>>>20:47:15,613][INFO][org.apache.kylin.job.impl.threadpool.DefaultSchedul
>>>>e
>>>>r
>>>>$FetcherRunner.run(DefaultScheduler.java:117)]
>>>> - Job Fetcher: 0 running, 0 actual running, 0 ready, 6 others
>>>> [http-bio-7070-exec-2]:[2015-02-24
>>>>
>>>>20:47:51,610][DEBUG][org.apache.kylin.rest.controller.CubeController.de
>>>>s
>>>>e
>>>>rializeDataModelDesc(CubeController.java:459)]
>>>> - Saving cube {
>>>>   "name": "",
>>>>   "fact_table": "RETAIL.FACT_SALES",
>>>>   "lookups": [],
>>>>   "filter_condition": "",
>>>>   "capacity": "SMALL",
>>>>   "partition_desc": {
>>>>     "partition_date_column": "",
>>>>     "partition_date_start": 0,
>>>>     "partition_type": "APPEND"
>>>>   },
>>>>   "last_modified": 0
>>>> }
>>>>
>>>>
>>>> local access logs all with 200 , so seems ok.
>>>>
>>>> 10.18.146.105 - - [24/Feb/2015:20:46:56 +0800] "GET
>>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>> 10.18.146.105 - - [24/Feb/2015:20:47:07 +0800] "GET
>>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>> 10.18.146.105 - - [24/Feb/2015:20:47:27 +0800] "GET
>>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>> 10.18.146.105 - - [24/Feb/2015:20:47:28 +0800] "GET
>>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>> 10.18.146.105 - - [24/Feb/2015:20:47:34 +0800] "GET
>>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>> 10.18.146.105 - - [24/Feb/2015:20:47:48 +0800] "GET
>>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>> 10.18.146.105 - - [24/Feb/2015:20:47:51 +0800] "POST /kylin/api/cubes
>>>> HTTP/1.1" 200 701
>>>>
>>>>
>>>> Regards,
>>>> Santosh Akhilesh
>>>> Bangalore R&D
>>>> HUAWEI TECHNOLOGIES CO.,LTD.
>>>>
>>>> www.huawei.com
>>>>
>>>>
>>>>-----------------------------------------------------------------------
>>>>-
>>>>-
>>>>------------------------------------------------------------
>>>> This e-mail and its attachments contain confidential information from
>>>> HUAWEI, which
>>>> is intended only for the person or entity whose address is listed
>>>>above.
>>>> Any use of the
>>>> information contained herein in any way (including, but not limited
>>>>to,
>>>> total or partial
>>>> disclosure, reproduction, or dissemination) by persons other than the
>>>> intended
>>>> recipient(s) is prohibited. If you receive this e-mail in error,
>>>>please
>>>> notify the sender by
>>>> phone or email immediately and delete it!
>>>>
>>>> ________________________________________
>>>> From: Santoshakhilesh [[email protected]]
>>>> Sent: Tuesday, February 24, 2015 12:09 PM
>>>> To: [email protected]
>>>> Cc: Kulbhushan Rana
>>>> Subject: Error while making cube & Measure option is not responding on
>>>>GUI
>>>>
>>>> Hi All ,
>>>>
>>>>     I am building a simple cube for test and using the binary build
>>>>0.7.1
>>>> . I have following hive tables with columns.
>>>>
>>>>
>>>>
>>>> fact_sales:
>>>>
>>>> storeid                 int
>>>> itemid                  int
>>>> custid                  int
>>>> qty                     int
>>>> price                   double
>>>>
>>>> dim_customer
>>>> custid                  int
>>>> name                    string
>>>>
>>>> dim_item
>>>>
>>>> itemid                  int
>>>> category                string
>>>> brand                   string
>>>> color                   string
>>>>
>>>> dim_store
>>>>
>>>> storeid                 int
>>>> city                    string
>>>> state                   string
>>>>
>>>> Please help me to answer following issues;
>>>>
>>>>
>>>>
>>>> 1. When I go to measure section and click on measure option , there is
>>>>no
>>>> response , I want add measure on qty and price with sum
>>>>
>>>> 2. If  I ignore and continue and try to save the cube I get an
>>>>exception
>>>> in Kylin.log , I have checked the path is set correctly and
>>>>HCatInputFormat
>>>> this file is present in hive-hcatalog-core-0.14.0.jar . Please let me
>>>>know
>>>> what can I do to resolve this ?
>>>>
>>>> 3. Also I have another question since this is a test and data is small
>>>>I
>>>> have not partitioned the fact table , is it ok to skip partition stage
>>>> while cube build ?
>>>>
>>>>
>>>>
>>>> Exception
>>>>
>>>> pool-4-thread-4]:[2015-02-24
>>>>
>>>>19:26:32,577][ERROR][org.apache.kylin.job.impl.threadpool.DefaultSchedu
>>>>l
>>>>e
>>>>r$JobRunner.run(DefaultScheduler.java:134)]
>>>> - ExecuteException job:c3532a6f-97ea-474a-b36a-218dd517cedb
>>>> org.apache.kylin.job.exception.ExecuteException:
>>>> org.apache.kylin.job.exception.ExecuteException:
>>>> java.lang.NoClassDefFoundError:
>>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat
>>>>  at
>>>>
>>>>org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecu
>>>>t
>>>>a
>>>>ble.java:102)
>>>>  at
>>>>
>>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(Def
>>>>a
>>>>u
>>>>ltScheduler.java:132)
>>>>  at
>>>>
>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.ja
>>>>v
>>>>a
>>>>:1145)
>>>>  at
>>>>
>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.j
>>>>a
>>>>v
>>>>a:615)
>>>>  at java.lang.Thread.run(Thread.java:745)
>>>> Caused by: org.apache.kylin.job.exception.ExecuteException:
>>>> java.lang.NoClassDefFoundError:
>>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat
>>>>
>>>>
>>>>
>>>> The JSON is as below.
>>>>
>>>>
>>>>
>>>> {
>>>>   "name": "Retail_Cube",
>>>>   "description": "",
>>>>   "dimensions": [
>>>>     {
>>>>       "name": "RETAIL.FACT_SALES.STOREID",
>>>>       "table": "RETAIL.FACT_SALES",
>>>>       "hierarchy": false,
>>>>       "derived": null,
>>>>       "column": [
>>>>         "STOREID"
>>>>       ],
>>>>       "id": 1
>>>>     },
>>>>     {
>>>>       "name": "RETAIL.FACT_SALES.ITEMID",
>>>>       "table": "RETAIL.FACT_SALES",
>>>>       "hierarchy": false,
>>>>       "derived": null,
>>>>       "column": [
>>>>         "ITEMID"
>>>>       ],
>>>>       "id": 2
>>>>     },
>>>>     {
>>>>       "name": "RETAIL.FACT_SALES.CUSTID",
>>>>       "table": "RETAIL.FACT_SALES",
>>>>       "hierarchy": false,
>>>>       "derived": null,
>>>>       "column": [
>>>>         "CUSTID"
>>>>       ],
>>>>       "id": 3
>>>>     }
>>>>   ],
>>>>   "measures": [
>>>>     {
>>>>       "id": 1,
>>>>       "name": "_COUNT_",
>>>>       "function": {
>>>>         "expression": "COUNT",
>>>>         "returntype": "bigint",
>>>>         "parameter": {
>>>>           "type": "constant",
>>>>           "value": 1
>>>>         }
>>>>       }
>>>>     }
>>>>   ],
>>>>   "rowkey": {
>>>>     "rowkey_columns": [
>>>>       {
>>>>         "column": "STOREID",
>>>>         "length": 0,
>>>>         "dictionary": "true",
>>>>         "mandatory": false
>>>>       },
>>>>       {
>>>>         "column": "ITEMID",
>>>>         "length": 0,
>>>>         "dictionary": "true",
>>>>         "mandatory": false
>>>>       },
>>>>       {
>>>>         "column": "CUSTID",
>>>>         "length": 0,
>>>>         "dictionary": "true",
>>>>         "mandatory": false
>>>>       }
>>>>     ],
>>>>     "aggregation_groups": [
>>>>       [
>>>>         "STOREID",
>>>>         "ITEMID",
>>>>         "CUSTID"
>>>>       ]
>>>>     ]
>>>>   },
>>>>   "notify_list": [],
>>>>   "capacity": "",
>>>>   "hbase_mapping": {
>>>>     "column_family": [
>>>>       {
>>>>         "name": "f1",
>>>>         "columns": [
>>>>           {
>>>>             "qualifier": "m",
>>>>             "measure_refs": [
>>>>               "_COUNT_"
>>>>             ]
>>>>           }
>>>>         ]
>>>>       }
>>>>     ]
>>>>   },
>>>>   "project": "RetailProject",
>>>>   "model_name": "Retail_Cube"
>>>> }
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> Regards,
>>>> Santosh Akhilesh
>>>> Bangalore R&D
>>>> HUAWEI TECHNOLOGIES CO.,LTD.
>>>>
>>>> www.huawei.com
>>>>
>>>>
>>>>-----------------------------------------------------------------------
>>>>-
>>>>-
>>>>------------------------------------------------------------
>>>> This e-mail and its attachments contain confidential information from
>>>> HUAWEI, which
>>>> is intended only for the person or entity whose address is listed
>>>>above.
>>>> Any use of the
>>>> information contained herein in any way (including, but not limited
>>>>to,
>>>> total or partial
>>>> disclosure, reproduction, or dissemination) by persons other than the
>>>> intended
>>>> recipient(s) is prohibited. If you receive this e-mail in error,
>>>>please
>>>> notify the sender by
>>>> phone or email immediately and delete it!
>>>>
>>>>
>>>>
>>>> --
>>>> Regards,
>>>> Santosh Akhilesh
>>>> +91-0-9845482201
>>>>
>>>
>>>
>>>
>>>--
>>>Regards,
>>>Santosh Akhilesh
>>>+91-0-9845482201

Reply via email to