Which version or code branch are you using? I assume you’re using the
stable version from master; Seems you’re trying to edit an existing cube
to add new measurement, try refresh your browser's cache; if it still
couldn’t be saved, try to create a new cube;

The two error traces in tomcat need be taken care:

1) java.lang.NoClassDefFoundError: org/apache/kylin/common/mr/KylinMapper
        Please check kylin.properties file, making sure the “kylin.job.jar”
points to a right jar file; It will be loaded in Map-reduce;

2) java.lang.IllegalArgumentException: No enum constant
org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS
        This indicates your hadoop version might be old; Please check and ensure
hadoop version is 2.2 or above.

On 2/26/15, 3:21 PM, "Santoshakhilesh" <[email protected]> wrote:

>Hi Shaofeng ,
>
>   I am using chrome , When I click on button to add measures , following
>is error on chrome console. When I try to save the cube there is no error
>in console. I just get a error dialog saying failed to take action and
>gives me the JSON cube schema.
>
>Error on chrome debug console is as below;
>
> ReferenceError: CubeDescModel is not defined
>    at h.$scope.addNewMeasure (scripts.min.0.js:15984)
>    at scripts.min.0.js:180
>    at scripts.min.0.js:197
>    at h.$eval (scripts.min.0.js:119)
>    at h.$apply (scripts.min.0.js:119)
>    at HTMLButtonElement.<anonymous> (scripts.min.0.js:197)
>    at HTMLButtonElement.m.event.dispatch (scripts.min.0.js:3)
>    at HTMLButtonElement.r.handle
>(scripts.min.0.js:3)scripts.min.0.js:100 (anonymous
>function)scripts.min.0.js:77 (anonymous function)scripts.min.0.js:119
>h.$applyscripts.min.0.js:197 (anonymous function)scripts.min.0.js:3
>m.event.dispatchscripts.min.0.js:3 r.handle
>
>   About the hive table import , I got pass the run shell command
>exception but it still fails the hadoop log is;
>2015-02-26 20:46:48,332 INFO [main] org.apache.hadoop.mapred.YarnChild:
>mapreduce.cluster.local.dir for child:
>/tmp/hadoop-root/nm-local-dir/usercache/root/appcache/application_14249530
>91340_0002
>2015-02-26 20:46:48,776 INFO [main]
>org.apache.hadoop.conf.Configuration.deprecation: session.id is
>deprecated. Instead, use dfs.metrics.session-id
>2015-02-26 20:46:49,310 INFO [main] org.apache.hadoop.mapred.Task:  Using
>ResourceCalculatorProcessTree : [ ]
>2015-02-26 20:46:49,386 FATAL [main] org.apache.hadoop.mapred.YarnChild:
>Error running child : java.lang.NoClassDefFoundError:
>org/apache/kylin/common/mr/KylinMapper
> at java.lang.ClassLoader.defineClass1(Native Method)
> at java.lang.ClassLoader.defineClass(ClassLoader.java:800)
> at 
>java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
> at java.net.URLClassLoader.defineClass(URLClassLoader.java:449)
> at java.net.URLClassLoader.access$100(URLClassLoader.java:71)
> at java.net.URLClassLoader$1.run(URLClassLoader.java:361)
> at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
> at java.security.AccessController.doPrivileged(Native Method)
> at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
>
>tomcat logs:
>usage: HiveColumnCardinalityJob
> -output <path>        Output path
> -table <table name>   The hive table name
>[pool-4-thread-2]:[2015-02-26
>20:47:49,936][ERROR][org.apache.kylin.job.common.HadoopShellExecutable.doW
>ork(HadoopShellExecutable.java:64)] - error execute
>HadoopShellExecutable{id=d4730d26-7fe6-412e-9841-3288ab362c5b-00,
>name=null, state=RUNNING}
>java.lang.IllegalArgumentException: No enum constant
>org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS
> at java.lang.Enum.valueOf(Enum.java:236)
> at 
>org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.valueOf(Framewo
>rkCounterGroup.java:148)
> at 
>org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.findCounter(Fra
>meworkCounterGroup.java:182)
> at 
>org.apache.hadoop.mapreduce.counters.AbstractCounters.findCounter(Abstract
>Counters.java:154)
> at 
>org.apache.hadoop.mapreduce.TypeConverter.fromYarn(TypeConverter.java:240)
> at 
>org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounters(ClientServic
>eDelegate.java:370)
> at 
>org.apache.hadoop.mapred.YARNRunner.getJobCounters(YARNRunner.java:511)
> at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756)
> at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:415)
> at 
>org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.
>java:1491)
> at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753)
> at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1361)
> at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1289)
> at 
>org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForCompletion(AbstractHa
>doopJob.java:134)
> at 
>org.apache.kylin.job.hadoop.cardinality.HiveColumnCardinalityJob.run(HiveC
>olumnCardinalityJob.java:114)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
> at 
>org.apache.kylin.job.common.HadoopShellExecutable.doWork(HadoopShellExecut
>able.java:62)
> at 
>org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutab
>le.java:99)
> at 
>org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChai
>nedExecutable.java:50)
> at 
>org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutab
>le.java:99)
> at 
>org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(Defaul
>tScheduler.java:132)
> at 
>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:
>1145)
> at 
>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java
>:615)
> at java.lang.Thread.run(Thread.java:745)
>
>Regards,
>Santosh Akhilesh
>Bangalore R&D
>HUAWEI TECHNOLOGIES CO.,LTD.
>
>www.huawei.com
>--------------------------------------------------------------------------
>-----------------------------------------------------------
>This e-mail and its attachments contain confidential information from
>HUAWEI, which
>is intended only for the person or entity whose address is listed above.
>Any use of the
>information contained herein in any way (including, but not limited to,
>total or partial
>disclosure, reproduction, or dissemination) by persons other than the
>intended
>recipient(s) is prohibited. If you receive this e-mail in error, please
>notify the sender by
>phone or email immediately and delete it!
>
>________________________________________
>From: Shi, Shaofeng [[email protected]]
>Sent: Thursday, February 26, 2015 11:32 AM
>To: [email protected]
>Cc: Kulbhushan Rana
>Subject: Re: Error while making cube & Measure option is not responding
>on GUI
>
>Hi Santosh, hive table importing issue should not impact on cube saving.
>
>If you couldn’t save the cube, firstly please check whether there is error
>in the tomcat’s log; If not please check your web browser; We suggest use
>Firefox (with firebug add-on) or Chrome, open the JS console (press F12)
>and then operate web UI, check whether there is any error reported in
>browser.
>
>
>On 2/26/15, 1:08 PM, "Santoshakhilesh" <[email protected]>
>wrote:
>
>>Hi Shaofeng ,
>>   Thanks for replying.
>>   Yes I am checking the yarn exception, But I find that this error comes
>>while importing the hive table to kylin.
>>   Even if this error comes , hive tables is exported successfully in
>>kylin. Is this the reason why cube saving has failed ?
>>
>>   Next step when I go on creating the cube for following schema  I get
>>error at last step while saving and  I am unable to add any measures ,
>>clicking on measure option just dont pop up any dialog,
>>
>>I am using star schema with fact_sales as fact table and dim_* as
>>dimension tables.
>>
>> fact_sales:
>> storeid                 int
>> itemid                  int
>> custid                  int
>> qty                     int
>> price                   double
>>
>>dim_customer
>> custid                  int
>> name                    string
>>
>> dim_item
>> itemid                  int
>> category                string
>> brand                   string
>> color                   string
>>
>>dim_store
>> storeid                 int
>> city                    string
>> state                   string
>>
>>JSON is as below.
>> The JSON is as below.
>>
>> {
>>   "name": "Retail_Cube",
>>   "description": "",
>>   "dimensions": [
>>     {
>>       "name": "RETAIL.FACT_SALES.STOREID",
>>       "table": "RETAIL.FACT_SALES",
>>       "hierarchy": false,
>>       "derived": null,
>>       "column": [
>>         "STOREID"
>>       ],
>>       "id": 1
>>     },
>>     {
>>       "name": "RETAIL.FACT_SALES.ITEMID",
>>       "table": "RETAIL.FACT_SALES",
>>       "hierarchy": false,
>>       "derived": null,
>>       "column": [
>>         "ITEMID"
>>       ],
>>       "id": 2
>>     },
>>     {
>>       "name": "RETAIL.FACT_SALES.CUSTID",
>>       "table": "RETAIL.FACT_SALES",
>>       "hierarchy": false,
>>       "derived": null,
>>       "column": [
>>         "CUSTID"
>>       ],
>>       "id": 3
>>     }
>>   ],
>>   "measures": [
>>     {
>>       "id": 1,
>>       "name": "_COUNT_",
>>       "function": {
>>         "expression": "COUNT",
>>         "returntype": "bigint",
>>         "parameter": {
>>           "type": "constant",
>>           "value": 1
>>         }
>>       }
>>     }
>>   ],
>>   "rowkey": {
>>     "rowkey_columns": [
>>       {
>>         "column": "STOREID",
>>         "length": 0,
>>         "dictionary": "true",
>>         "mandatory": false
>>       },
>>       {
>>         "column": "ITEMID",
>>         "length": 0,
>>         "dictionary": "true",
>>         "mandatory": false
>>       },
>>       {
>>         "column": "CUSTID",
>>         "length": 0,
>>         "dictionary": "true",
>>         "mandatory": false
>>       }
>>     ],
>>     "aggregation_groups": [
>>       [
>>         "STOREID",
>>         "ITEMID",
>>         "CUSTID"
>>       ]
>>     ]
>>   },
>>   "notify_list": [],
>>   "capacity": "",
>>   "hbase_mapping": {
>>     "column_family": [
>>       {
>>         "name": "f1",
>>         "columns": [
>>           {
>>             "qualifier": "m",
>>             "measure_refs": [
>>               "_COUNT_"
>>             ]
>>           }
>>         ]
>>       }
>>     ]
>>   },
>>   "project": "RetailProject",
>>   "model_name": "Retail_Cube"
>> }
>>
>>Regards,
>>Santosh Akhilesh
>>Bangalore R&D
>>HUAWEI TECHNOLOGIES CO.,LTD.
>>
>>www.huawei.com
>>-------------------------------------------------------------------------
>>-
>>-----------------------------------------------------------
>>This e-mail and its attachments contain confidential information from
>>HUAWEI, which
>>is intended only for the person or entity whose address is listed above.
>>Any use of the
>>information contained herein in any way (including, but not limited to,
>>total or partial
>>disclosure, reproduction, or dissemination) by persons other than the
>>intended
>>recipient(s) is prohibited. If you receive this e-mail in error, please
>>notify the sender by
>>phone or email immediately and delete it!
>>
>>________________________________________
>>From: Shi, Shaofeng [[email protected]]
>>Sent: Thursday, February 26, 2015 7:01 AM
>>To: [email protected]
>>Subject: Re: Error while making cube & Measure option is not responding
>>on GUI
>>
>>Hi Santosh,
>>
>>It looks like hadoop failed to execute some shell command in the
>>container; You need dive into hadoop to see what¹s the concrete error.
>>You
>>can use yarn logs command to fetch all logs:
>>
>>yarn logs -applicationId <app_id>
>>
>>
>>On 2/25/15, 7:39 PM, "Santosh Akhilesh" <[email protected]>
>>wrote:
>>
>>>Hi Luke / Shaofeng ,
>>>           Can you please help me to check this issue.
>>>Regards,
>>>Santosh Akhilesh
>>>
>>>On Tue, Feb 24, 2015 at 10:41 PM, Santosh Akhilesh <
>>>[email protected]> wrote:
>>>
>>>> Hi All ,
>>>>         is it because of following error in map reduce job ? what
>>>>could
>>>>be
>>>> way to resolve this , a google search says that its issue of Yarn
>>>>class
>>>> path , but I am not sure what it is ?
>>>>
>>>> Kylin Hive Column Cardinality Job table=RETAIL.FACT_SALES
>>>> output=/tmp/cardinality/RETAIL.FACT_SALES
>>>>
>>>> Application application_1424791969399_0008 failed 2 times due to AM
>>>> Container for appattempt_1424791969399_0008_000002 exited with
>>>>exitCode: 1
>>>> For more detailed output, check application tracking page:
>>>> http://santosh:8088/proxy/application_1424791969399_0008/Then, click
>>>>on
>>>> links to logs of each attempt.
>>>> Diagnostics: Exception from container-launch.
>>>> Container id: container_1424791969399_0008_02_000001
>>>> Exit code: 1
>>>> Stack trace: ExitCodeException exitCode=1:
>>>> at org.apache.hadoop.util.Shell.runCommand(Shell.java:538)
>>>> at org.apache.hadoop.util.Shell.run(Shell.java:455)
>>>> at
>>>>
>>>>org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:71
>>>>5
>>>>)
>>>> at
>>>>
>>>>org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.laun
>>>>c
>>>>h
>>>>Container(DefaultContainerExecutor.java:211)
>>>> at
>>>>
>>>>org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.Con
>>>>t
>>>>a
>>>>inerLaunch.call(ContainerLaunch.java:302)
>>>> at
>>>>
>>>>org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.Con
>>>>t
>>>>a
>>>>inerLaunch.call(ContainerLaunch.java:82)
>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>>>> at
>>>>
>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.ja
>>>>v
>>>>a
>>>>:1145)
>>>> at
>>>>
>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.j
>>>>a
>>>>v
>>>>a:615)
>>>> at java.lang.Thread.run(Thread.java:745)
>>>> Container exited with a non-zero exit code 1
>>>> Failing this attempt. Failing the application.
>>>>
>>>> ---------- Forwarded message ----------
>>>> From: Santoshakhilesh <[email protected]>
>>>> Date: Tue, Feb 24, 2015 at 7:41 PM
>>>> Subject: FW: Error while making cube & Measure option is not
>>>>responding
>>>>on
>>>> GUI
>>>> To: "[email protected]" <[email protected]>
>>>>
>>>>
>>>> hi ,
>>>>    please someone give me a hand to resolve this issue , thanks.
>>>>
>>>> Regards,
>>>> Santosh Akhilesh
>>>> Bangalore R&D
>>>> HUAWEI TECHNOLOGIES CO.,LTD.
>>>>
>>>> www.huawei.com
>>>>
>>>>
>>>>-----------------------------------------------------------------------
>>>>-
>>>>-
>>>>------------------------------------------------------------
>>>> This e-mail and its attachments contain confidential information from
>>>> HUAWEI, which
>>>> is intended only for the person or entity whose address is listed
>>>>above.
>>>> Any use of the
>>>> information contained herein in any way (including, but not limited
>>>>to,
>>>> total or partial
>>>> disclosure, reproduction, or dissemination) by persons other than the
>>>> intended
>>>> recipient(s) is prohibited. If you receive this e-mail in error,
>>>>please
>>>> notify the sender by
>>>> phone or email immediately and delete it!
>>>>
>>>> ________________________________________
>>>> From: Santoshakhilesh [[email protected]]
>>>> Sent: Tuesday, February 24, 2015 12:55 PM
>>>> To: [email protected]
>>>> Cc: Kulbhushan Rana
>>>> Subject: FW: Error while making cube & Measure option is not
>>>>responding
>>>>on
>>>> GUI
>>>>
>>>> 2. If  I ignore and continue and try to save the cube I get an
>>>>exception
>>>> in Kylin.log , I have checked the path is set correctly and
>>>>HCatInputFormat
>>>> this file is present in hive-hcatalog-core-0.14.0.jar . Please let me
>>>>know
>>>> what can I do to resolve this ?
>>>>
>>>>  -- This was path issue , now no more exception in kylin.log
>>>>
>>>> But saveing cube still fails with error. And still can't add measures.
>>>>
>>>> Error Message
>>>> Failed to take action.
>>>>
>>>> In log I can find no exception. Following is the last log in kylin.log
>>>>
>>>> [pool-3-thread-1]:[2015-02-24
>>>>
>>>>20:47:15,613][INFO][org.apache.kylin.job.impl.threadpool.DefaultSchedul
>>>>e
>>>>r
>>>>$FetcherRunner.run(DefaultScheduler.java:117)]
>>>> - Job Fetcher: 0 running, 0 actual running, 0 ready, 6 others
>>>> [http-bio-7070-exec-2]:[2015-02-24
>>>>
>>>>20:47:51,610][DEBUG][org.apache.kylin.rest.controller.CubeController.de
>>>>s
>>>>e
>>>>rializeDataModelDesc(CubeController.java:459)]
>>>> - Saving cube {
>>>>   "name": "",
>>>>   "fact_table": "RETAIL.FACT_SALES",
>>>>   "lookups": [],
>>>>   "filter_condition": "",
>>>>   "capacity": "SMALL",
>>>>   "partition_desc": {
>>>>     "partition_date_column": "",
>>>>     "partition_date_start": 0,
>>>>     "partition_type": "APPEND"
>>>>   },
>>>>   "last_modified": 0
>>>> }
>>>>
>>>>
>>>> local access logs all with 200 , so seems ok.
>>>>
>>>> 10.18.146.105 - - [24/Feb/2015:20:46:56 +0800] "GET
>>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>> 10.18.146.105 - - [24/Feb/2015:20:47:07 +0800] "GET
>>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>> 10.18.146.105 - - [24/Feb/2015:20:47:27 +0800] "GET
>>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>> 10.18.146.105 - - [24/Feb/2015:20:47:28 +0800] "GET
>>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>> 10.18.146.105 - - [24/Feb/2015:20:47:34 +0800] "GET
>>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>> 10.18.146.105 - - [24/Feb/2015:20:47:48 +0800] "GET
>>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>>> 10.18.146.105 - - [24/Feb/2015:20:47:51 +0800] "POST /kylin/api/cubes
>>>> HTTP/1.1" 200 701
>>>>
>>>>
>>>> Regards,
>>>> Santosh Akhilesh
>>>> Bangalore R&D
>>>> HUAWEI TECHNOLOGIES CO.,LTD.
>>>>
>>>> www.huawei.com
>>>>
>>>>
>>>>-----------------------------------------------------------------------
>>>>-
>>>>-
>>>>------------------------------------------------------------
>>>> This e-mail and its attachments contain confidential information from
>>>> HUAWEI, which
>>>> is intended only for the person or entity whose address is listed
>>>>above.
>>>> Any use of the
>>>> information contained herein in any way (including, but not limited
>>>>to,
>>>> total or partial
>>>> disclosure, reproduction, or dissemination) by persons other than the
>>>> intended
>>>> recipient(s) is prohibited. If you receive this e-mail in error,
>>>>please
>>>> notify the sender by
>>>> phone or email immediately and delete it!
>>>>
>>>> ________________________________________
>>>> From: Santoshakhilesh [[email protected]]
>>>> Sent: Tuesday, February 24, 2015 12:09 PM
>>>> To: [email protected]
>>>> Cc: Kulbhushan Rana
>>>> Subject: Error while making cube & Measure option is not responding on
>>>>GUI
>>>>
>>>> Hi All ,
>>>>
>>>>     I am building a simple cube for test and using the binary build
>>>>0.7.1
>>>> . I have following hive tables with columns.
>>>>
>>>>
>>>>
>>>> fact_sales:
>>>>
>>>> storeid                 int
>>>> itemid                  int
>>>> custid                  int
>>>> qty                     int
>>>> price                   double
>>>>
>>>> dim_customer
>>>> custid                  int
>>>> name                    string
>>>>
>>>> dim_item
>>>>
>>>> itemid                  int
>>>> category                string
>>>> brand                   string
>>>> color                   string
>>>>
>>>> dim_store
>>>>
>>>> storeid                 int
>>>> city                    string
>>>> state                   string
>>>>
>>>> Please help me to answer following issues;
>>>>
>>>>
>>>>
>>>> 1. When I go to measure section and click on measure option , there is
>>>>no
>>>> response , I want add measure on qty and price with sum
>>>>
>>>> 2. If  I ignore and continue and try to save the cube I get an
>>>>exception
>>>> in Kylin.log , I have checked the path is set correctly and
>>>>HCatInputFormat
>>>> this file is present in hive-hcatalog-core-0.14.0.jar . Please let me
>>>>know
>>>> what can I do to resolve this ?
>>>>
>>>> 3. Also I have another question since this is a test and data is small
>>>>I
>>>> have not partitioned the fact table , is it ok to skip partition stage
>>>> while cube build ?
>>>>
>>>>
>>>>
>>>> Exception
>>>>
>>>> pool-4-thread-4]:[2015-02-24
>>>>
>>>>19:26:32,577][ERROR][org.apache.kylin.job.impl.threadpool.DefaultSchedu
>>>>l
>>>>e
>>>>r$JobRunner.run(DefaultScheduler.java:134)]
>>>> - ExecuteException job:c3532a6f-97ea-474a-b36a-218dd517cedb
>>>> org.apache.kylin.job.exception.ExecuteException:
>>>> org.apache.kylin.job.exception.ExecuteException:
>>>> java.lang.NoClassDefFoundError:
>>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat
>>>>  at
>>>>
>>>>org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecu
>>>>t
>>>>a
>>>>ble.java:102)
>>>>  at
>>>>
>>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(Def
>>>>a
>>>>u
>>>>ltScheduler.java:132)
>>>>  at
>>>>
>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.ja
>>>>v
>>>>a
>>>>:1145)
>>>>  at
>>>>
>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.j
>>>>a
>>>>v
>>>>a:615)
>>>>  at java.lang.Thread.run(Thread.java:745)
>>>> Caused by: org.apache.kylin.job.exception.ExecuteException:
>>>> java.lang.NoClassDefFoundError:
>>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat
>>>>
>>>>
>>>>
>>>> The JSON is as below.
>>>>
>>>>
>>>>
>>>> {
>>>>   "name": "Retail_Cube",
>>>>   "description": "",
>>>>   "dimensions": [
>>>>     {
>>>>       "name": "RETAIL.FACT_SALES.STOREID",
>>>>       "table": "RETAIL.FACT_SALES",
>>>>       "hierarchy": false,
>>>>       "derived": null,
>>>>       "column": [
>>>>         "STOREID"
>>>>       ],
>>>>       "id": 1
>>>>     },
>>>>     {
>>>>       "name": "RETAIL.FACT_SALES.ITEMID",
>>>>       "table": "RETAIL.FACT_SALES",
>>>>       "hierarchy": false,
>>>>       "derived": null,
>>>>       "column": [
>>>>         "ITEMID"
>>>>       ],
>>>>       "id": 2
>>>>     },
>>>>     {
>>>>       "name": "RETAIL.FACT_SALES.CUSTID",
>>>>       "table": "RETAIL.FACT_SALES",
>>>>       "hierarchy": false,
>>>>       "derived": null,
>>>>       "column": [
>>>>         "CUSTID"
>>>>       ],
>>>>       "id": 3
>>>>     }
>>>>   ],
>>>>   "measures": [
>>>>     {
>>>>       "id": 1,
>>>>       "name": "_COUNT_",
>>>>       "function": {
>>>>         "expression": "COUNT",
>>>>         "returntype": "bigint",
>>>>         "parameter": {
>>>>           "type": "constant",
>>>>           "value": 1
>>>>         }
>>>>       }
>>>>     }
>>>>   ],
>>>>   "rowkey": {
>>>>     "rowkey_columns": [
>>>>       {
>>>>         "column": "STOREID",
>>>>         "length": 0,
>>>>         "dictionary": "true",
>>>>         "mandatory": false
>>>>       },
>>>>       {
>>>>         "column": "ITEMID",
>>>>         "length": 0,
>>>>         "dictionary": "true",
>>>>         "mandatory": false
>>>>       },
>>>>       {
>>>>         "column": "CUSTID",
>>>>         "length": 0,
>>>>         "dictionary": "true",
>>>>         "mandatory": false
>>>>       }
>>>>     ],
>>>>     "aggregation_groups": [
>>>>       [
>>>>         "STOREID",
>>>>         "ITEMID",
>>>>         "CUSTID"
>>>>       ]
>>>>     ]
>>>>   },
>>>>   "notify_list": [],
>>>>   "capacity": "",
>>>>   "hbase_mapping": {
>>>>     "column_family": [
>>>>       {
>>>>         "name": "f1",
>>>>         "columns": [
>>>>           {
>>>>             "qualifier": "m",
>>>>             "measure_refs": [
>>>>               "_COUNT_"
>>>>             ]
>>>>           }
>>>>         ]
>>>>       }
>>>>     ]
>>>>   },
>>>>   "project": "RetailProject",
>>>>   "model_name": "Retail_Cube"
>>>> }
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> Regards,
>>>> Santosh Akhilesh
>>>> Bangalore R&D
>>>> HUAWEI TECHNOLOGIES CO.,LTD.
>>>>
>>>> www.huawei.com
>>>>
>>>>
>>>>-----------------------------------------------------------------------
>>>>-
>>>>-
>>>>------------------------------------------------------------
>>>> This e-mail and its attachments contain confidential information from
>>>> HUAWEI, which
>>>> is intended only for the person or entity whose address is listed
>>>>above.
>>>> Any use of the
>>>> information contained herein in any way (including, but not limited
>>>>to,
>>>> total or partial
>>>> disclosure, reproduction, or dissemination) by persons other than the
>>>> intended
>>>> recipient(s) is prohibited. If you receive this e-mail in error,
>>>>please
>>>> notify the sender by
>>>> phone or email immediately and delete it!
>>>>
>>>>
>>>>
>>>> --
>>>> Regards,
>>>> Santosh Akhilesh
>>>> +91-0-9845482201
>>>>
>>>
>>>
>>>
>>>--
>>>Regards,
>>>Santosh Akhilesh
>>>+91-0-9845482201

Reply via email to