Hi Santosh, 

0.7.1 hasn’t been formally released; from 0.6.x to 0.7.x we have metadata
structure change; While, the web UI (cube wizard) for this change hasn’t
been stabilized; So it is not strange that you got trouble when saving the
cube; 

@Jason, any idea about the JS error?

On 2/26/15, 5:08 PM, "Santosh Akhilesh" <[email protected]> wrote:

>Hi Shaofeng,
>
>I am using the binary distribution 0.7.1. I have not been able to save
>cube
>even once. I have tried creating new project and from local machine and
>server machine. But I am always stuck with this error. I am never allowed
>to add measures and never been able to save the cube. I also see the
>kylin.log and it always tries to save cube with append mode. One thing I
>need to tell that at partition stage since I don't have a big fact table
>now I have not partititioned the fact table and I skip this step. Does
>this
>have affect in saving the cube. Is this because some metadata is available
>and it tries to modify the cube? I am using latest Hadoop 2.6.6. Yes kylin
>propert I have not added the jar. I will add them and check. But cube
>creation failure is really puzzling me. I could see no error logs in
>kylin.log.
>Regards
>Santosh
>On Thu, 26 Feb 2015 at 1:40 pm, Shi, Shaofeng <[email protected]> wrote:
>
>> Which version or code branch are you using? I assume you’re using the
>> stable version from master; Seems you’re trying to edit an existing cube
>> to add new measurement, try refresh your browser's cache; if it still
>> couldn’t be saved, try to create a new cube;
>>
>> The two error traces in tomcat need be taken care:
>>
>> 1) java.lang.NoClassDefFoundError:
>>org/apache/kylin/common/mr/KylinMapper
>>         Please check kylin.properties file, making sure the
>>“kylin.job.jar”
>> points to a right jar file; It will be loaded in Map-reduce;
>>
>> 2) java.lang.IllegalArgumentException: No enum constant
>> org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS
>>         This indicates your hadoop version might be old; Please check
>>and
>> ensure
>> hadoop version is 2.2 or above.
>>
>> On 2/26/15, 3:21 PM, "Santoshakhilesh" <[email protected]>
>> wrote:
>>
>> >Hi Shaofeng ,
>> >
>> >   I am using chrome , When I click on button to add measures ,
>>following
>> >is error on chrome console. When I try to save the cube there is no
>>error
>> >in console. I just get a error dialog saying failed to take action and
>> >gives me the JSON cube schema.
>> >
>> >Error on chrome debug console is as below;
>> >
>> > ReferenceError: CubeDescModel is not defined
>> >    at h.$scope.addNewMeasure (scripts.min.0.js:15984)
>> >    at scripts.min.0.js:180
>> >    at scripts.min.0.js:197
>> >    at h.$eval (scripts.min.0.js:119)
>> >    at h.$apply (scripts.min.0.js:119)
>> >    at HTMLButtonElement.<anonymous> (scripts.min.0.js:197)
>> >    at HTMLButtonElement.m.event.dispatch (scripts.min.0.js:3)
>> >    at HTMLButtonElement.r.handle
>> >(scripts.min.0.js:3)scripts.min.0.js:100 (anonymous
>> >function)scripts.min.0.js:77 (anonymous function)scripts.min.0.js:119
>> >h.$applyscripts.min.0.js:197 (anonymous function)scripts.min.0.js:3
>> >m.event.dispatchscripts.min.0.js:3 r.handle
>> >
>> >   About the hive table import , I got pass the run shell command
>> >exception but it still fails the hadoop log is;
>> >2015-02-26 20:46:48,332 INFO [main] org.apache.hadoop.mapred.YarnChild:
>> >mapreduce.cluster.local.dir for child:
>> >/tmp/hadoop-root/nm-local-dir/usercache/root/appcache/appli
>> cation_14249530
>> >91340_0002
>> >2015-02-26 20:46:48,776 INFO [main]
>> >org.apache.hadoop.conf.Configuration.deprecation: session.id is
>> >deprecated. Instead, use dfs.metrics.session-id
>> >2015-02-26 20:46:49,310 INFO [main] org.apache.hadoop.mapred.Task:
>>Using
>> >ResourceCalculatorProcessTree : [ ]
>> >2015-02-26 20:46:49,386 FATAL [main]
>>org.apache.hadoop.mapred.YarnChild:
>> >Error running child : java.lang.NoClassDefFoundError:
>> >org/apache/kylin/common/mr/KylinMapper
>> > at java.lang.ClassLoader.defineClass1(Native Method)
>> > at java.lang.ClassLoader.defineClass(ClassLoader.java:800)
>> > at
>> >java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
>> > at java.net.URLClassLoader.defineClass(URLClassLoader.java:449)
>> > at java.net.URLClassLoader.access$100(URLClassLoader.java:71)
>> > at java.net.URLClassLoader$1.run(URLClassLoader.java:361)
>> > at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
>> > at java.security.AccessController.doPrivileged(Native Method)
>> > at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
>> > at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
>> > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
>> >
>> >tomcat logs:
>> >usage: HiveColumnCardinalityJob
>> > -output <path>        Output path
>> > -table <table name>   The hive table name
>> >[pool-4-thread-2]:[2015-02-26
>> >20:47:49,936][ERROR][org.apache.kylin.job.common.HadoopShel
>> lExecutable.doW
>> >ork(HadoopShellExecutable.java:64)] - error execute
>> >HadoopShellExecutable{id=d4730d26-7fe6-412e-9841-3288ab362c5b-00,
>> >name=null, state=RUNNING}
>> >java.lang.IllegalArgumentException: No enum constant
>> >org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS
>> > at java.lang.Enum.valueOf(Enum.java:236)
>> > at
>> >org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.valueOf(
>> Framewo
>> >rkCounterGroup.java:148)
>> > at
>> >org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.
>> findCounter(Fra
>> >meworkCounterGroup.java:182)
>> > at
>> >org.apache.hadoop.mapreduce.counters.AbstractCounters.findC
>> ounter(Abstract
>> >Counters.java:154)
>> > at
>> >org.apache.hadoop.mapreduce.TypeConverter.fromYarn(TypeConv
>> erter.java:240)
>> > at
>> >org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounte
>> rs(ClientServic
>> >eDelegate.java:370)
>> > at
>> >org.apache.hadoop.mapred.YARNRunner.getJobCounters(YARNRunner.java:511)
>> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756)
>> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753)
>> > at java.security.AccessController.doPrivileged(Native Method)
>> > at javax.security.auth.Subject.doAs(Subject.java:415)
>> > at
>> >org.apache.hadoop.security.UserGroupInformation.doAs(UserGr
>> oupInformation.
>> >java:1491)
>> > at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753)
>> > at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1361)
>> > at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1289)
>> > at
>> >org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForComple
>> tion(AbstractHa
>> >doopJob.java:134)
>> > at
>> >org.apache.kylin.job.hadoop.cardinality.HiveColumnCardinalityJob.run(
>> HiveC
>> >olumnCardinalityJob.java:114)
>> > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>> > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
>> > at
>> >org.apache.kylin.job.common.HadoopShellExecutable.doWork(Ha
>> doopShellExecut
>> >able.java:62)
>> > at
>> >org.apache.kylin.job.execution.AbstractExecutable.execute(A
>> bstractExecutab
>> >le.java:99)
>> > at
>> >org.apache.kylin.job.execution.DefaultChainedExecutable.doW
>> ork(DefaultChai
>> >nedExecutable.java:50)
>> > at
>> >org.apache.kylin.job.execution.AbstractExecutable.execute(A
>> bstractExecutab
>> >le.java:99)
>> > at
>> >org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRu
>> nner.run(Defaul
>> >tScheduler.java:132)
>> > at
>> >java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoo
>> lExecutor.java:
>> >1145)
>> > at
>> >java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPo
>> olExecutor.java
>> >:615)
>> > at java.lang.Thread.run(Thread.java:745)
>> >
>> >Regards,
>> >Santosh Akhilesh
>> >Bangalore R&D
>> >HUAWEI TECHNOLOGIES CO.,LTD.
>> >
>> >www.huawei.com
>> >-----------------------------------------------------------
>> ---------------
>> >-----------------------------------------------------------
>> >This e-mail and its attachments contain confidential information from
>> >HUAWEI, which
>> >is intended only for the person or entity whose address is listed
>>above.
>> >Any use of the
>> >information contained herein in any way (including, but not limited to,
>> >total or partial
>> >disclosure, reproduction, or dissemination) by persons other than the
>> >intended
>> >recipient(s) is prohibited. If you receive this e-mail in error, please
>> >notify the sender by
>> >phone or email immediately and delete it!
>> >
>> >________________________________________
>> >From: Shi, Shaofeng [[email protected]]
>> >Sent: Thursday, February 26, 2015 11:32 AM
>> >To: [email protected]
>> >Cc: Kulbhushan Rana
>> >Subject: Re: Error while making cube & Measure option is not responding
>> >on GUI
>> >
>> >Hi Santosh, hive table importing issue should not impact on cube
>>saving.
>> >
>> >If you couldn’t save the cube, firstly please check whether there is
>>error
>> >in the tomcat’s log; If not please check your web browser; We suggest
>>use
>> >Firefox (with firebug add-on) or Chrome, open the JS console (press
>>F12)
>> >and then operate web UI, check whether there is any error reported in
>> >browser.
>> >
>> >
>> >On 2/26/15, 1:08 PM, "Santoshakhilesh" <[email protected]>
>> >wrote:
>> >
>> >>Hi Shaofeng ,
>> >>   Thanks for replying.
>> >>   Yes I am checking the yarn exception, But I find that this error
>>comes
>> >>while importing the hive table to kylin.
>> >>   Even if this error comes , hive tables is exported successfully in
>> >>kylin. Is this the reason why cube saving has failed ?
>> >>
>> >>   Next step when I go on creating the cube for following schema  I
>>get
>> >>error at last step while saving and  I am unable to add any measures ,
>> >>clicking on measure option just dont pop up any dialog,
>> >>
>> >>I am using star schema with fact_sales as fact table and dim_* as
>> >>dimension tables.
>> >>
>> >> fact_sales:
>> >> storeid                 int
>> >> itemid                  int
>> >> custid                  int
>> >> qty                     int
>> >> price                   double
>> >>
>> >>dim_customer
>> >> custid                  int
>> >> name                    string
>> >>
>> >> dim_item
>> >> itemid                  int
>> >> category                string
>> >> brand                   string
>> >> color                   string
>> >>
>> >>dim_store
>> >> storeid                 int
>> >> city                    string
>> >> state                   string
>> >>
>> >>JSON is as below.
>> >> The JSON is as below.
>> >>
>> >> {
>> >>   "name": "Retail_Cube",
>> >>   "description": "",
>> >>   "dimensions": [
>> >>     {
>> >>       "name": "RETAIL.FACT_SALES.STOREID",
>> >>       "table": "RETAIL.FACT_SALES",
>> >>       "hierarchy": false,
>> >>       "derived": null,
>> >>       "column": [
>> >>         "STOREID"
>> >>       ],
>> >>       "id": 1
>> >>     },
>> >>     {
>> >>       "name": "RETAIL.FACT_SALES.ITEMID",
>> >>       "table": "RETAIL.FACT_SALES",
>> >>       "hierarchy": false,
>> >>       "derived": null,
>> >>       "column": [
>> >>         "ITEMID"
>> >>       ],
>> >>       "id": 2
>> >>     },
>> >>     {
>> >>       "name": "RETAIL.FACT_SALES.CUSTID",
>> >>       "table": "RETAIL.FACT_SALES",
>> >>       "hierarchy": false,
>> >>       "derived": null,
>> >>       "column": [
>> >>         "CUSTID"
>> >>       ],
>> >>       "id": 3
>> >>     }
>> >>   ],
>> >>   "measures": [
>> >>     {
>> >>       "id": 1,
>> >>       "name": "_COUNT_",
>> >>       "function": {
>> >>         "expression": "COUNT",
>> >>         "returntype": "bigint",
>> >>         "parameter": {
>> >>           "type": "constant",
>> >>           "value": 1
>> >>         }
>> >>       }
>> >>     }
>> >>   ],
>> >>   "rowkey": {
>> >>     "rowkey_columns": [
>> >>       {
>> >>         "column": "STOREID",
>> >>         "length": 0,
>> >>         "dictionary": "true",
>> >>         "mandatory": false
>> >>       },
>> >>       {
>> >>         "column": "ITEMID",
>> >>         "length": 0,
>> >>         "dictionary": "true",
>> >>         "mandatory": false
>> >>       },
>> >>       {
>> >>         "column": "CUSTID",
>> >>         "length": 0,
>> >>         "dictionary": "true",
>> >>         "mandatory": false
>> >>       }
>> >>     ],
>> >>     "aggregation_groups": [
>> >>       [
>> >>         "STOREID",
>> >>         "ITEMID",
>> >>         "CUSTID"
>> >>       ]
>> >>     ]
>> >>   },
>> >>   "notify_list": [],
>> >>   "capacity": "",
>> >>   "hbase_mapping": {
>> >>     "column_family": [
>> >>       {
>> >>         "name": "f1",
>> >>         "columns": [
>> >>           {
>> >>             "qualifier": "m",
>> >>             "measure_refs": [
>> >>               "_COUNT_"
>> >>             ]
>> >>           }
>> >>         ]
>> >>       }
>> >>     ]
>> >>   },
>> >>   "project": "RetailProject",
>> >>   "model_name": "Retail_Cube"
>> >> }
>> >>
>> >>Regards,
>> >>Santosh Akhilesh
>> >>Bangalore R&D
>> >>HUAWEI TECHNOLOGIES CO.,LTD.
>> >>
>> >>www.huawei.com
>> >>----------------------------------------------------------
>> ---------------
>> >>-
>> >>-----------------------------------------------------------
>> >>This e-mail and its attachments contain confidential information from
>> >>HUAWEI, which
>> >>is intended only for the person or entity whose address is listed
>>above.
>> >>Any use of the
>> >>information contained herein in any way (including, but not limited
>>to,
>> >>total or partial
>> >>disclosure, reproduction, or dissemination) by persons other than the
>> >>intended
>> >>recipient(s) is prohibited. If you receive this e-mail in error,
>>please
>> >>notify the sender by
>> >>phone or email immediately and delete it!
>> >>
>> >>________________________________________
>> >>From: Shi, Shaofeng [[email protected]]
>> >>Sent: Thursday, February 26, 2015 7:01 AM
>> >>To: [email protected]
>> >>Subject: Re: Error while making cube & Measure option is not
>>responding
>> >>on GUI
>> >>
>> >>Hi Santosh,
>> >>
>> >>It looks like hadoop failed to execute some shell command in the
>> >>container; You need dive into hadoop to see what¹s the concrete error.
>> >>You
>> >>can use yarn logs command to fetch all logs:
>> >>
>> >>yarn logs -applicationId <app_id>
>> >>
>> >>
>> >>On 2/25/15, 7:39 PM, "Santosh Akhilesh" <[email protected]>
>> >>wrote:
>> >>
>> >>>Hi Luke / Shaofeng ,
>> >>>           Can you please help me to check this issue.
>> >>>Regards,
>> >>>Santosh Akhilesh
>> >>>
>> >>>On Tue, Feb 24, 2015 at 10:41 PM, Santosh Akhilesh <
>> >>>[email protected]> wrote:
>> >>>
>> >>>> Hi All ,
>> >>>>         is it because of following error in map reduce job ? what
>> >>>>could
>> >>>>be
>> >>>> way to resolve this , a google search says that its issue of Yarn
>> >>>>class
>> >>>> path , but I am not sure what it is ?
>> >>>>
>> >>>> Kylin Hive Column Cardinality Job table=RETAIL.FACT_SALES
>> >>>> output=/tmp/cardinality/RETAIL.FACT_SALES
>> >>>>
>> >>>> Application application_1424791969399_0008 failed 2 times due to AM
>> >>>> Container for appattempt_1424791969399_0008_000002 exited with
>> >>>>exitCode: 1
>> >>>> For more detailed output, check application tracking page:
>> >>>> http://santosh:8088/proxy/application_1424791969399_0008/Then,
>>click
>> >>>>on
>> >>>> links to logs of each attempt.
>> >>>> Diagnostics: Exception from container-launch.
>> >>>> Container id: container_1424791969399_0008_02_000001
>> >>>> Exit code: 1
>> >>>> Stack trace: ExitCodeException exitCode=1:
>> >>>> at org.apache.hadoop.util.Shell.runCommand(Shell.java:538)
>> >>>> at org.apache.hadoop.util.Shell.run(Shell.java:455)
>> >>>> at
>> >>>>
>> >>>>org.apache.hadoop.util.Shell$ShellCommandExecutor.execut
>> e(Shell.java:71
>> >>>>5
>> >>>>)
>> >>>> at
>> >>>>
>> >>>>org.apache.hadoop.yarn.server.nodemanager.DefaultContain
>> erExecutor.laun
>> >>>>c
>> >>>>h
>> >>>>Container(DefaultContainerExecutor.java:211)
>> >>>> at
>> >>>>
>> >>>>org.apache.hadoop.yarn.server.nodemanager.containermanag
>> er.launcher.Con
>> >>>>t
>> >>>>a
>> >>>>inerLaunch.call(ContainerLaunch.java:302)
>> >>>> at
>> >>>>
>> >>>>org.apache.hadoop.yarn.server.nodemanager.containermanag
>> er.launcher.Con
>> >>>>t
>> >>>>a
>> >>>>inerLaunch.call(ContainerLaunch.java:82)
>> >>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>> >>>> at
>> >>>>
>> >>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread
>> PoolExecutor.ja
>> >>>>v
>> >>>>a
>> >>>>:1145)
>> >>>> at
>> >>>>
>> >>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa
>> dPoolExecutor.j
>> >>>>a
>> >>>>v
>> >>>>a:615)
>> >>>> at java.lang.Thread.run(Thread.java:745)
>> >>>> Container exited with a non-zero exit code 1
>> >>>> Failing this attempt. Failing the application.
>> >>>>
>> >>>> ---------- Forwarded message ----------
>> >>>> From: Santoshakhilesh <[email protected]>
>> >>>> Date: Tue, Feb 24, 2015 at 7:41 PM
>> >>>> Subject: FW: Error while making cube & Measure option is not
>> >>>>responding
>> >>>>on
>> >>>> GUI
>> >>>> To: "[email protected]"
>><[email protected]>
>> >>>>
>> >>>>
>> >>>> hi ,
>> >>>>    please someone give me a hand to resolve this issue , thanks.
>> >>>>
>> >>>> Regards,
>> >>>> Santosh Akhilesh
>> >>>> Bangalore R&D
>> >>>> HUAWEI TECHNOLOGIES CO.,LTD.
>> >>>>
>> >>>> www.huawei.com
>> >>>>
>> >>>>
>> >>>>--------------------------------------------------------
>> ---------------
>> >>>>-
>> >>>>-
>> >>>>------------------------------------------------------------
>> >>>> This e-mail and its attachments contain confidential information
>>from
>> >>>> HUAWEI, which
>> >>>> is intended only for the person or entity whose address is listed
>> >>>>above.
>> >>>> Any use of the
>> >>>> information contained herein in any way (including, but not limited
>> >>>>to,
>> >>>> total or partial
>> >>>> disclosure, reproduction, or dissemination) by persons other than
>>the
>> >>>> intended
>> >>>> recipient(s) is prohibited. If you receive this e-mail in error,
>> >>>>please
>> >>>> notify the sender by
>> >>>> phone or email immediately and delete it!
>> >>>>
>> >>>> ________________________________________
>> >>>> From: Santoshakhilesh [[email protected]]
>> >>>> Sent: Tuesday, February 24, 2015 12:55 PM
>> >>>> To: [email protected]
>> >>>> Cc: Kulbhushan Rana
>> >>>> Subject: FW: Error while making cube & Measure option is not
>> >>>>responding
>> >>>>on
>> >>>> GUI
>> >>>>
>> >>>> 2. If  I ignore and continue and try to save the cube I get an
>> >>>>exception
>> >>>> in Kylin.log , I have checked the path is set correctly and
>> >>>>HCatInputFormat
>> >>>> this file is present in hive-hcatalog-core-0.14.0.jar . Please let
>>me
>> >>>>know
>> >>>> what can I do to resolve this ?
>> >>>>
>> >>>>  -- This was path issue , now no more exception in kylin.log
>> >>>>
>> >>>> But saveing cube still fails with error. And still can't add
>>measures.
>> >>>>
>> >>>> Error Message
>> >>>> Failed to take action.
>> >>>>
>> >>>> In log I can find no exception. Following is the last log in
>>kylin.log
>> >>>>
>> >>>> [pool-3-thread-1]:[2015-02-24
>> >>>>
>> >>>>20:47:15,613][INFO][org.apache.kylin.job.impl.threadpool
>> .DefaultSchedul
>> >>>>e
>> >>>>r
>> >>>>$FetcherRunner.run(DefaultScheduler.java:117)]
>> >>>> - Job Fetcher: 0 running, 0 actual running, 0 ready, 6 others
>> >>>> [http-bio-7070-exec-2]:[2015-02-24
>> >>>>
>> >>>>20:47:51,610][DEBUG][org.apache.kylin.rest.controller.Cu
>> beController.de
>> >>>>s
>> >>>>e
>> >>>>rializeDataModelDesc(CubeController.java:459)]
>> >>>> - Saving cube {
>> >>>>   "name": "",
>> >>>>   "fact_table": "RETAIL.FACT_SALES",
>> >>>>   "lookups": [],
>> >>>>   "filter_condition": "",
>> >>>>   "capacity": "SMALL",
>> >>>>   "partition_desc": {
>> >>>>     "partition_date_column": "",
>> >>>>     "partition_date_start": 0,
>> >>>>     "partition_type": "APPEND"
>> >>>>   },
>> >>>>   "last_modified": 0
>> >>>> }
>> >>>>
>> >>>>
>> >>>> local access logs all with 200 , so seems ok.
>> >>>>
>> >>>> 10.18.146.105 - - [24/Feb/2015:20:46:56 +0800] "GET
>> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:07 +0800] "GET
>> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:27 +0800] "GET
>> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:28 +0800] "GET
>> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:34 +0800] "GET
>> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:48 +0800] "GET
>> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:51 +0800] "POST
>>/kylin/api/cubes
>> >>>> HTTP/1.1" 200 701
>> >>>>
>> >>>>
>> >>>> Regards,
>> >>>> Santosh Akhilesh
>> >>>> Bangalore R&D
>> >>>> HUAWEI TECHNOLOGIES CO.,LTD.
>> >>>>
>> >>>> www.huawei.com
>> >>>>
>> >>>>
>> >>>>--------------------------------------------------------
>> ---------------
>> >>>>-
>> >>>>-
>> >>>>------------------------------------------------------------
>> >>>> This e-mail and its attachments contain confidential information
>>from
>> >>>> HUAWEI, which
>> >>>> is intended only for the person or entity whose address is listed
>> >>>>above.
>> >>>> Any use of the
>> >>>> information contained herein in any way (including, but not limited
>> >>>>to,
>> >>>> total or partial
>> >>>> disclosure, reproduction, or dissemination) by persons other than
>>the
>> >>>> intended
>> >>>> recipient(s) is prohibited. If you receive this e-mail in error,
>> >>>>please
>> >>>> notify the sender by
>> >>>> phone or email immediately and delete it!
>> >>>>
>> >>>> ________________________________________
>> >>>> From: Santoshakhilesh [[email protected]]
>> >>>> Sent: Tuesday, February 24, 2015 12:09 PM
>> >>>> To: [email protected]
>> >>>> Cc: Kulbhushan Rana
>> >>>> Subject: Error while making cube & Measure option is not
>>responding on
>> >>>>GUI
>> >>>>
>> >>>> Hi All ,
>> >>>>
>> >>>>     I am building a simple cube for test and using the binary build
>> >>>>0.7.1
>> >>>> . I have following hive tables with columns.
>> >>>>
>> >>>>
>> >>>>
>> >>>> fact_sales:
>> >>>>
>> >>>> storeid                 int
>> >>>> itemid                  int
>> >>>> custid                  int
>> >>>> qty                     int
>> >>>> price                   double
>> >>>>
>> >>>> dim_customer
>> >>>> custid                  int
>> >>>> name                    string
>> >>>>
>> >>>> dim_item
>> >>>>
>> >>>> itemid                  int
>> >>>> category                string
>> >>>> brand                   string
>> >>>> color                   string
>> >>>>
>> >>>> dim_store
>> >>>>
>> >>>> storeid                 int
>> >>>> city                    string
>> >>>> state                   string
>> >>>>
>> >>>> Please help me to answer following issues;
>> >>>>
>> >>>>
>> >>>>
>> >>>> 1. When I go to measure section and click on measure option ,
>>there is
>> >>>>no
>> >>>> response , I want add measure on qty and price with sum
>> >>>>
>> >>>> 2. If  I ignore and continue and try to save the cube I get an
>> >>>>exception
>> >>>> in Kylin.log , I have checked the path is set correctly and
>> >>>>HCatInputFormat
>> >>>> this file is present in hive-hcatalog-core-0.14.0.jar . Please let
>>me
>> >>>>know
>> >>>> what can I do to resolve this ?
>> >>>>
>> >>>> 3. Also I have another question since this is a test and data is
>>small
>> >>>>I
>> >>>> have not partitioned the fact table , is it ok to skip partition
>>stage
>> >>>> while cube build ?
>> >>>>
>> >>>>
>> >>>>
>> >>>> Exception
>> >>>>
>> >>>> pool-4-thread-4]:[2015-02-24
>> >>>>
>> >>>>19:26:32,577][ERROR][org.apache.kylin.job.impl.threadpoo
>> l.DefaultSchedu
>> >>>>l
>> >>>>e
>> >>>>r$JobRunner.run(DefaultScheduler.java:134)]
>> >>>> - ExecuteException job:c3532a6f-97ea-474a-b36a-218dd517cedb
>> >>>> org.apache.kylin.job.exception.ExecuteException:
>> >>>> org.apache.kylin.job.exception.ExecuteException:
>> >>>> java.lang.NoClassDefFoundError:
>> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat
>> >>>>  at
>> >>>>
>> >>>>org.apache.kylin.job.execution.AbstractExecutable.execut
>> e(AbstractExecu
>> >>>>t
>> >>>>a
>> >>>>ble.java:102)
>> >>>>  at
>> >>>>
>> >>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$Jo
>> bRunner.run(Def
>> >>>>a
>> >>>>u
>> >>>>ltScheduler.java:132)
>> >>>>  at
>> >>>>
>> >>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread
>> PoolExecutor.ja
>> >>>>v
>> >>>>a
>> >>>>:1145)
>> >>>>  at
>> >>>>
>> >>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa
>> dPoolExecutor.j
>> >>>>a
>> >>>>v
>> >>>>a:615)
>> >>>>  at java.lang.Thread.run(Thread.java:745)
>> >>>> Caused by: org.apache.kylin.job.exception.ExecuteException:
>> >>>> java.lang.NoClassDefFoundError:
>> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat
>> >>>>
>> >>>>
>> >>>>
>> >>>> The JSON is as below.
>> >>>>
>> >>>>
>> >>>>
>> >>>> {
>> >>>>   "name": "Retail_Cube",
>> >>>>   "description": "",
>> >>>>   "dimensions": [
>> >>>>     {
>> >>>>       "name": "RETAIL.FACT_SALES.STOREID",
>> >>>>       "table": "RETAIL.FACT_SALES",
>> >>>>       "hierarchy": false,
>> >>>>       "derived": null,
>> >>>>       "column": [
>> >>>>         "STOREID"
>> >>>>       ],
>> >>>>       "id": 1
>> >>>>     },
>> >>>>     {
>> >>>>       "name": "RETAIL.FACT_SALES.ITEMID",
>> >>>>       "table": "RETAIL.FACT_SALES",
>> >>>>       "hierarchy": false,
>> >>>>       "derived": null,
>> >>>>       "column": [
>> >>>>         "ITEMID"
>> >>>>       ],
>> >>>>       "id": 2
>> >>>>     },
>> >>>>     {
>> >>>>       "name": "RETAIL.FACT_SALES.CUSTID",
>> >>>>       "table": "RETAIL.FACT_SALES",
>> >>>>       "hierarchy": false,
>> >>>>       "derived": null,
>> >>>>       "column": [
>> >>>>         "CUSTID"
>> >>>>       ],
>> >>>>       "id": 3
>> >>>>     }
>> >>>>   ],
>> >>>>   "measures": [
>> >>>>     {
>> >>>>       "id": 1,
>> >>>>       "name": "_COUNT_",
>> >>>>       "function": {
>> >>>>         "expression": "COUNT",
>> >>>>         "returntype": "bigint",
>> >>>>         "parameter": {
>> >>>>           "type": "constant",
>> >>>>           "value": 1
>> >>>>         }
>> >>>>       }
>> >>>>     }
>> >>>>   ],
>> >>>>   "rowkey": {
>> >>>>     "rowkey_columns": [
>> >>>>       {
>> >>>>         "column": "STOREID",
>> >>>>         "length": 0,
>> >>>>         "dictionary": "true",
>> >>>>         "mandatory": false
>> >>>>       },
>> >>>>       {
>> >>>>         "column": "ITEMID",
>> >>>>         "length": 0,
>> >>>>         "dictionary": "true",
>> >>>>         "mandatory": false
>> >>>>       },
>> >>>>       {
>> >>>>         "column": "CUSTID",
>> >>>>         "length": 0,
>> >>>>         "dictionary": "true",
>> >>>>         "mandatory": false
>> >>>>       }
>> >>>>     ],
>> >>>>     "aggregation_groups": [
>> >>>>       [
>> >>>>         "STOREID",
>> >>>>         "ITEMID",
>> >>>>         "CUSTID"
>> >>>>       ]
>> >>>>     ]
>> >>>>   },
>> >>>>   "notify_list": [],
>> >>>>   "capacity": "",
>> >>>>   "hbase_mapping": {
>> >>>>     "column_family": [
>> >>>>       {
>> >>>>         "name": "f1",
>> >>>>         "columns": [
>> >>>>           {
>> >>>>             "qualifier": "m",
>> >>>>             "measure_refs": [
>> >>>>               "_COUNT_"
>> >>>>             ]
>> >>>>           }
>> >>>>         ]
>> >>>>       }
>> >>>>     ]
>> >>>>   },
>> >>>>   "project": "RetailProject",
>> >>>>   "model_name": "Retail_Cube"
>> >>>> }
>> >>>>
>> >>>>
>> >>>>
>> >>>>
>> >>>>
>> >>>> Regards,
>> >>>> Santosh Akhilesh
>> >>>> Bangalore R&D
>> >>>> HUAWEI TECHNOLOGIES CO.,LTD.
>> >>>>
>> >>>> www.huawei.com
>> >>>>
>> >>>>
>> >>>>--------------------------------------------------------
>> ---------------
>> >>>>-
>> >>>>-
>> >>>>------------------------------------------------------------
>> >>>> This e-mail and its attachments contain confidential information
>>from
>> >>>> HUAWEI, which
>> >>>> is intended only for the person or entity whose address is listed
>> >>>>above.
>> >>>> Any use of the
>> >>>> information contained herein in any way (including, but not limited
>> >>>>to,
>> >>>> total or partial
>> >>>> disclosure, reproduction, or dissemination) by persons other than
>>the
>> >>>> intended
>> >>>> recipient(s) is prohibited. If you receive this e-mail in error,
>> >>>>please
>> >>>> notify the sender by
>> >>>> phone or email immediately and delete it!
>> >>>>
>> >>>>
>> >>>>
>> >>>> --
>> >>>> Regards,
>> >>>> Santosh Akhilesh
>> >>>> +91-0-9845482201
>> >>>>
>> >>>
>> >>>
>> >>>
>> >>>--
>> >>>Regards,
>> >>>Santosh Akhilesh
>> >>>+91-0-9845482201
>>
>>

Reply via email to