Hi Shaofeng
So what do you suggest , how should I proceed further with this release?
Will there be a patch? Any alternate way I can create cube?
Please suggest.
Regards
Santosh
On Thu, 26 Feb 2015 at 3:04 pm, Shi, Shaofeng <[email protected]> wrote:

> Hi Santosh,
>
> 0.7.1 hasn’t been formally released; from 0.6.x to 0.7.x we have metadata
> structure change; While, the web UI (cube wizard) for this change hasn’t
> been stabilized; So it is not strange that you got trouble when saving the
> cube;
>
> @Jason, any idea about the JS error?
>
> On 2/26/15, 5:08 PM, "Santosh Akhilesh" <[email protected]> wrote:
>
> >Hi Shaofeng,
> >
> >I am using the binary distribution 0.7.1. I have not been able to save
> >cube
> >even once. I have tried creating new project and from local machine and
> >server machine. But I am always stuck with this error. I am never allowed
> >to add measures and never been able to save the cube. I also see the
> >kylin.log and it always tries to save cube with append mode. One thing I
> >need to tell that at partition stage since I don't have a big fact table
> >now I have not partititioned the fact table and I skip this step. Does
> >this
> >have affect in saving the cube. Is this because some metadata is available
> >and it tries to modify the cube? I am using latest Hadoop 2.6.6. Yes kylin
> >propert I have not added the jar. I will add them and check. But cube
> >creation failure is really puzzling me. I could see no error logs in
> >kylin.log.
> >Regards
> >Santosh
> >On Thu, 26 Feb 2015 at 1:40 pm, Shi, Shaofeng <[email protected]> wrote:
> >
> >> Which version or code branch are you using? I assume you’re using the
> >> stable version from master; Seems you’re trying to edit an existing cube
> >> to add new measurement, try refresh your browser's cache; if it still
> >> couldn’t be saved, try to create a new cube;
> >>
> >> The two error traces in tomcat need be taken care:
> >>
> >> 1) java.lang.NoClassDefFoundError:
> >>org/apache/kylin/common/mr/KylinMapper
> >>         Please check kylin.properties file, making sure the
> >>“kylin.job.jar”
> >> points to a right jar file; It will be loaded in Map-reduce;
> >>
> >> 2) java.lang.IllegalArgumentException: No enum constant
> >> org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS
> >>         This indicates your hadoop version might be old; Please check
> >>and
> >> ensure
> >> hadoop version is 2.2 or above.
> >>
> >> On 2/26/15, 3:21 PM, "Santoshakhilesh" <[email protected]>
> >> wrote:
> >>
> >> >Hi Shaofeng ,
> >> >
> >> >   I am using chrome , When I click on button to add measures ,
> >>following
> >> >is error on chrome console. When I try to save the cube there is no
> >>error
> >> >in console. I just get a error dialog saying failed to take action and
> >> >gives me the JSON cube schema.
> >> >
> >> >Error on chrome debug console is as below;
> >> >
> >> > ReferenceError: CubeDescModel is not defined
> >> >    at h.$scope.addNewMeasure (scripts.min.0.js:15984)
> >> >    at scripts.min.0.js:180
> >> >    at scripts.min.0.js:197
> >> >    at h.$eval (scripts.min.0.js:119)
> >> >    at h.$apply (scripts.min.0.js:119)
> >> >    at HTMLButtonElement.<anonymous> (scripts.min.0.js:197)
> >> >    at HTMLButtonElement.m.event.dispatch (scripts.min.0.js:3)
> >> >    at HTMLButtonElement.r.handle
> >> >(scripts.min.0.js:3)scripts.min.0.js:100 (anonymous
> >> >function)scripts.min.0.js:77 (anonymous function)scripts.min.0.js:119
> >> >h.$applyscripts.min.0.js:197 (anonymous function)scripts.min.0.js:3
> >> >m.event.dispatchscripts.min.0.js:3 r.handle
> >> >
> >> >   About the hive table import , I got pass the run shell command
> >> >exception but it still fails the hadoop log is;
> >> >2015-02-26 20:46:48,332 INFO [main] org.apache.hadoop.mapred.
> YarnChild:
> >> >mapreduce.cluster.local.dir for child:
> >> >/tmp/hadoop-root/nm-local-dir/usercache/root/appcache/appli
> >> cation_14249530
> >> >91340_0002
> >> >2015-02-26 20:46:48,776 INFO [main]
> >> >org.apache.hadoop.conf.Configuration.deprecation: session.id is
> >> >deprecated. Instead, use dfs.metrics.session-id
> >> >2015-02-26 20:46:49,310 INFO [main] org.apache.hadoop.mapred.Task:
> >>Using
> >> >ResourceCalculatorProcessTree : [ ]
> >> >2015-02-26 20:46:49,386 FATAL [main]
> >>org.apache.hadoop.mapred.YarnChild:
> >> >Error running child : java.lang.NoClassDefFoundError:
> >> >org/apache/kylin/common/mr/KylinMapper
> >> > at java.lang.ClassLoader.defineClass1(Native Method)
> >> > at java.lang.ClassLoader.defineClass(ClassLoader.java:800)
> >> > at
> >> >java.security.SecureClassLoader.defineClass(
> SecureClassLoader.java:142)
> >> > at java.net.URLClassLoader.defineClass(URLClassLoader.java:449)
> >> > at java.net.URLClassLoader.access$100(URLClassLoader.java:71)
> >> > at java.net.URLClassLoader$1.run(URLClassLoader.java:361)
> >> > at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
> >> > at java.security.AccessController.doPrivileged(Native Method)
> >> > at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
> >> > at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
> >> > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
> >> >
> >> >tomcat logs:
> >> >usage: HiveColumnCardinalityJob
> >> > -output <path>        Output path
> >> > -table <table name>   The hive table name
> >> >[pool-4-thread-2]:[2015-02-26
> >> >20:47:49,936][ERROR][org.apache.kylin.job.common.HadoopShel
> >> lExecutable.doW
> >> >ork(HadoopShellExecutable.java:64)] - error execute
> >> >HadoopShellExecutable{id=d4730d26-7fe6-412e-9841-3288ab362c5b-00,
> >> >name=null, state=RUNNING}
> >> >java.lang.IllegalArgumentException: No enum constant
> >> >org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS
> >> > at java.lang.Enum.valueOf(Enum.java:236)
> >> > at
> >> >org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.valueOf(
> >> Framewo
> >> >rkCounterGroup.java:148)
> >> > at
> >> >org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.
> >> findCounter(Fra
> >> >meworkCounterGroup.java:182)
> >> > at
> >> >org.apache.hadoop.mapreduce.counters.AbstractCounters.findC
> >> ounter(Abstract
> >> >Counters.java:154)
> >> > at
> >> >org.apache.hadoop.mapreduce.TypeConverter.fromYarn(TypeConv
> >> erter.java:240)
> >> > at
> >> >org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounte
> >> rs(ClientServic
> >> >eDelegate.java:370)
> >> > at
> >> >org.apache.hadoop.mapred.YARNRunner.getJobCounters(
> YARNRunner.java:511)
> >> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756)
> >> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753)
> >> > at java.security.AccessController.doPrivileged(Native Method)
> >> > at javax.security.auth.Subject.doAs(Subject.java:415)
> >> > at
> >> >org.apache.hadoop.security.UserGroupInformation.doAs(UserGr
> >> oupInformation.
> >> >java:1491)
> >> > at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753)
> >> > at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1361)
> >> > at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1289)
> >> > at
> >> >org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForComple
> >> tion(AbstractHa
> >> >doopJob.java:134)
> >> > at
> >> >org.apache.kylin.job.hadoop.cardinality.HiveColumnCardinalityJob.run(
> >> HiveC
> >> >olumnCardinalityJob.java:114)
> >> > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> >> > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
> >> > at
> >> >org.apache.kylin.job.common.HadoopShellExecutable.doWork(Ha
> >> doopShellExecut
> >> >able.java:62)
> >> > at
> >> >org.apache.kylin.job.execution.AbstractExecutable.execute(A
> >> bstractExecutab
> >> >le.java:99)
> >> > at
> >> >org.apache.kylin.job.execution.DefaultChainedExecutable.doW
> >> ork(DefaultChai
> >> >nedExecutable.java:50)
> >> > at
> >> >org.apache.kylin.job.execution.AbstractExecutable.execute(A
> >> bstractExecutab
> >> >le.java:99)
> >> > at
> >> >org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRu
> >> nner.run(Defaul
> >> >tScheduler.java:132)
> >> > at
> >> >java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoo
> >> lExecutor.java:
> >> >1145)
> >> > at
> >> >java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPo
> >> olExecutor.java
> >> >:615)
> >> > at java.lang.Thread.run(Thread.java:745)
> >> >
> >> >Regards,
> >> >Santosh Akhilesh
> >> >Bangalore R&D
> >> >HUAWEI TECHNOLOGIES CO.,LTD.
> >> >
> >> >www.huawei.com
> >> >-----------------------------------------------------------
> >> ---------------
> >> >-----------------------------------------------------------
> >> >This e-mail and its attachments contain confidential information from
> >> >HUAWEI, which
> >> >is intended only for the person or entity whose address is listed
> >>above.
> >> >Any use of the
> >> >information contained herein in any way (including, but not limited to,
> >> >total or partial
> >> >disclosure, reproduction, or dissemination) by persons other than the
> >> >intended
> >> >recipient(s) is prohibited. If you receive this e-mail in error, please
> >> >notify the sender by
> >> >phone or email immediately and delete it!
> >> >
> >> >________________________________________
> >> >From: Shi, Shaofeng [[email protected]]
> >> >Sent: Thursday, February 26, 2015 11:32 AM
> >> >To: [email protected]
> >> >Cc: Kulbhushan Rana
> >> >Subject: Re: Error while making cube & Measure option is not responding
> >> >on GUI
> >> >
> >> >Hi Santosh, hive table importing issue should not impact on cube
> >>saving.
> >> >
> >> >If you couldn’t save the cube, firstly please check whether there is
> >>error
> >> >in the tomcat’s log; If not please check your web browser; We suggest
> >>use
> >> >Firefox (with firebug add-on) or Chrome, open the JS console (press
> >>F12)
> >> >and then operate web UI, check whether there is any error reported in
> >> >browser.
> >> >
> >> >
> >> >On 2/26/15, 1:08 PM, "Santoshakhilesh" <[email protected]>
> >> >wrote:
> >> >
> >> >>Hi Shaofeng ,
> >> >>   Thanks for replying.
> >> >>   Yes I am checking the yarn exception, But I find that this error
> >>comes
> >> >>while importing the hive table to kylin.
> >> >>   Even if this error comes , hive tables is exported successfully in
> >> >>kylin. Is this the reason why cube saving has failed ?
> >> >>
> >> >>   Next step when I go on creating the cube for following schema  I
> >>get
> >> >>error at last step while saving and  I am unable to add any measures ,
> >> >>clicking on measure option just dont pop up any dialog,
> >> >>
> >> >>I am using star schema with fact_sales as fact table and dim_* as
> >> >>dimension tables.
> >> >>
> >> >> fact_sales:
> >> >> storeid                 int
> >> >> itemid                  int
> >> >> custid                  int
> >> >> qty                     int
> >> >> price                   double
> >> >>
> >> >>dim_customer
> >> >> custid                  int
> >> >> name                    string
> >> >>
> >> >> dim_item
> >> >> itemid                  int
> >> >> category                string
> >> >> brand                   string
> >> >> color                   string
> >> >>
> >> >>dim_store
> >> >> storeid                 int
> >> >> city                    string
> >> >> state                   string
> >> >>
> >> >>JSON is as below.
> >> >> The JSON is as below.
> >> >>
> >> >> {
> >> >>   "name": "Retail_Cube",
> >> >>   "description": "",
> >> >>   "dimensions": [
> >> >>     {
> >> >>       "name": "RETAIL.FACT_SALES.STOREID",
> >> >>       "table": "RETAIL.FACT_SALES",
> >> >>       "hierarchy": false,
> >> >>       "derived": null,
> >> >>       "column": [
> >> >>         "STOREID"
> >> >>       ],
> >> >>       "id": 1
> >> >>     },
> >> >>     {
> >> >>       "name": "RETAIL.FACT_SALES.ITEMID",
> >> >>       "table": "RETAIL.FACT_SALES",
> >> >>       "hierarchy": false,
> >> >>       "derived": null,
> >> >>       "column": [
> >> >>         "ITEMID"
> >> >>       ],
> >> >>       "id": 2
> >> >>     },
> >> >>     {
> >> >>       "name": "RETAIL.FACT_SALES.CUSTID",
> >> >>       "table": "RETAIL.FACT_SALES",
> >> >>       "hierarchy": false,
> >> >>       "derived": null,
> >> >>       "column": [
> >> >>         "CUSTID"
> >> >>       ],
> >> >>       "id": 3
> >> >>     }
> >> >>   ],
> >> >>   "measures": [
> >> >>     {
> >> >>       "id": 1,
> >> >>       "name": "_COUNT_",
> >> >>       "function": {
> >> >>         "expression": "COUNT",
> >> >>         "returntype": "bigint",
> >> >>         "parameter": {
> >> >>           "type": "constant",
> >> >>           "value": 1
> >> >>         }
> >> >>       }
> >> >>     }
> >> >>   ],
> >> >>   "rowkey": {
> >> >>     "rowkey_columns": [
> >> >>       {
> >> >>         "column": "STOREID",
> >> >>         "length": 0,
> >> >>         "dictionary": "true",
> >> >>         "mandatory": false
> >> >>       },
> >> >>       {
> >> >>         "column": "ITEMID",
> >> >>         "length": 0,
> >> >>         "dictionary": "true",
> >> >>         "mandatory": false
> >> >>       },
> >> >>       {
> >> >>         "column": "CUSTID",
> >> >>         "length": 0,
> >> >>         "dictionary": "true",
> >> >>         "mandatory": false
> >> >>       }
> >> >>     ],
> >> >>     "aggregation_groups": [
> >> >>       [
> >> >>         "STOREID",
> >> >>         "ITEMID",
> >> >>         "CUSTID"
> >> >>       ]
> >> >>     ]
> >> >>   },
> >> >>   "notify_list": [],
> >> >>   "capacity": "",
> >> >>   "hbase_mapping": {
> >> >>     "column_family": [
> >> >>       {
> >> >>         "name": "f1",
> >> >>         "columns": [
> >> >>           {
> >> >>             "qualifier": "m",
> >> >>             "measure_refs": [
> >> >>               "_COUNT_"
> >> >>             ]
> >> >>           }
> >> >>         ]
> >> >>       }
> >> >>     ]
> >> >>   },
> >> >>   "project": "RetailProject",
> >> >>   "model_name": "Retail_Cube"
> >> >> }
> >> >>
> >> >>Regards,
> >> >>Santosh Akhilesh
> >> >>Bangalore R&D
> >> >>HUAWEI TECHNOLOGIES CO.,LTD.
> >> >>
> >> >>www.huawei.com
> >> >>----------------------------------------------------------
> >> ---------------
> >> >>-
> >> >>-----------------------------------------------------------
> >> >>This e-mail and its attachments contain confidential information from
> >> >>HUAWEI, which
> >> >>is intended only for the person or entity whose address is listed
> >>above.
> >> >>Any use of the
> >> >>information contained herein in any way (including, but not limited
> >>to,
> >> >>total or partial
> >> >>disclosure, reproduction, or dissemination) by persons other than the
> >> >>intended
> >> >>recipient(s) is prohibited. If you receive this e-mail in error,
> >>please
> >> >>notify the sender by
> >> >>phone or email immediately and delete it!
> >> >>
> >> >>________________________________________
> >> >>From: Shi, Shaofeng [[email protected]]
> >> >>Sent: Thursday, February 26, 2015 7:01 AM
> >> >>To: [email protected]
> >> >>Subject: Re: Error while making cube & Measure option is not
> >>responding
> >> >>on GUI
> >> >>
> >> >>Hi Santosh,
> >> >>
> >> >>It looks like hadoop failed to execute some shell command in the
> >> >>container; You need dive into hadoop to see what¹s the concrete error.
> >> >>You
> >> >>can use yarn logs command to fetch all logs:
> >> >>
> >> >>yarn logs -applicationId <app_id>
> >> >>
> >> >>
> >> >>On 2/25/15, 7:39 PM, "Santosh Akhilesh" <[email protected]>
> >> >>wrote:
> >> >>
> >> >>>Hi Luke / Shaofeng ,
> >> >>>           Can you please help me to check this issue.
> >> >>>Regards,
> >> >>>Santosh Akhilesh
> >> >>>
> >> >>>On Tue, Feb 24, 2015 at 10:41 PM, Santosh Akhilesh <
> >> >>>[email protected]> wrote:
> >> >>>
> >> >>>> Hi All ,
> >> >>>>         is it because of following error in map reduce job ? what
> >> >>>>could
> >> >>>>be
> >> >>>> way to resolve this , a google search says that its issue of Yarn
> >> >>>>class
> >> >>>> path , but I am not sure what it is ?
> >> >>>>
> >> >>>> Kylin Hive Column Cardinality Job table=RETAIL.FACT_SALES
> >> >>>> output=/tmp/cardinality/RETAIL.FACT_SALES
> >> >>>>
> >> >>>> Application application_1424791969399_0008 failed 2 times due to AM
> >> >>>> Container for appattempt_1424791969399_0008_000002 exited with
> >> >>>>exitCode: 1
> >> >>>> For more detailed output, check application tracking page:
> >> >>>> http://santosh:8088/proxy/application_1424791969399_0008/Then,
> >>click
> >> >>>>on
> >> >>>> links to logs of each attempt.
> >> >>>> Diagnostics: Exception from container-launch.
> >> >>>> Container id: container_1424791969399_0008_02_000001
> >> >>>> Exit code: 1
> >> >>>> Stack trace: ExitCodeException exitCode=1:
> >> >>>> at org.apache.hadoop.util.Shell.runCommand(Shell.java:538)
> >> >>>> at org.apache.hadoop.util.Shell.run(Shell.java:455)
> >> >>>> at
> >> >>>>
> >> >>>>org.apache.hadoop.util.Shell$ShellCommandExecutor.execut
> >> e(Shell.java:71
> >> >>>>5
> >> >>>>)
> >> >>>> at
> >> >>>>
> >> >>>>org.apache.hadoop.yarn.server.nodemanager.DefaultContain
> >> erExecutor.laun
> >> >>>>c
> >> >>>>h
> >> >>>>Container(DefaultContainerExecutor.java:211)
> >> >>>> at
> >> >>>>
> >> >>>>org.apache.hadoop.yarn.server.nodemanager.containermanag
> >> er.launcher.Con
> >> >>>>t
> >> >>>>a
> >> >>>>inerLaunch.call(ContainerLaunch.java:302)
> >> >>>> at
> >> >>>>
> >> >>>>org.apache.hadoop.yarn.server.nodemanager.containermanag
> >> er.launcher.Con
> >> >>>>t
> >> >>>>a
> >> >>>>inerLaunch.call(ContainerLaunch.java:82)
> >> >>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262)
> >> >>>> at
> >> >>>>
> >> >>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread
> >> PoolExecutor.ja
> >> >>>>v
> >> >>>>a
> >> >>>>:1145)
> >> >>>> at
> >> >>>>
> >> >>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa
> >> dPoolExecutor.j
> >> >>>>a
> >> >>>>v
> >> >>>>a:615)
> >> >>>> at java.lang.Thread.run(Thread.java:745)
> >> >>>> Container exited with a non-zero exit code 1
> >> >>>> Failing this attempt. Failing the application.
> >> >>>>
> >> >>>> ---------- Forwarded message ----------
> >> >>>> From: Santoshakhilesh <[email protected]>
> >> >>>> Date: Tue, Feb 24, 2015 at 7:41 PM
> >> >>>> Subject: FW: Error while making cube & Measure option is not
> >> >>>>responding
> >> >>>>on
> >> >>>> GUI
> >> >>>> To: "[email protected]"
> >><[email protected]>
> >> >>>>
> >> >>>>
> >> >>>> hi ,
> >> >>>>    please someone give me a hand to resolve this issue , thanks.
> >> >>>>
> >> >>>> Regards,
> >> >>>> Santosh Akhilesh
> >> >>>> Bangalore R&D
> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD.
> >> >>>>
> >> >>>> www.huawei.com
> >> >>>>
> >> >>>>
> >> >>>>--------------------------------------------------------
> >> ---------------
> >> >>>>-
> >> >>>>-
> >> >>>>------------------------------------------------------------
> >> >>>> This e-mail and its attachments contain confidential information
> >>from
> >> >>>> HUAWEI, which
> >> >>>> is intended only for the person or entity whose address is listed
> >> >>>>above.
> >> >>>> Any use of the
> >> >>>> information contained herein in any way (including, but not limited
> >> >>>>to,
> >> >>>> total or partial
> >> >>>> disclosure, reproduction, or dissemination) by persons other than
> >>the
> >> >>>> intended
> >> >>>> recipient(s) is prohibited. If you receive this e-mail in error,
> >> >>>>please
> >> >>>> notify the sender by
> >> >>>> phone or email immediately and delete it!
> >> >>>>
> >> >>>> ________________________________________
> >> >>>> From: Santoshakhilesh [[email protected]]
> >> >>>> Sent: Tuesday, February 24, 2015 12:55 PM
> >> >>>> To: [email protected]
> >> >>>> Cc: Kulbhushan Rana
> >> >>>> Subject: FW: Error while making cube & Measure option is not
> >> >>>>responding
> >> >>>>on
> >> >>>> GUI
> >> >>>>
> >> >>>> 2. If  I ignore and continue and try to save the cube I get an
> >> >>>>exception
> >> >>>> in Kylin.log , I have checked the path is set correctly and
> >> >>>>HCatInputFormat
> >> >>>> this file is present in hive-hcatalog-core-0.14.0.jar . Please let
> >>me
> >> >>>>know
> >> >>>> what can I do to resolve this ?
> >> >>>>
> >> >>>>  -- This was path issue , now no more exception in kylin.log
> >> >>>>
> >> >>>> But saveing cube still fails with error. And still can't add
> >>measures.
> >> >>>>
> >> >>>> Error Message
> >> >>>> Failed to take action.
> >> >>>>
> >> >>>> In log I can find no exception. Following is the last log in
> >>kylin.log
> >> >>>>
> >> >>>> [pool-3-thread-1]:[2015-02-24
> >> >>>>
> >> >>>>20:47:15,613][INFO][org.apache.kylin.job.impl.threadpool
> >> .DefaultSchedul
> >> >>>>e
> >> >>>>r
> >> >>>>$FetcherRunner.run(DefaultScheduler.java:117)]
> >> >>>> - Job Fetcher: 0 running, 0 actual running, 0 ready, 6 others
> >> >>>> [http-bio-7070-exec-2]:[2015-02-24
> >> >>>>
> >> >>>>20:47:51,610][DEBUG][org.apache.kylin.rest.controller.Cu
> >> beController.de
> >> >>>>s
> >> >>>>e
> >> >>>>rializeDataModelDesc(CubeController.java:459)]
> >> >>>> - Saving cube {
> >> >>>>   "name": "",
> >> >>>>   "fact_table": "RETAIL.FACT_SALES",
> >> >>>>   "lookups": [],
> >> >>>>   "filter_condition": "",
> >> >>>>   "capacity": "SMALL",
> >> >>>>   "partition_desc": {
> >> >>>>     "partition_date_column": "",
> >> >>>>     "partition_date_start": 0,
> >> >>>>     "partition_type": "APPEND"
> >> >>>>   },
> >> >>>>   "last_modified": 0
> >> >>>> }
> >> >>>>
> >> >>>>
> >> >>>> local access logs all with 200 , so seems ok.
> >> >>>>
> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:46:56 +0800] "GET
> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:07 +0800] "GET
> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:27 +0800] "GET
> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:28 +0800] "GET
> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:34 +0800] "GET
> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:48 +0800] "GET
> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:51 +0800] "POST
> >>/kylin/api/cubes
> >> >>>> HTTP/1.1" 200 701
> >> >>>>
> >> >>>>
> >> >>>> Regards,
> >> >>>> Santosh Akhilesh
> >> >>>> Bangalore R&D
> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD.
> >> >>>>
> >> >>>> www.huawei.com
> >> >>>>
> >> >>>>
> >> >>>>--------------------------------------------------------
> >> ---------------
> >> >>>>-
> >> >>>>-
> >> >>>>------------------------------------------------------------
> >> >>>> This e-mail and its attachments contain confidential information
> >>from
> >> >>>> HUAWEI, which
> >> >>>> is intended only for the person or entity whose address is listed
> >> >>>>above.
> >> >>>> Any use of the
> >> >>>> information contained herein in any way (including, but not limited
> >> >>>>to,
> >> >>>> total or partial
> >> >>>> disclosure, reproduction, or dissemination) by persons other than
> >>the
> >> >>>> intended
> >> >>>> recipient(s) is prohibited. If you receive this e-mail in error,
> >> >>>>please
> >> >>>> notify the sender by
> >> >>>> phone or email immediately and delete it!
> >> >>>>
> >> >>>> ________________________________________
> >> >>>> From: Santoshakhilesh [[email protected]]
> >> >>>> Sent: Tuesday, February 24, 2015 12:09 PM
> >> >>>> To: [email protected]
> >> >>>> Cc: Kulbhushan Rana
> >> >>>> Subject: Error while making cube & Measure option is not
> >>responding on
> >> >>>>GUI
> >> >>>>
> >> >>>> Hi All ,
> >> >>>>
> >> >>>>     I am building a simple cube for test and using the binary build
> >> >>>>0.7.1
> >> >>>> . I have following hive tables with columns.
> >> >>>>
> >> >>>>
> >> >>>>
> >> >>>> fact_sales:
> >> >>>>
> >> >>>> storeid                 int
> >> >>>> itemid                  int
> >> >>>> custid                  int
> >> >>>> qty                     int
> >> >>>> price                   double
> >> >>>>
> >> >>>> dim_customer
> >> >>>> custid                  int
> >> >>>> name                    string
> >> >>>>
> >> >>>> dim_item
> >> >>>>
> >> >>>> itemid                  int
> >> >>>> category                string
> >> >>>> brand                   string
> >> >>>> color                   string
> >> >>>>
> >> >>>> dim_store
> >> >>>>
> >> >>>> storeid                 int
> >> >>>> city                    string
> >> >>>> state                   string
> >> >>>>
> >> >>>> Please help me to answer following issues;
> >> >>>>
> >> >>>>
> >> >>>>
> >> >>>> 1. When I go to measure section and click on measure option ,
> >>there is
> >> >>>>no
> >> >>>> response , I want add measure on qty and price with sum
> >> >>>>
> >> >>>> 2. If  I ignore and continue and try to save the cube I get an
> >> >>>>exception
> >> >>>> in Kylin.log , I have checked the path is set correctly and
> >> >>>>HCatInputFormat
> >> >>>> this file is present in hive-hcatalog-core-0.14.0.jar . Please let
> >>me
> >> >>>>know
> >> >>>> what can I do to resolve this ?
> >> >>>>
> >> >>>> 3. Also I have another question since this is a test and data is
> >>small
> >> >>>>I
> >> >>>> have not partitioned the fact table , is it ok to skip partition
> >>stage
> >> >>>> while cube build ?
> >> >>>>
> >> >>>>
> >> >>>>
> >> >>>> Exception
> >> >>>>
> >> >>>> pool-4-thread-4]:[2015-02-24
> >> >>>>
> >> >>>>19:26:32,577][ERROR][org.apache.kylin.job.impl.threadpoo
> >> l.DefaultSchedu
> >> >>>>l
> >> >>>>e
> >> >>>>r$JobRunner.run(DefaultScheduler.java:134)]
> >> >>>> - ExecuteException job:c3532a6f-97ea-474a-b36a-218dd517cedb
> >> >>>> org.apache.kylin.job.exception.ExecuteException:
> >> >>>> org.apache.kylin.job.exception.ExecuteException:
> >> >>>> java.lang.NoClassDefFoundError:
> >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat
> >> >>>>  at
> >> >>>>
> >> >>>>org.apache.kylin.job.execution.AbstractExecutable.execut
> >> e(AbstractExecu
> >> >>>>t
> >> >>>>a
> >> >>>>ble.java:102)
> >> >>>>  at
> >> >>>>
> >> >>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$Jo
> >> bRunner.run(Def
> >> >>>>a
> >> >>>>u
> >> >>>>ltScheduler.java:132)
> >> >>>>  at
> >> >>>>
> >> >>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread
> >> PoolExecutor.ja
> >> >>>>v
> >> >>>>a
> >> >>>>:1145)
> >> >>>>  at
> >> >>>>
> >> >>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa
> >> dPoolExecutor.j
> >> >>>>a
> >> >>>>v
> >> >>>>a:615)
> >> >>>>  at java.lang.Thread.run(Thread.java:745)
> >> >>>> Caused by: org.apache.kylin.job.exception.ExecuteException:
> >> >>>> java.lang.NoClassDefFoundError:
> >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat
> >> >>>>
> >> >>>>
> >> >>>>
> >> >>>> The JSON is as below.
> >> >>>>
> >> >>>>
> >> >>>>
> >> >>>> {
> >> >>>>   "name": "Retail_Cube",
> >> >>>>   "description": "",
> >> >>>>   "dimensions": [
> >> >>>>     {
> >> >>>>       "name": "RETAIL.FACT_SALES.STOREID",
> >> >>>>       "table": "RETAIL.FACT_SALES",
> >> >>>>       "hierarchy": false,
> >> >>>>       "derived": null,
> >> >>>>       "column": [
> >> >>>>         "STOREID"
> >> >>>>       ],
> >> >>>>       "id": 1
> >> >>>>     },
> >> >>>>     {
> >> >>>>       "name": "RETAIL.FACT_SALES.ITEMID",
> >> >>>>       "table": "RETAIL.FACT_SALES",
> >> >>>>       "hierarchy": false,
> >> >>>>       "derived": null,
> >> >>>>       "column": [
> >> >>>>         "ITEMID"
> >> >>>>       ],
> >> >>>>       "id": 2
> >> >>>>     },
> >> >>>>     {
> >> >>>>       "name": "RETAIL.FACT_SALES.CUSTID",
> >> >>>>       "table": "RETAIL.FACT_SALES",
> >> >>>>       "hierarchy": false,
> >> >>>>       "derived": null,
> >> >>>>       "column": [
> >> >>>>         "CUSTID"
> >> >>>>       ],
> >> >>>>       "id": 3
> >> >>>>     }
> >> >>>>   ],
> >> >>>>   "measures": [
> >> >>>>     {
> >> >>>>       "id": 1,
> >> >>>>       "name": "_COUNT_",
> >> >>>>       "function": {
> >> >>>>         "expression": "COUNT",
> >> >>>>         "returntype": "bigint",
> >> >>>>         "parameter": {
> >> >>>>           "type": "constant",
> >> >>>>           "value": 1
> >> >>>>         }
> >> >>>>       }
> >> >>>>     }
> >> >>>>   ],
> >> >>>>   "rowkey": {
> >> >>>>     "rowkey_columns": [
> >> >>>>       {
> >> >>>>         "column": "STOREID",
> >> >>>>         "length": 0,
> >> >>>>         "dictionary": "true",
> >> >>>>         "mandatory": false
> >> >>>>       },
> >> >>>>       {
> >> >>>>         "column": "ITEMID",
> >> >>>>         "length": 0,
> >> >>>>         "dictionary": "true",
> >> >>>>         "mandatory": false
> >> >>>>       },
> >> >>>>       {
> >> >>>>         "column": "CUSTID",
> >> >>>>         "length": 0,
> >> >>>>         "dictionary": "true",
> >> >>>>         "mandatory": false
> >> >>>>       }
> >> >>>>     ],
> >> >>>>     "aggregation_groups": [
> >> >>>>       [
> >> >>>>         "STOREID",
> >> >>>>         "ITEMID",
> >> >>>>         "CUSTID"
> >> >>>>       ]
> >> >>>>     ]
> >> >>>>   },
> >> >>>>   "notify_list": [],
> >> >>>>   "capacity": "",
> >> >>>>   "hbase_mapping": {
> >> >>>>     "column_family": [
> >> >>>>       {
> >> >>>>         "name": "f1",
> >> >>>>         "columns": [
> >> >>>>           {
> >> >>>>             "qualifier": "m",
> >> >>>>             "measure_refs": [
> >> >>>>               "_COUNT_"
> >> >>>>             ]
> >> >>>>           }
> >> >>>>         ]
> >> >>>>       }
> >> >>>>     ]
> >> >>>>   },
> >> >>>>   "project": "RetailProject",
> >> >>>>   "model_name": "Retail_Cube"
> >> >>>> }
> >> >>>>
> >> >>>>
> >> >>>>
> >> >>>>
> >> >>>>
> >> >>>> Regards,
> >> >>>> Santosh Akhilesh
> >> >>>> Bangalore R&D
> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD.
> >> >>>>
> >> >>>> www.huawei.com
> >> >>>>
> >> >>>>
> >> >>>>--------------------------------------------------------
> >> ---------------
> >> >>>>-
> >> >>>>-
> >> >>>>------------------------------------------------------------
> >> >>>> This e-mail and its attachments contain confidential information
> >>from
> >> >>>> HUAWEI, which
> >> >>>> is intended only for the person or entity whose address is listed
> >> >>>>above.
> >> >>>> Any use of the
> >> >>>> information contained herein in any way (including, but not limited
> >> >>>>to,
> >> >>>> total or partial
> >> >>>> disclosure, reproduction, or dissemination) by persons other than
> >>the
> >> >>>> intended
> >> >>>> recipient(s) is prohibited. If you receive this e-mail in error,
> >> >>>>please
> >> >>>> notify the sender by
> >> >>>> phone or email immediately and delete it!
> >> >>>>
> >> >>>>
> >> >>>>
> >> >>>> --
> >> >>>> Regards,
> >> >>>> Santosh Akhilesh
> >> >>>> +91-0-9845482201
> >> >>>>
> >> >>>
> >> >>>
> >> >>>
> >> >>>--
> >> >>>Regards,
> >> >>>Santosh Akhilesh
> >> >>>+91-0-9845482201
> >>
> >>
>
>

Reply via email to