Thanks for the links. Is there link for creating cube using rest ?
Before building I should ensure cube is created successfully.
And now I am downloading from download link of
http://www.kylin.io/
I hope this is latest , I have a zip file Kylin-Master.zip
Thanks.

On Thu, Feb 26, 2015 at 8:18 PM, Shi, Shaofeng <[email protected]> wrote:

> For the REST API, please refer to these two docs in kylin’s wiki:
>
> https://github.com/KylinOLAP/Kylin/wiki/Build-Cube-with-Restful-API
>
> https://github.com/KylinOLAP/Kylin/wiki/Restful-Service-Doc
>
>
> For the sample code, there is a RestClient.java in Kylin code base:
>
> https://github.com/KylinOLAP/Kylin/blob/master/common/src/main/java/com/kyl
> inolap/common/restclient/RestClient.java
>
> In it there is only 1 action which is “wipeCache”, but it can be a sample
> for how to invoking the REST services from Java.
>
>
> On 2/26/15, 7:03 PM, "Santosh Akhilesh" <[email protected]> wrote:
>
> >Hi Saofeng,
> >Can I use CLI or RESTI interface for 0.7.1?
> >Any sample code I can refer?
> >When can we expect a working Web interface?
> >Regards
> >Santosh Akhilesh
> >On Thu, 26 Feb 2015 at 3:09 pm, Santosh Akhilesh
> ><[email protected]>
> >wrote:
> >
> >> Hi Shaofeng
> >> So what do you suggest , how should I proceed further with this release?
> >> Will there be a patch? Any alternate way I can create cube?
> >> Please suggest.
> >> Regards
> >> Santosh
> >> On Thu, 26 Feb 2015 at 3:04 pm, Shi, Shaofeng <[email protected]> wrote:
> >>
> >>> Hi Santosh,
> >>>
> >>> 0.7.1 hasn’t been formally released; from 0.6.x to 0.7.x we have
> >>>metadata
> >>> structure change; While, the web UI (cube wizard) for this change
> >>>hasn’t
> >>> been stabilized; So it is not strange that you got trouble when saving
> >>>the
> >>> cube;
> >>>
> >>> @Jason, any idea about the JS error?
> >>>
> >>> On 2/26/15, 5:08 PM, "Santosh Akhilesh" <[email protected]>
> >>> wrote:
> >>>
> >>> >Hi Shaofeng,
> >>> >
> >>> >I am using the binary distribution 0.7.1. I have not been able to save
> >>> >cube
> >>> >even once. I have tried creating new project and from local machine
> >>>and
> >>> >server machine. But I am always stuck with this error. I am never
> >>>allowed
> >>> >to add measures and never been able to save the cube. I also see the
> >>> >kylin.log and it always tries to save cube with append mode. One
> >>>thing I
> >>> >need to tell that at partition stage since I don't have a big fact
> >>>table
> >>> >now I have not partititioned the fact table and I skip this step. Does
> >>> >this
> >>> >have affect in saving the cube. Is this because some metadata is
> >>> available
> >>> >and it tries to modify the cube? I am using latest Hadoop 2.6.6. Yes
> >>> kylin
> >>> >propert I have not added the jar. I will add them and check. But cube
> >>> >creation failure is really puzzling me. I could see no error logs in
> >>> >kylin.log.
> >>> >Regards
> >>> >Santosh
> >>> >On Thu, 26 Feb 2015 at 1:40 pm, Shi, Shaofeng <[email protected]>
> >>>wrote:
> >>> >
> >>> >> Which version or code branch are you using? I assume you’re using
> >>>the
> >>> >> stable version from master; Seems you’re trying to edit an existing
> >>> cube
> >>> >> to add new measurement, try refresh your browser's cache; if it
> >>>still
> >>> >> couldn’t be saved, try to create a new cube;
> >>> >>
> >>> >> The two error traces in tomcat need be taken care:
> >>> >>
> >>> >> 1) java.lang.NoClassDefFoundError:
> >>> >>org/apache/kylin/common/mr/KylinMapper
> >>> >>         Please check kylin.properties file, making sure the
> >>> >>“kylin.job.jar”
> >>> >> points to a right jar file; It will be loaded in Map-reduce;
> >>> >>
> >>> >> 2) java.lang.IllegalArgumentException: No enum constant
> >>> >> org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS
> >>> >>         This indicates your hadoop version might be old; Please
> >>>check
> >>> >>and
> >>> >> ensure
> >>> >> hadoop version is 2.2 or above.
> >>> >>
> >>> >> On 2/26/15, 3:21 PM, "Santoshakhilesh" <[email protected]
> >
> >>> >> wrote:
> >>> >>
> >>> >> >Hi Shaofeng ,
> >>> >> >
> >>> >> >   I am using chrome , When I click on button to add measures ,
> >>> >>following
> >>> >> >is error on chrome console. When I try to save the cube there is no
> >>> >>error
> >>> >> >in console. I just get a error dialog saying failed to take action
> >>>and
> >>> >> >gives me the JSON cube schema.
> >>> >> >
> >>> >> >Error on chrome debug console is as below;
> >>> >> >
> >>> >> > ReferenceError: CubeDescModel is not defined
> >>> >> >    at h.$scope.addNewMeasure (scripts.min.0.js:15984)
> >>> >> >    at scripts.min.0.js:180
> >>> >> >    at scripts.min.0.js:197
> >>> >> >    at h.$eval (scripts.min.0.js:119)
> >>> >> >    at h.$apply (scripts.min.0.js:119)
> >>> >> >    at HTMLButtonElement.<anonymous> (scripts.min.0.js:197)
> >>> >> >    at HTMLButtonElement.m.event.dispatch (scripts.min.0.js:3)
> >>> >> >    at HTMLButtonElement.r.handle
> >>> >> >(scripts.min.0.js:3)scripts.min.0.js:100 (anonymous
> >>> >> >function)scripts.min.0.js:77 (anonymous
> >>>function)scripts.min.0.js:119
> >>> >> >h.$applyscripts.min.0.js:197 (anonymous function)scripts.min.0.js:3
> >>> >> >m.event.dispatchscripts.min.0.js:3 r.handle
> >>> >> >
> >>> >> >   About the hive table import , I got pass the run shell command
> >>> >> >exception but it still fails the hadoop log is;
> >>> >> >2015-02-26 20:46:48,332 INFO [main] org.apache.hadoop.mapred.YarnC
> >>> hild:
> >>> >> >mapreduce.cluster.local.dir for child:
> >>> >> >/tmp/hadoop-root/nm-local-dir/usercache/root/appcache/appli
> >>> >> cation_14249530
> >>> >> >91340_0002
> >>> >> >2015-02-26 20:46:48,776 INFO [main]
> >>> >> >org.apache.hadoop.conf.Configuration.deprecation: session.id is
> >>> >> >deprecated. Instead, use dfs.metrics.session-id
> >>> >> >2015-02-26 20:46:49,310 INFO [main] org.apache.hadoop.mapred.Task:
> >>> >>Using
> >>> >> >ResourceCalculatorProcessTree : [ ]
> >>> >> >2015-02-26 20:46:49,386 FATAL [main]
> >>> >>org.apache.hadoop.mapred.YarnChild:
> >>> >> >Error running child : java.lang.NoClassDefFoundError:
> >>> >> >org/apache/kylin/common/mr/KylinMapper
> >>> >> > at java.lang.ClassLoader.defineClass1(Native Method)
> >>> >> > at java.lang.ClassLoader.defineClass(ClassLoader.java:800)
> >>> >> > at
> >>> >> >java.security.SecureClassLoader.defineClass(SecureClassLoad
> >>> er.java:142)
> >>> >> > at java.net.URLClassLoader.defineClass(URLClassLoader.java:449)
> >>> >> > at java.net.URLClassLoader.access$100(URLClassLoader.java:71)
> >>> >> > at java.net.URLClassLoader$1.run(URLClassLoader.java:361)
> >>> >> > at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
> >>> >> > at java.security.AccessController.doPrivileged(Native Method)
> >>> >> > at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
> >>> >> > at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
> >>> >> > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
> >>> >> >
> >>> >> >tomcat logs:
> >>> >> >usage: HiveColumnCardinalityJob
> >>> >> > -output <path>        Output path
> >>> >> > -table <table name>   The hive table name
> >>> >> >[pool-4-thread-2]:[2015-02-26
> >>> >> >20:47:49,936][ERROR][org.apache.kylin.job.common.HadoopShel
> >>> >> lExecutable.doW
> >>> >> >ork(HadoopShellExecutable.java:64)] - error execute
> >>> >> >HadoopShellExecutable{id=d4730d26-7fe6-412e-9841-3288ab362c5b-00,
> >>> >> >name=null, state=RUNNING}
> >>> >> >java.lang.IllegalArgumentException: No enum constant
> >>> >> >org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS
> >>> >> > at java.lang.Enum.valueOf(Enum.java:236)
> >>> >> > at
> >>> >> >org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.valueOf(
> >>> >> Framewo
> >>> >> >rkCounterGroup.java:148)
> >>> >> > at
> >>> >> >org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.
> >>> >> findCounter(Fra
> >>> >> >meworkCounterGroup.java:182)
> >>> >> > at
> >>> >> >org.apache.hadoop.mapreduce.counters.AbstractCounters.findC
> >>> >> ounter(Abstract
> >>> >> >Counters.java:154)
> >>> >> > at
> >>> >> >org.apache.hadoop.mapreduce.TypeConverter.fromYarn(TypeConv
> >>> >> erter.java:240)
> >>> >> > at
> >>> >> >org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounte
> >>> >> rs(ClientServic
> >>> >> >eDelegate.java:370)
> >>> >> > at
> >>> >> >org.apache.hadoop.mapred.YARNRunner.getJobCounters(YARNRunn
> >>> er.java:511)
> >>> >> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756)
> >>> >> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753)
> >>> >> > at java.security.AccessController.doPrivileged(Native Method)
> >>> >> > at javax.security.auth.Subject.doAs(Subject.java:415)
> >>> >> > at
> >>> >> >org.apache.hadoop.security.UserGroupInformation.doAs(UserGr
> >>> >> oupInformation.
> >>> >> >java:1491)
> >>> >> > at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753)
> >>> >> > at
> >>>org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1361)
> >>> >> > at
> >>>org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1289)
> >>> >> > at
> >>> >> >org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForComple
> >>> >> tion(AbstractHa
> >>> >> >doopJob.java:134)
> >>> >> > at
> >>> >>
> >>>>org.apache.kylin.job.hadoop.cardinality.HiveColumnCardinalityJob.run(
> >>> >> HiveC
> >>> >> >olumnCardinalityJob.java:114)
> >>> >> > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> >>> >> > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
> >>> >> > at
> >>> >> >org.apache.kylin.job.common.HadoopShellExecutable.doWork(Ha
> >>> >> doopShellExecut
> >>> >> >able.java:62)
> >>> >> > at
> >>> >> >org.apache.kylin.job.execution.AbstractExecutable.execute(A
> >>> >> bstractExecutab
> >>> >> >le.java:99)
> >>> >> > at
> >>> >> >org.apache.kylin.job.execution.DefaultChainedExecutable.doW
> >>> >> ork(DefaultChai
> >>> >> >nedExecutable.java:50)
> >>> >> > at
> >>> >> >org.apache.kylin.job.execution.AbstractExecutable.execute(A
> >>> >> bstractExecutab
> >>> >> >le.java:99)
> >>> >> > at
> >>> >> >org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRu
> >>> >> nner.run(Defaul
> >>> >> >tScheduler.java:132)
> >>> >> > at
> >>> >> >java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoo
> >>> >> lExecutor.java:
> >>> >> >1145)
> >>> >> > at
> >>> >> >java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPo
> >>> >> olExecutor.java
> >>> >> >:615)
> >>> >> > at java.lang.Thread.run(Thread.java:745)
> >>> >> >
> >>> >> >Regards,
> >>> >> >Santosh Akhilesh
> >>> >> >Bangalore R&D
> >>> >> >HUAWEI TECHNOLOGIES CO.,LTD.
> >>> >> >
> >>> >> >www.huawei.com
> >>> >> >-----------------------------------------------------------
> >>> >> ---------------
> >>> >> >-----------------------------------------------------------
> >>> >> >This e-mail and its attachments contain confidential information
> >>>from
> >>> >> >HUAWEI, which
> >>> >> >is intended only for the person or entity whose address is listed
> >>> >>above.
> >>> >> >Any use of the
> >>> >> >information contained herein in any way (including, but not limited
> >>> to,
> >>> >> >total or partial
> >>> >> >disclosure, reproduction, or dissemination) by persons other than
> >>>the
> >>> >> >intended
> >>> >> >recipient(s) is prohibited. If you receive this e-mail in error,
> >>> please
> >>> >> >notify the sender by
> >>> >> >phone or email immediately and delete it!
> >>> >> >
> >>> >> >________________________________________
> >>> >> >From: Shi, Shaofeng [[email protected]]
> >>> >> >Sent: Thursday, February 26, 2015 11:32 AM
> >>> >> >To: [email protected]
> >>> >> >Cc: Kulbhushan Rana
> >>> >> >Subject: Re: Error while making cube & Measure option is not
> >>> responding
> >>> >> >on GUI
> >>> >> >
> >>> >> >Hi Santosh, hive table importing issue should not impact on cube
> >>> >>saving.
> >>> >> >
> >>> >> >If you couldn’t save the cube, firstly please check whether there
> >>>is
> >>> >>error
> >>> >> >in the tomcat’s log; If not please check your web browser; We
> >>>suggest
> >>> >>use
> >>> >> >Firefox (with firebug add-on) or Chrome, open the JS console (press
> >>> >>F12)
> >>> >> >and then operate web UI, check whether there is any error reported
> >>>in
> >>> >> >browser.
> >>> >> >
> >>> >> >
> >>> >> >On 2/26/15, 1:08 PM, "Santoshakhilesh"
> >>><[email protected]>
> >>> >> >wrote:
> >>> >> >
> >>> >> >>Hi Shaofeng ,
> >>> >> >>   Thanks for replying.
> >>> >> >>   Yes I am checking the yarn exception, But I find that this
> >>>error
> >>> >>comes
> >>> >> >>while importing the hive table to kylin.
> >>> >> >>   Even if this error comes , hive tables is exported
> >>>successfully in
> >>> >> >>kylin. Is this the reason why cube saving has failed ?
> >>> >> >>
> >>> >> >>   Next step when I go on creating the cube for following schema
> >>>I
> >>> >>get
> >>> >> >>error at last step while saving and  I am unable to add any
> >>>measures
> >>> ,
> >>> >> >>clicking on measure option just dont pop up any dialog,
> >>> >> >>
> >>> >> >>I am using star schema with fact_sales as fact table and dim_* as
> >>> >> >>dimension tables.
> >>> >> >>
> >>> >> >> fact_sales:
> >>> >> >> storeid                 int
> >>> >> >> itemid                  int
> >>> >> >> custid                  int
> >>> >> >> qty                     int
> >>> >> >> price                   double
> >>> >> >>
> >>> >> >>dim_customer
> >>> >> >> custid                  int
> >>> >> >> name                    string
> >>> >> >>
> >>> >> >> dim_item
> >>> >> >> itemid                  int
> >>> >> >> category                string
> >>> >> >> brand                   string
> >>> >> >> color                   string
> >>> >> >>
> >>> >> >>dim_store
> >>> >> >> storeid                 int
> >>> >> >> city                    string
> >>> >> >> state                   string
> >>> >> >>
> >>> >> >>JSON is as below.
> >>> >> >> The JSON is as below.
> >>> >> >>
> >>> >> >> {
> >>> >> >>   "name": "Retail_Cube",
> >>> >> >>   "description": "",
> >>> >> >>   "dimensions": [
> >>> >> >>     {
> >>> >> >>       "name": "RETAIL.FACT_SALES.STOREID",
> >>> >> >>       "table": "RETAIL.FACT_SALES",
> >>> >> >>       "hierarchy": false,
> >>> >> >>       "derived": null,
> >>> >> >>       "column": [
> >>> >> >>         "STOREID"
> >>> >> >>       ],
> >>> >> >>       "id": 1
> >>> >> >>     },
> >>> >> >>     {
> >>> >> >>       "name": "RETAIL.FACT_SALES.ITEMID",
> >>> >> >>       "table": "RETAIL.FACT_SALES",
> >>> >> >>       "hierarchy": false,
> >>> >> >>       "derived": null,
> >>> >> >>       "column": [
> >>> >> >>         "ITEMID"
> >>> >> >>       ],
> >>> >> >>       "id": 2
> >>> >> >>     },
> >>> >> >>     {
> >>> >> >>       "name": "RETAIL.FACT_SALES.CUSTID",
> >>> >> >>       "table": "RETAIL.FACT_SALES",
> >>> >> >>       "hierarchy": false,
> >>> >> >>       "derived": null,
> >>> >> >>       "column": [
> >>> >> >>         "CUSTID"
> >>> >> >>       ],
> >>> >> >>       "id": 3
> >>> >> >>     }
> >>> >> >>   ],
> >>> >> >>   "measures": [
> >>> >> >>     {
> >>> >> >>       "id": 1,
> >>> >> >>       "name": "_COUNT_",
> >>> >> >>       "function": {
> >>> >> >>         "expression": "COUNT",
> >>> >> >>         "returntype": "bigint",
> >>> >> >>         "parameter": {
> >>> >> >>           "type": "constant",
> >>> >> >>           "value": 1
> >>> >> >>         }
> >>> >> >>       }
> >>> >> >>     }
> >>> >> >>   ],
> >>> >> >>   "rowkey": {
> >>> >> >>     "rowkey_columns": [
> >>> >> >>       {
> >>> >> >>         "column": "STOREID",
> >>> >> >>         "length": 0,
> >>> >> >>         "dictionary": "true",
> >>> >> >>         "mandatory": false
> >>> >> >>       },
> >>> >> >>       {
> >>> >> >>         "column": "ITEMID",
> >>> >> >>         "length": 0,
> >>> >> >>         "dictionary": "true",
> >>> >> >>         "mandatory": false
> >>> >> >>       },
> >>> >> >>       {
> >>> >> >>         "column": "CUSTID",
> >>> >> >>         "length": 0,
> >>> >> >>         "dictionary": "true",
> >>> >> >>         "mandatory": false
> >>> >> >>       }
> >>> >> >>     ],
> >>> >> >>     "aggregation_groups": [
> >>> >> >>       [
> >>> >> >>         "STOREID",
> >>> >> >>         "ITEMID",
> >>> >> >>         "CUSTID"
> >>> >> >>       ]
> >>> >> >>     ]
> >>> >> >>   },
> >>> >> >>   "notify_list": [],
> >>> >> >>   "capacity": "",
> >>> >> >>   "hbase_mapping": {
> >>> >> >>     "column_family": [
> >>> >> >>       {
> >>> >> >>         "name": "f1",
> >>> >> >>         "columns": [
> >>> >> >>           {
> >>> >> >>             "qualifier": "m",
> >>> >> >>             "measure_refs": [
> >>> >> >>               "_COUNT_"
> >>> >> >>             ]
> >>> >> >>           }
> >>> >> >>         ]
> >>> >> >>       }
> >>> >> >>     ]
> >>> >> >>   },
> >>> >> >>   "project": "RetailProject",
> >>> >> >>   "model_name": "Retail_Cube"
> >>> >> >> }
> >>> >> >>
> >>> >> >>Regards,
> >>> >> >>Santosh Akhilesh
> >>> >> >>Bangalore R&D
> >>> >> >>HUAWEI TECHNOLOGIES CO.,LTD.
> >>> >> >>
> >>> >> >>www.huawei.com
> >>> >> >>----------------------------------------------------------
> >>> >> ---------------
> >>> >> >>-
> >>> >> >>-----------------------------------------------------------
> >>> >> >>This e-mail and its attachments contain confidential information
> >>>from
> >>> >> >>HUAWEI, which
> >>> >> >>is intended only for the person or entity whose address is listed
> >>> >>above.
> >>> >> >>Any use of the
> >>> >> >>information contained herein in any way (including, but not
> >>>limited
> >>> >>to,
> >>> >> >>total or partial
> >>> >> >>disclosure, reproduction, or dissemination) by persons other than
> >>>the
> >>> >> >>intended
> >>> >> >>recipient(s) is prohibited. If you receive this e-mail in error,
> >>> >>please
> >>> >> >>notify the sender by
> >>> >> >>phone or email immediately and delete it!
> >>> >> >>
> >>> >> >>________________________________________
> >>> >> >>From: Shi, Shaofeng [[email protected]]
> >>> >> >>Sent: Thursday, February 26, 2015 7:01 AM
> >>> >> >>To: [email protected]
> >>> >> >>Subject: Re: Error while making cube & Measure option is not
> >>> >>responding
> >>> >> >>on GUI
> >>> >> >>
> >>> >> >>Hi Santosh,
> >>> >> >>
> >>> >> >>It looks like hadoop failed to execute some shell command in the
> >>> >> >>container; You need dive into hadoop to see what¹s the concrete
> >>> error.
> >>> >> >>You
> >>> >> >>can use yarn logs command to fetch all logs:
> >>> >> >>
> >>> >> >>yarn logs -applicationId <app_id>
> >>> >> >>
> >>> >> >>
> >>> >> >>On 2/25/15, 7:39 PM, "Santosh Akhilesh"
> >>><[email protected]>
> >>> >> >>wrote:
> >>> >> >>
> >>> >> >>>Hi Luke / Shaofeng ,
> >>> >> >>>           Can you please help me to check this issue.
> >>> >> >>>Regards,
> >>> >> >>>Santosh Akhilesh
> >>> >> >>>
> >>> >> >>>On Tue, Feb 24, 2015 at 10:41 PM, Santosh Akhilesh <
> >>> >> >>>[email protected]> wrote:
> >>> >> >>>
> >>> >> >>>> Hi All ,
> >>> >> >>>>         is it because of following error in map reduce job ?
> >>>what
> >>> >> >>>>could
> >>> >> >>>>be
> >>> >> >>>> way to resolve this , a google search says that its issue of
> >>>Yarn
> >>> >> >>>>class
> >>> >> >>>> path , but I am not sure what it is ?
> >>> >> >>>>
> >>> >> >>>> Kylin Hive Column Cardinality Job table=RETAIL.FACT_SALES
> >>> >> >>>> output=/tmp/cardinality/RETAIL.FACT_SALES
> >>> >> >>>>
> >>> >> >>>> Application application_1424791969399_0008 failed 2 times due
> >>>to
> >>> AM
> >>> >> >>>> Container for appattempt_1424791969399_0008_000002 exited with
> >>> >> >>>>exitCode: 1
> >>> >> >>>> For more detailed output, check application tracking page:
> >>> >> >>>> http://santosh:8088/proxy/application_1424791969399_0008/Then,
> >>> >>click
> >>> >> >>>>on
> >>> >> >>>> links to logs of each attempt.
> >>> >> >>>> Diagnostics: Exception from container-launch.
> >>> >> >>>> Container id: container_1424791969399_0008_02_000001
> >>> >> >>>> Exit code: 1
> >>> >> >>>> Stack trace: ExitCodeException exitCode=1:
> >>> >> >>>> at org.apache.hadoop.util.Shell.runCommand(Shell.java:538)
> >>> >> >>>> at org.apache.hadoop.util.Shell.run(Shell.java:455)
> >>> >> >>>> at
> >>> >> >>>>
> >>> >> >>>>org.apache.hadoop.util.Shell$ShellCommandExecutor.execut
> >>> >> e(Shell.java:71
> >>> >> >>>>5
> >>> >> >>>>)
> >>> >> >>>> at
> >>> >> >>>>
> >>> >> >>>>org.apache.hadoop.yarn.server.nodemanager.DefaultContain
> >>> >> erExecutor.laun
> >>> >> >>>>c
> >>> >> >>>>h
> >>> >> >>>>Container(DefaultContainerExecutor.java:211)
> >>> >> >>>> at
> >>> >> >>>>
> >>> >> >>>>org.apache.hadoop.yarn.server.nodemanager.containermanag
> >>> >> er.launcher.Con
> >>> >> >>>>t
> >>> >> >>>>a
> >>> >> >>>>inerLaunch.call(ContainerLaunch.java:302)
> >>> >> >>>> at
> >>> >> >>>>
> >>> >> >>>>org.apache.hadoop.yarn.server.nodemanager.containermanag
> >>> >> er.launcher.Con
> >>> >> >>>>t
> >>> >> >>>>a
> >>> >> >>>>inerLaunch.call(ContainerLaunch.java:82)
> >>> >> >>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262)
> >>> >> >>>> at
> >>> >> >>>>
> >>> >> >>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread
> >>> >> PoolExecutor.ja
> >>> >> >>>>v
> >>> >> >>>>a
> >>> >> >>>>:1145)
> >>> >> >>>> at
> >>> >> >>>>
> >>> >> >>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa
> >>> >> dPoolExecutor.j
> >>> >> >>>>a
> >>> >> >>>>v
> >>> >> >>>>a:615)
> >>> >> >>>> at java.lang.Thread.run(Thread.java:745)
> >>> >> >>>> Container exited with a non-zero exit code 1
> >>> >> >>>> Failing this attempt. Failing the application.
> >>> >> >>>>
> >>> >> >>>> ---------- Forwarded message ----------
> >>> >> >>>> From: Santoshakhilesh <[email protected]>
> >>> >> >>>> Date: Tue, Feb 24, 2015 at 7:41 PM
> >>> >> >>>> Subject: FW: Error while making cube & Measure option is not
> >>> >> >>>>responding
> >>> >> >>>>on
> >>> >> >>>> GUI
> >>> >> >>>> To: "[email protected]"
> >>> >><[email protected]>
> >>> >> >>>>
> >>> >> >>>>
> >>> >> >>>> hi ,
> >>> >> >>>>    please someone give me a hand to resolve this issue ,
> >>>thanks.
> >>> >> >>>>
> >>> >> >>>> Regards,
> >>> >> >>>> Santosh Akhilesh
> >>> >> >>>> Bangalore R&D
> >>> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD.
> >>> >> >>>>
> >>> >> >>>> www.huawei.com
> >>> >> >>>>
> >>> >> >>>>
> >>> >> >>>>--------------------------------------------------------
> >>> >> ---------------
> >>> >> >>>>-
> >>> >> >>>>-
> >>> >> >>>>------------------------------------------------------------
> >>> >> >>>> This e-mail and its attachments contain confidential
> >>>information
> >>> >>from
> >>> >> >>>> HUAWEI, which
> >>> >> >>>> is intended only for the person or entity whose address is
> >>>listed
> >>> >> >>>>above.
> >>> >> >>>> Any use of the
> >>> >> >>>> information contained herein in any way (including, but not
> >>> limited
> >>> >> >>>>to,
> >>> >> >>>> total or partial
> >>> >> >>>> disclosure, reproduction, or dissemination) by persons other
> >>>than
> >>> >>the
> >>> >> >>>> intended
> >>> >> >>>> recipient(s) is prohibited. If you receive this e-mail in
> >>>error,
> >>> >> >>>>please
> >>> >> >>>> notify the sender by
> >>> >> >>>> phone or email immediately and delete it!
> >>> >> >>>>
> >>> >> >>>> ________________________________________
> >>> >> >>>> From: Santoshakhilesh [[email protected]]
> >>> >> >>>> Sent: Tuesday, February 24, 2015 12:55 PM
> >>> >> >>>> To: [email protected]
> >>> >> >>>> Cc: Kulbhushan Rana
> >>> >> >>>> Subject: FW: Error while making cube & Measure option is not
> >>> >> >>>>responding
> >>> >> >>>>on
> >>> >> >>>> GUI
> >>> >> >>>>
> >>> >> >>>> 2. If  I ignore and continue and try to save the cube I get an
> >>> >> >>>>exception
> >>> >> >>>> in Kylin.log , I have checked the path is set correctly and
> >>> >> >>>>HCatInputFormat
> >>> >> >>>> this file is present in hive-hcatalog-core-0.14.0.jar . Please
> >>>let
> >>> >>me
> >>> >> >>>>know
> >>> >> >>>> what can I do to resolve this ?
> >>> >> >>>>
> >>> >> >>>>  -- This was path issue , now no more exception in kylin.log
> >>> >> >>>>
> >>> >> >>>> But saveing cube still fails with error. And still can't add
> >>> >>measures.
> >>> >> >>>>
> >>> >> >>>> Error Message
> >>> >> >>>> Failed to take action.
> >>> >> >>>>
> >>> >> >>>> In log I can find no exception. Following is the last log in
> >>> >>kylin.log
> >>> >> >>>>
> >>> >> >>>> [pool-3-thread-1]:[2015-02-24
> >>> >> >>>>
> >>> >> >>>>20:47:15,613][INFO][org.apache.kylin.job.impl.threadpool
> >>> >> .DefaultSchedul
> >>> >> >>>>e
> >>> >> >>>>r
> >>> >> >>>>$FetcherRunner.run(DefaultScheduler.java:117)]
> >>> >> >>>> - Job Fetcher: 0 running, 0 actual running, 0 ready, 6 others
> >>> >> >>>> [http-bio-7070-exec-2]:[2015-02-24
> >>> >> >>>>
> >>> >> >>>>20:47:51,610][DEBUG][org.apache.kylin.rest.controller.Cu
> >>> >> beController.de
> >>> >> >>>>s
> >>> >> >>>>e
> >>> >> >>>>rializeDataModelDesc(CubeController.java:459)]
> >>> >> >>>> - Saving cube {
> >>> >> >>>>   "name": "",
> >>> >> >>>>   "fact_table": "RETAIL.FACT_SALES",
> >>> >> >>>>   "lookups": [],
> >>> >> >>>>   "filter_condition": "",
> >>> >> >>>>   "capacity": "SMALL",
> >>> >> >>>>   "partition_desc": {
> >>> >> >>>>     "partition_date_column": "",
> >>> >> >>>>     "partition_date_start": 0,
> >>> >> >>>>     "partition_type": "APPEND"
> >>> >> >>>>   },
> >>> >> >>>>   "last_modified": 0
> >>> >> >>>> }
> >>> >> >>>>
> >>> >> >>>>
> >>> >> >>>> local access logs all with 200 , so seems ok.
> >>> >> >>>>
> >>> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:46:56 +0800] "GET
> >>> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
> >>> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:07 +0800] "GET
> >>> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
> >>> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:27 +0800] "GET
> >>> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
> >>> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:28 +0800] "GET
> >>> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
> >>> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:34 +0800] "GET
> >>> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
> >>> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:48 +0800] "GET
> >>> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
> >>> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:51 +0800] "POST
> >>> >>/kylin/api/cubes
> >>> >> >>>> HTTP/1.1" 200 701
> >>> >> >>>>
> >>> >> >>>>
> >>> >> >>>> Regards,
> >>> >> >>>> Santosh Akhilesh
> >>> >> >>>> Bangalore R&D
> >>> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD.
> >>> >> >>>>
> >>> >> >>>> www.huawei.com
> >>> >> >>>>
> >>> >> >>>>
> >>> >> >>>>--------------------------------------------------------
> >>> >> ---------------
> >>> >> >>>>-
> >>> >> >>>>-
> >>> >> >>>>------------------------------------------------------------
> >>> >> >>>> This e-mail and its attachments contain confidential
> >>>information
> >>> >>from
> >>> >> >>>> HUAWEI, which
> >>> >> >>>> is intended only for the person or entity whose address is
> >>>listed
> >>> >> >>>>above.
> >>> >> >>>> Any use of the
> >>> >> >>>> information contained herein in any way (including, but not
> >>> limited
> >>> >> >>>>to,
> >>> >> >>>> total or partial
> >>> >> >>>> disclosure, reproduction, or dissemination) by persons other
> >>>than
> >>> >>the
> >>> >> >>>> intended
> >>> >> >>>> recipient(s) is prohibited. If you receive this e-mail in
> >>>error,
> >>> >> >>>>please
> >>> >> >>>> notify the sender by
> >>> >> >>>> phone or email immediately and delete it!
> >>> >> >>>>
> >>> >> >>>> ________________________________________
> >>> >> >>>> From: Santoshakhilesh [[email protected]]
> >>> >> >>>> Sent: Tuesday, February 24, 2015 12:09 PM
> >>> >> >>>> To: [email protected]
> >>> >> >>>> Cc: Kulbhushan Rana
> >>> >> >>>> Subject: Error while making cube & Measure option is not
> >>> >>responding on
> >>> >> >>>>GUI
> >>> >> >>>>
> >>> >> >>>> Hi All ,
> >>> >> >>>>
> >>> >> >>>>     I am building a simple cube for test and using the binary
> >>> build
> >>> >> >>>>0.7.1
> >>> >> >>>> . I have following hive tables with columns.
> >>> >> >>>>
> >>> >> >>>>
> >>> >> >>>>
> >>> >> >>>> fact_sales:
> >>> >> >>>>
> >>> >> >>>> storeid                 int
> >>> >> >>>> itemid                  int
> >>> >> >>>> custid                  int
> >>> >> >>>> qty                     int
> >>> >> >>>> price                   double
> >>> >> >>>>
> >>> >> >>>> dim_customer
> >>> >> >>>> custid                  int
> >>> >> >>>> name                    string
> >>> >> >>>>
> >>> >> >>>> dim_item
> >>> >> >>>>
> >>> >> >>>> itemid                  int
> >>> >> >>>> category                string
> >>> >> >>>> brand                   string
> >>> >> >>>> color                   string
> >>> >> >>>>
> >>> >> >>>> dim_store
> >>> >> >>>>
> >>> >> >>>> storeid                 int
> >>> >> >>>> city                    string
> >>> >> >>>> state                   string
> >>> >> >>>>
> >>> >> >>>> Please help me to answer following issues;
> >>> >> >>>>
> >>> >> >>>>
> >>> >> >>>>
> >>> >> >>>> 1. When I go to measure section and click on measure option ,
> >>> >>there is
> >>> >> >>>>no
> >>> >> >>>> response , I want add measure on qty and price with sum
> >>> >> >>>>
> >>> >> >>>> 2. If  I ignore and continue and try to save the cube I get an
> >>> >> >>>>exception
> >>> >> >>>> in Kylin.log , I have checked the path is set correctly and
> >>> >> >>>>HCatInputFormat
> >>> >> >>>> this file is present in hive-hcatalog-core-0.14.0.jar . Please
> >>>let
> >>> >>me
> >>> >> >>>>know
> >>> >> >>>> what can I do to resolve this ?
> >>> >> >>>>
> >>> >> >>>> 3. Also I have another question since this is a test and data
> >>>is
> >>> >>small
> >>> >> >>>>I
> >>> >> >>>> have not partitioned the fact table , is it ok to skip
> >>>partition
> >>> >>stage
> >>> >> >>>> while cube build ?
> >>> >> >>>>
> >>> >> >>>>
> >>> >> >>>>
> >>> >> >>>> Exception
> >>> >> >>>>
> >>> >> >>>> pool-4-thread-4]:[2015-02-24
> >>> >> >>>>
> >>> >> >>>>19:26:32,577][ERROR][org.apache.kylin.job.impl.threadpoo
> >>> >> l.DefaultSchedu
> >>> >> >>>>l
> >>> >> >>>>e
> >>> >> >>>>r$JobRunner.run(DefaultScheduler.java:134)]
> >>> >> >>>> - ExecuteException job:c3532a6f-97ea-474a-b36a-218dd517cedb
> >>> >> >>>> org.apache.kylin.job.exception.ExecuteException:
> >>> >> >>>> org.apache.kylin.job.exception.ExecuteException:
> >>> >> >>>> java.lang.NoClassDefFoundError:
> >>> >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat
> >>> >> >>>>  at
> >>> >> >>>>
> >>> >> >>>>org.apache.kylin.job.execution.AbstractExecutable.execut
> >>> >> e(AbstractExecu
> >>> >> >>>>t
> >>> >> >>>>a
> >>> >> >>>>ble.java:102)
> >>> >> >>>>  at
> >>> >> >>>>
> >>> >> >>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$Jo
> >>> >> bRunner.run(Def
> >>> >> >>>>a
> >>> >> >>>>u
> >>> >> >>>>ltScheduler.java:132)
> >>> >> >>>>  at
> >>> >> >>>>
> >>> >> >>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread
> >>> >> PoolExecutor.ja
> >>> >> >>>>v
> >>> >> >>>>a
> >>> >> >>>>:1145)
> >>> >> >>>>  at
> >>> >> >>>>
> >>> >> >>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa
> >>> >> dPoolExecutor.j
> >>> >> >>>>a
> >>> >> >>>>v
> >>> >> >>>>a:615)
> >>> >> >>>>  at java.lang.Thread.run(Thread.java:745)
> >>> >> >>>> Caused by: org.apache.kylin.job.exception.ExecuteException:
> >>> >> >>>> java.lang.NoClassDefFoundError:
> >>> >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat
> >>> >> >>>>
> >>> >> >>>>
> >>> >> >>>>
> >>> >> >>>> The JSON is as below.
> >>> >> >>>>
> >>> >> >>>>
> >>> >> >>>>
> >>> >> >>>> {
> >>> >> >>>>   "name": "Retail_Cube",
> >>> >> >>>>   "description": "",
> >>> >> >>>>   "dimensions": [
> >>> >> >>>>     {
> >>> >> >>>>       "name": "RETAIL.FACT_SALES.STOREID",
> >>> >> >>>>       "table": "RETAIL.FACT_SALES",
> >>> >> >>>>       "hierarchy": false,
> >>> >> >>>>       "derived": null,
> >>> >> >>>>       "column": [
> >>> >> >>>>         "STOREID"
> >>> >> >>>>       ],
> >>> >> >>>>       "id": 1
> >>> >> >>>>     },
> >>> >> >>>>     {
> >>> >> >>>>       "name": "RETAIL.FACT_SALES.ITEMID",
> >>> >> >>>>       "table": "RETAIL.FACT_SALES",
> >>> >> >>>>       "hierarchy": false,
> >>> >> >>>>       "derived": null,
> >>> >> >>>>       "column": [
> >>> >> >>>>         "ITEMID"
> >>> >> >>>>       ],
> >>> >> >>>>       "id": 2
> >>> >> >>>>     },
> >>> >> >>>>     {
> >>> >> >>>>       "name": "RETAIL.FACT_SALES.CUSTID",
> >>> >> >>>>       "table": "RETAIL.FACT_SALES",
> >>> >> >>>>       "hierarchy": false,
> >>> >> >>>>       "derived": null,
> >>> >> >>>>       "column": [
> >>> >> >>>>         "CUSTID"
> >>> >> >>>>       ],
> >>> >> >>>>       "id": 3
> >>> >> >>>>     }
> >>> >> >>>>   ],
> >>> >> >>>>   "measures": [
> >>> >> >>>>     {
> >>> >> >>>>       "id": 1,
> >>> >> >>>>       "name": "_COUNT_",
> >>> >> >>>>       "function": {
> >>> >> >>>>         "expression": "COUNT",
> >>> >> >>>>         "returntype": "bigint",
> >>> >> >>>>         "parameter": {
> >>> >> >>>>           "type": "constant",
> >>> >> >>>>           "value": 1
> >>> >> >>>>         }
> >>> >> >>>>       }
> >>> >> >>>>     }
> >>> >> >>>>   ],
> >>> >> >>>>   "rowkey": {
> >>> >> >>>>     "rowkey_columns": [
> >>> >> >>>>       {
> >>> >> >>>>         "column": "STOREID",
> >>> >> >>>>         "length": 0,
> >>> >> >>>>         "dictionary": "true",
> >>> >> >>>>         "mandatory": false
> >>> >> >>>>       },
> >>> >> >>>>       {
> >>> >> >>>>         "column": "ITEMID",
> >>> >> >>>>         "length": 0,
> >>> >> >>>>         "dictionary": "true",
> >>> >> >>>>         "mandatory": false
> >>> >> >>>>       },
> >>> >> >>>>       {
> >>> >> >>>>         "column": "CUSTID",
> >>> >> >>>>         "length": 0,
> >>> >> >>>>         "dictionary": "true",
> >>> >> >>>>         "mandatory": false
> >>> >> >>>>       }
> >>> >> >>>>     ],
> >>> >> >>>>     "aggregation_groups": [
> >>> >> >>>>       [
> >>> >> >>>>         "STOREID",
> >>> >> >>>>         "ITEMID",
> >>> >> >>>>         "CUSTID"
> >>> >> >>>>       ]
> >>> >> >>>>     ]
> >>> >> >>>>   },
> >>> >> >>>>   "notify_list": [],
> >>> >> >>>>   "capacity": "",
> >>> >> >>>>   "hbase_mapping": {
> >>> >> >>>>     "column_family": [
> >>> >> >>>>       {
> >>> >> >>>>         "name": "f1",
> >>> >> >>>>         "columns": [
> >>> >> >>>>           {
> >>> >> >>>>             "qualifier": "m",
> >>> >> >>>>             "measure_refs": [
> >>> >> >>>>               "_COUNT_"
> >>> >> >>>>             ]
> >>> >> >>>>           }
> >>> >> >>>>         ]
> >>> >> >>>>       }
> >>> >> >>>>     ]
> >>> >> >>>>   },
> >>> >> >>>>   "project": "RetailProject",
> >>> >> >>>>   "model_name": "Retail_Cube"
> >>> >> >>>> }
> >>> >> >>>>
> >>> >> >>>>
> >>> >> >>>>
> >>> >> >>>>
> >>> >> >>>>
> >>> >> >>>> Regards,
> >>> >> >>>> Santosh Akhilesh
> >>> >> >>>> Bangalore R&D
> >>> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD.
> >>> >> >>>>
> >>> >> >>>> www.huawei.com
> >>> >> >>>>
> >>> >> >>>>
> >>> >> >>>>--------------------------------------------------------
> >>> >> ---------------
> >>> >> >>>>-
> >>> >> >>>>-
> >>> >> >>>>------------------------------------------------------------
> >>> >> >>>> This e-mail and its attachments contain confidential
> >>>information
> >>> >>from
> >>> >> >>>> HUAWEI, which
> >>> >> >>>> is intended only for the person or entity whose address is
> >>>listed
> >>> >> >>>>above.
> >>> >> >>>> Any use of the
> >>> >> >>>> information contained herein in any way (including, but not
> >>> limited
> >>> >> >>>>to,
> >>> >> >>>> total or partial
> >>> >> >>>> disclosure, reproduction, or dissemination) by persons other
> >>>than
> >>> >>the
> >>> >> >>>> intended
> >>> >> >>>> recipient(s) is prohibited. If you receive this e-mail in
> >>>error,
> >>> >> >>>>please
> >>> >> >>>> notify the sender by
> >>> >> >>>> phone or email immediately and delete it!
> >>> >> >>>>
> >>> >> >>>>
> >>> >> >>>>
> >>> >> >>>> --
> >>> >> >>>> Regards,
> >>> >> >>>> Santosh Akhilesh
> >>> >> >>>> +91-0-9845482201
> >>> >> >>>>
> >>> >> >>>
> >>> >> >>>
> >>> >> >>>
> >>> >> >>>--
> >>> >> >>>Regards,
> >>> >> >>>Santosh Akhilesh
> >>> >> >>>+91-0-9845482201
> >>> >>
> >>> >>
> >>>
> >>>
>
>


-- 
Regards,
Santosh Akhilesh
+91-0-9845482201

Reply via email to