For the REST API, please refer to these two docs in kylin’s wiki:

https://github.com/KylinOLAP/Kylin/wiki/Build-Cube-with-Restful-API

https://github.com/KylinOLAP/Kylin/wiki/Restful-Service-Doc


For the sample code, there is a RestClient.java in Kylin code base:

https://github.com/KylinOLAP/Kylin/blob/master/common/src/main/java/com/kyl
inolap/common/restclient/RestClient.java

In it there is only 1 action which is “wipeCache”, but it can be a sample
for how to invoking the REST services from Java.


On 2/26/15, 7:03 PM, "Santosh Akhilesh" <[email protected]> wrote:

>Hi Saofeng,
>Can I use CLI or RESTI interface for 0.7.1?
>Any sample code I can refer?
>When can we expect a working Web interface?
>Regards
>Santosh Akhilesh
>On Thu, 26 Feb 2015 at 3:09 pm, Santosh Akhilesh
><[email protected]>
>wrote:
>
>> Hi Shaofeng
>> So what do you suggest , how should I proceed further with this release?
>> Will there be a patch? Any alternate way I can create cube?
>> Please suggest.
>> Regards
>> Santosh
>> On Thu, 26 Feb 2015 at 3:04 pm, Shi, Shaofeng <[email protected]> wrote:
>>
>>> Hi Santosh,
>>>
>>> 0.7.1 hasn’t been formally released; from 0.6.x to 0.7.x we have
>>>metadata
>>> structure change; While, the web UI (cube wizard) for this change
>>>hasn’t
>>> been stabilized; So it is not strange that you got trouble when saving
>>>the
>>> cube;
>>>
>>> @Jason, any idea about the JS error?
>>>
>>> On 2/26/15, 5:08 PM, "Santosh Akhilesh" <[email protected]>
>>> wrote:
>>>
>>> >Hi Shaofeng,
>>> >
>>> >I am using the binary distribution 0.7.1. I have not been able to save
>>> >cube
>>> >even once. I have tried creating new project and from local machine
>>>and
>>> >server machine. But I am always stuck with this error. I am never
>>>allowed
>>> >to add measures and never been able to save the cube. I also see the
>>> >kylin.log and it always tries to save cube with append mode. One
>>>thing I
>>> >need to tell that at partition stage since I don't have a big fact
>>>table
>>> >now I have not partititioned the fact table and I skip this step. Does
>>> >this
>>> >have affect in saving the cube. Is this because some metadata is
>>> available
>>> >and it tries to modify the cube? I am using latest Hadoop 2.6.6. Yes
>>> kylin
>>> >propert I have not added the jar. I will add them and check. But cube
>>> >creation failure is really puzzling me. I could see no error logs in
>>> >kylin.log.
>>> >Regards
>>> >Santosh
>>> >On Thu, 26 Feb 2015 at 1:40 pm, Shi, Shaofeng <[email protected]>
>>>wrote:
>>> >
>>> >> Which version or code branch are you using? I assume you’re using
>>>the
>>> >> stable version from master; Seems you’re trying to edit an existing
>>> cube
>>> >> to add new measurement, try refresh your browser's cache; if it
>>>still
>>> >> couldn’t be saved, try to create a new cube;
>>> >>
>>> >> The two error traces in tomcat need be taken care:
>>> >>
>>> >> 1) java.lang.NoClassDefFoundError:
>>> >>org/apache/kylin/common/mr/KylinMapper
>>> >>         Please check kylin.properties file, making sure the
>>> >>“kylin.job.jar”
>>> >> points to a right jar file; It will be loaded in Map-reduce;
>>> >>
>>> >> 2) java.lang.IllegalArgumentException: No enum constant
>>> >> org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS
>>> >>         This indicates your hadoop version might be old; Please
>>>check
>>> >>and
>>> >> ensure
>>> >> hadoop version is 2.2 or above.
>>> >>
>>> >> On 2/26/15, 3:21 PM, "Santoshakhilesh" <[email protected]>
>>> >> wrote:
>>> >>
>>> >> >Hi Shaofeng ,
>>> >> >
>>> >> >   I am using chrome , When I click on button to add measures ,
>>> >>following
>>> >> >is error on chrome console. When I try to save the cube there is no
>>> >>error
>>> >> >in console. I just get a error dialog saying failed to take action
>>>and
>>> >> >gives me the JSON cube schema.
>>> >> >
>>> >> >Error on chrome debug console is as below;
>>> >> >
>>> >> > ReferenceError: CubeDescModel is not defined
>>> >> >    at h.$scope.addNewMeasure (scripts.min.0.js:15984)
>>> >> >    at scripts.min.0.js:180
>>> >> >    at scripts.min.0.js:197
>>> >> >    at h.$eval (scripts.min.0.js:119)
>>> >> >    at h.$apply (scripts.min.0.js:119)
>>> >> >    at HTMLButtonElement.<anonymous> (scripts.min.0.js:197)
>>> >> >    at HTMLButtonElement.m.event.dispatch (scripts.min.0.js:3)
>>> >> >    at HTMLButtonElement.r.handle
>>> >> >(scripts.min.0.js:3)scripts.min.0.js:100 (anonymous
>>> >> >function)scripts.min.0.js:77 (anonymous
>>>function)scripts.min.0.js:119
>>> >> >h.$applyscripts.min.0.js:197 (anonymous function)scripts.min.0.js:3
>>> >> >m.event.dispatchscripts.min.0.js:3 r.handle
>>> >> >
>>> >> >   About the hive table import , I got pass the run shell command
>>> >> >exception but it still fails the hadoop log is;
>>> >> >2015-02-26 20:46:48,332 INFO [main] org.apache.hadoop.mapred.YarnC
>>> hild:
>>> >> >mapreduce.cluster.local.dir for child:
>>> >> >/tmp/hadoop-root/nm-local-dir/usercache/root/appcache/appli
>>> >> cation_14249530
>>> >> >91340_0002
>>> >> >2015-02-26 20:46:48,776 INFO [main]
>>> >> >org.apache.hadoop.conf.Configuration.deprecation: session.id is
>>> >> >deprecated. Instead, use dfs.metrics.session-id
>>> >> >2015-02-26 20:46:49,310 INFO [main] org.apache.hadoop.mapred.Task:
>>> >>Using
>>> >> >ResourceCalculatorProcessTree : [ ]
>>> >> >2015-02-26 20:46:49,386 FATAL [main]
>>> >>org.apache.hadoop.mapred.YarnChild:
>>> >> >Error running child : java.lang.NoClassDefFoundError:
>>> >> >org/apache/kylin/common/mr/KylinMapper
>>> >> > at java.lang.ClassLoader.defineClass1(Native Method)
>>> >> > at java.lang.ClassLoader.defineClass(ClassLoader.java:800)
>>> >> > at
>>> >> >java.security.SecureClassLoader.defineClass(SecureClassLoad
>>> er.java:142)
>>> >> > at java.net.URLClassLoader.defineClass(URLClassLoader.java:449)
>>> >> > at java.net.URLClassLoader.access$100(URLClassLoader.java:71)
>>> >> > at java.net.URLClassLoader$1.run(URLClassLoader.java:361)
>>> >> > at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
>>> >> > at java.security.AccessController.doPrivileged(Native Method)
>>> >> > at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
>>> >> > at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
>>> >> > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
>>> >> >
>>> >> >tomcat logs:
>>> >> >usage: HiveColumnCardinalityJob
>>> >> > -output <path>        Output path
>>> >> > -table <table name>   The hive table name
>>> >> >[pool-4-thread-2]:[2015-02-26
>>> >> >20:47:49,936][ERROR][org.apache.kylin.job.common.HadoopShel
>>> >> lExecutable.doW
>>> >> >ork(HadoopShellExecutable.java:64)] - error execute
>>> >> >HadoopShellExecutable{id=d4730d26-7fe6-412e-9841-3288ab362c5b-00,
>>> >> >name=null, state=RUNNING}
>>> >> >java.lang.IllegalArgumentException: No enum constant
>>> >> >org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS
>>> >> > at java.lang.Enum.valueOf(Enum.java:236)
>>> >> > at
>>> >> >org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.valueOf(
>>> >> Framewo
>>> >> >rkCounterGroup.java:148)
>>> >> > at
>>> >> >org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.
>>> >> findCounter(Fra
>>> >> >meworkCounterGroup.java:182)
>>> >> > at
>>> >> >org.apache.hadoop.mapreduce.counters.AbstractCounters.findC
>>> >> ounter(Abstract
>>> >> >Counters.java:154)
>>> >> > at
>>> >> >org.apache.hadoop.mapreduce.TypeConverter.fromYarn(TypeConv
>>> >> erter.java:240)
>>> >> > at
>>> >> >org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounte
>>> >> rs(ClientServic
>>> >> >eDelegate.java:370)
>>> >> > at
>>> >> >org.apache.hadoop.mapred.YARNRunner.getJobCounters(YARNRunn
>>> er.java:511)
>>> >> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756)
>>> >> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753)
>>> >> > at java.security.AccessController.doPrivileged(Native Method)
>>> >> > at javax.security.auth.Subject.doAs(Subject.java:415)
>>> >> > at
>>> >> >org.apache.hadoop.security.UserGroupInformation.doAs(UserGr
>>> >> oupInformation.
>>> >> >java:1491)
>>> >> > at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753)
>>> >> > at 
>>>org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1361)
>>> >> > at 
>>>org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1289)
>>> >> > at
>>> >> >org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForComple
>>> >> tion(AbstractHa
>>> >> >doopJob.java:134)
>>> >> > at
>>> >> 
>>>>org.apache.kylin.job.hadoop.cardinality.HiveColumnCardinalityJob.run(
>>> >> HiveC
>>> >> >olumnCardinalityJob.java:114)
>>> >> > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>>> >> > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
>>> >> > at
>>> >> >org.apache.kylin.job.common.HadoopShellExecutable.doWork(Ha
>>> >> doopShellExecut
>>> >> >able.java:62)
>>> >> > at
>>> >> >org.apache.kylin.job.execution.AbstractExecutable.execute(A
>>> >> bstractExecutab
>>> >> >le.java:99)
>>> >> > at
>>> >> >org.apache.kylin.job.execution.DefaultChainedExecutable.doW
>>> >> ork(DefaultChai
>>> >> >nedExecutable.java:50)
>>> >> > at
>>> >> >org.apache.kylin.job.execution.AbstractExecutable.execute(A
>>> >> bstractExecutab
>>> >> >le.java:99)
>>> >> > at
>>> >> >org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRu
>>> >> nner.run(Defaul
>>> >> >tScheduler.java:132)
>>> >> > at
>>> >> >java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoo
>>> >> lExecutor.java:
>>> >> >1145)
>>> >> > at
>>> >> >java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPo
>>> >> olExecutor.java
>>> >> >:615)
>>> >> > at java.lang.Thread.run(Thread.java:745)
>>> >> >
>>> >> >Regards,
>>> >> >Santosh Akhilesh
>>> >> >Bangalore R&D
>>> >> >HUAWEI TECHNOLOGIES CO.,LTD.
>>> >> >
>>> >> >www.huawei.com
>>> >> >-----------------------------------------------------------
>>> >> ---------------
>>> >> >-----------------------------------------------------------
>>> >> >This e-mail and its attachments contain confidential information
>>>from
>>> >> >HUAWEI, which
>>> >> >is intended only for the person or entity whose address is listed
>>> >>above.
>>> >> >Any use of the
>>> >> >information contained herein in any way (including, but not limited
>>> to,
>>> >> >total or partial
>>> >> >disclosure, reproduction, or dissemination) by persons other than
>>>the
>>> >> >intended
>>> >> >recipient(s) is prohibited. If you receive this e-mail in error,
>>> please
>>> >> >notify the sender by
>>> >> >phone or email immediately and delete it!
>>> >> >
>>> >> >________________________________________
>>> >> >From: Shi, Shaofeng [[email protected]]
>>> >> >Sent: Thursday, February 26, 2015 11:32 AM
>>> >> >To: [email protected]
>>> >> >Cc: Kulbhushan Rana
>>> >> >Subject: Re: Error while making cube & Measure option is not
>>> responding
>>> >> >on GUI
>>> >> >
>>> >> >Hi Santosh, hive table importing issue should not impact on cube
>>> >>saving.
>>> >> >
>>> >> >If you couldn’t save the cube, firstly please check whether there
>>>is
>>> >>error
>>> >> >in the tomcat’s log; If not please check your web browser; We
>>>suggest
>>> >>use
>>> >> >Firefox (with firebug add-on) or Chrome, open the JS console (press
>>> >>F12)
>>> >> >and then operate web UI, check whether there is any error reported
>>>in
>>> >> >browser.
>>> >> >
>>> >> >
>>> >> >On 2/26/15, 1:08 PM, "Santoshakhilesh"
>>><[email protected]>
>>> >> >wrote:
>>> >> >
>>> >> >>Hi Shaofeng ,
>>> >> >>   Thanks for replying.
>>> >> >>   Yes I am checking the yarn exception, But I find that this
>>>error
>>> >>comes
>>> >> >>while importing the hive table to kylin.
>>> >> >>   Even if this error comes , hive tables is exported
>>>successfully in
>>> >> >>kylin. Is this the reason why cube saving has failed ?
>>> >> >>
>>> >> >>   Next step when I go on creating the cube for following schema
>>>I
>>> >>get
>>> >> >>error at last step while saving and  I am unable to add any
>>>measures
>>> ,
>>> >> >>clicking on measure option just dont pop up any dialog,
>>> >> >>
>>> >> >>I am using star schema with fact_sales as fact table and dim_* as
>>> >> >>dimension tables.
>>> >> >>
>>> >> >> fact_sales:
>>> >> >> storeid                 int
>>> >> >> itemid                  int
>>> >> >> custid                  int
>>> >> >> qty                     int
>>> >> >> price                   double
>>> >> >>
>>> >> >>dim_customer
>>> >> >> custid                  int
>>> >> >> name                    string
>>> >> >>
>>> >> >> dim_item
>>> >> >> itemid                  int
>>> >> >> category                string
>>> >> >> brand                   string
>>> >> >> color                   string
>>> >> >>
>>> >> >>dim_store
>>> >> >> storeid                 int
>>> >> >> city                    string
>>> >> >> state                   string
>>> >> >>
>>> >> >>JSON is as below.
>>> >> >> The JSON is as below.
>>> >> >>
>>> >> >> {
>>> >> >>   "name": "Retail_Cube",
>>> >> >>   "description": "",
>>> >> >>   "dimensions": [
>>> >> >>     {
>>> >> >>       "name": "RETAIL.FACT_SALES.STOREID",
>>> >> >>       "table": "RETAIL.FACT_SALES",
>>> >> >>       "hierarchy": false,
>>> >> >>       "derived": null,
>>> >> >>       "column": [
>>> >> >>         "STOREID"
>>> >> >>       ],
>>> >> >>       "id": 1
>>> >> >>     },
>>> >> >>     {
>>> >> >>       "name": "RETAIL.FACT_SALES.ITEMID",
>>> >> >>       "table": "RETAIL.FACT_SALES",
>>> >> >>       "hierarchy": false,
>>> >> >>       "derived": null,
>>> >> >>       "column": [
>>> >> >>         "ITEMID"
>>> >> >>       ],
>>> >> >>       "id": 2
>>> >> >>     },
>>> >> >>     {
>>> >> >>       "name": "RETAIL.FACT_SALES.CUSTID",
>>> >> >>       "table": "RETAIL.FACT_SALES",
>>> >> >>       "hierarchy": false,
>>> >> >>       "derived": null,
>>> >> >>       "column": [
>>> >> >>         "CUSTID"
>>> >> >>       ],
>>> >> >>       "id": 3
>>> >> >>     }
>>> >> >>   ],
>>> >> >>   "measures": [
>>> >> >>     {
>>> >> >>       "id": 1,
>>> >> >>       "name": "_COUNT_",
>>> >> >>       "function": {
>>> >> >>         "expression": "COUNT",
>>> >> >>         "returntype": "bigint",
>>> >> >>         "parameter": {
>>> >> >>           "type": "constant",
>>> >> >>           "value": 1
>>> >> >>         }
>>> >> >>       }
>>> >> >>     }
>>> >> >>   ],
>>> >> >>   "rowkey": {
>>> >> >>     "rowkey_columns": [
>>> >> >>       {
>>> >> >>         "column": "STOREID",
>>> >> >>         "length": 0,
>>> >> >>         "dictionary": "true",
>>> >> >>         "mandatory": false
>>> >> >>       },
>>> >> >>       {
>>> >> >>         "column": "ITEMID",
>>> >> >>         "length": 0,
>>> >> >>         "dictionary": "true",
>>> >> >>         "mandatory": false
>>> >> >>       },
>>> >> >>       {
>>> >> >>         "column": "CUSTID",
>>> >> >>         "length": 0,
>>> >> >>         "dictionary": "true",
>>> >> >>         "mandatory": false
>>> >> >>       }
>>> >> >>     ],
>>> >> >>     "aggregation_groups": [
>>> >> >>       [
>>> >> >>         "STOREID",
>>> >> >>         "ITEMID",
>>> >> >>         "CUSTID"
>>> >> >>       ]
>>> >> >>     ]
>>> >> >>   },
>>> >> >>   "notify_list": [],
>>> >> >>   "capacity": "",
>>> >> >>   "hbase_mapping": {
>>> >> >>     "column_family": [
>>> >> >>       {
>>> >> >>         "name": "f1",
>>> >> >>         "columns": [
>>> >> >>           {
>>> >> >>             "qualifier": "m",
>>> >> >>             "measure_refs": [
>>> >> >>               "_COUNT_"
>>> >> >>             ]
>>> >> >>           }
>>> >> >>         ]
>>> >> >>       }
>>> >> >>     ]
>>> >> >>   },
>>> >> >>   "project": "RetailProject",
>>> >> >>   "model_name": "Retail_Cube"
>>> >> >> }
>>> >> >>
>>> >> >>Regards,
>>> >> >>Santosh Akhilesh
>>> >> >>Bangalore R&D
>>> >> >>HUAWEI TECHNOLOGIES CO.,LTD.
>>> >> >>
>>> >> >>www.huawei.com
>>> >> >>----------------------------------------------------------
>>> >> ---------------
>>> >> >>-
>>> >> >>-----------------------------------------------------------
>>> >> >>This e-mail and its attachments contain confidential information
>>>from
>>> >> >>HUAWEI, which
>>> >> >>is intended only for the person or entity whose address is listed
>>> >>above.
>>> >> >>Any use of the
>>> >> >>information contained herein in any way (including, but not
>>>limited
>>> >>to,
>>> >> >>total or partial
>>> >> >>disclosure, reproduction, or dissemination) by persons other than
>>>the
>>> >> >>intended
>>> >> >>recipient(s) is prohibited. If you receive this e-mail in error,
>>> >>please
>>> >> >>notify the sender by
>>> >> >>phone or email immediately and delete it!
>>> >> >>
>>> >> >>________________________________________
>>> >> >>From: Shi, Shaofeng [[email protected]]
>>> >> >>Sent: Thursday, February 26, 2015 7:01 AM
>>> >> >>To: [email protected]
>>> >> >>Subject: Re: Error while making cube & Measure option is not
>>> >>responding
>>> >> >>on GUI
>>> >> >>
>>> >> >>Hi Santosh,
>>> >> >>
>>> >> >>It looks like hadoop failed to execute some shell command in the
>>> >> >>container; You need dive into hadoop to see what¹s the concrete
>>> error.
>>> >> >>You
>>> >> >>can use yarn logs command to fetch all logs:
>>> >> >>
>>> >> >>yarn logs -applicationId <app_id>
>>> >> >>
>>> >> >>
>>> >> >>On 2/25/15, 7:39 PM, "Santosh Akhilesh"
>>><[email protected]>
>>> >> >>wrote:
>>> >> >>
>>> >> >>>Hi Luke / Shaofeng ,
>>> >> >>>           Can you please help me to check this issue.
>>> >> >>>Regards,
>>> >> >>>Santosh Akhilesh
>>> >> >>>
>>> >> >>>On Tue, Feb 24, 2015 at 10:41 PM, Santosh Akhilesh <
>>> >> >>>[email protected]> wrote:
>>> >> >>>
>>> >> >>>> Hi All ,
>>> >> >>>>         is it because of following error in map reduce job ?
>>>what
>>> >> >>>>could
>>> >> >>>>be
>>> >> >>>> way to resolve this , a google search says that its issue of
>>>Yarn
>>> >> >>>>class
>>> >> >>>> path , but I am not sure what it is ?
>>> >> >>>>
>>> >> >>>> Kylin Hive Column Cardinality Job table=RETAIL.FACT_SALES
>>> >> >>>> output=/tmp/cardinality/RETAIL.FACT_SALES
>>> >> >>>>
>>> >> >>>> Application application_1424791969399_0008 failed 2 times due
>>>to
>>> AM
>>> >> >>>> Container for appattempt_1424791969399_0008_000002 exited with
>>> >> >>>>exitCode: 1
>>> >> >>>> For more detailed output, check application tracking page:
>>> >> >>>> http://santosh:8088/proxy/application_1424791969399_0008/Then,
>>> >>click
>>> >> >>>>on
>>> >> >>>> links to logs of each attempt.
>>> >> >>>> Diagnostics: Exception from container-launch.
>>> >> >>>> Container id: container_1424791969399_0008_02_000001
>>> >> >>>> Exit code: 1
>>> >> >>>> Stack trace: ExitCodeException exitCode=1:
>>> >> >>>> at org.apache.hadoop.util.Shell.runCommand(Shell.java:538)
>>> >> >>>> at org.apache.hadoop.util.Shell.run(Shell.java:455)
>>> >> >>>> at
>>> >> >>>>
>>> >> >>>>org.apache.hadoop.util.Shell$ShellCommandExecutor.execut
>>> >> e(Shell.java:71
>>> >> >>>>5
>>> >> >>>>)
>>> >> >>>> at
>>> >> >>>>
>>> >> >>>>org.apache.hadoop.yarn.server.nodemanager.DefaultContain
>>> >> erExecutor.laun
>>> >> >>>>c
>>> >> >>>>h
>>> >> >>>>Container(DefaultContainerExecutor.java:211)
>>> >> >>>> at
>>> >> >>>>
>>> >> >>>>org.apache.hadoop.yarn.server.nodemanager.containermanag
>>> >> er.launcher.Con
>>> >> >>>>t
>>> >> >>>>a
>>> >> >>>>inerLaunch.call(ContainerLaunch.java:302)
>>> >> >>>> at
>>> >> >>>>
>>> >> >>>>org.apache.hadoop.yarn.server.nodemanager.containermanag
>>> >> er.launcher.Con
>>> >> >>>>t
>>> >> >>>>a
>>> >> >>>>inerLaunch.call(ContainerLaunch.java:82)
>>> >> >>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>>> >> >>>> at
>>> >> >>>>
>>> >> >>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread
>>> >> PoolExecutor.ja
>>> >> >>>>v
>>> >> >>>>a
>>> >> >>>>:1145)
>>> >> >>>> at
>>> >> >>>>
>>> >> >>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa
>>> >> dPoolExecutor.j
>>> >> >>>>a
>>> >> >>>>v
>>> >> >>>>a:615)
>>> >> >>>> at java.lang.Thread.run(Thread.java:745)
>>> >> >>>> Container exited with a non-zero exit code 1
>>> >> >>>> Failing this attempt. Failing the application.
>>> >> >>>>
>>> >> >>>> ---------- Forwarded message ----------
>>> >> >>>> From: Santoshakhilesh <[email protected]>
>>> >> >>>> Date: Tue, Feb 24, 2015 at 7:41 PM
>>> >> >>>> Subject: FW: Error while making cube & Measure option is not
>>> >> >>>>responding
>>> >> >>>>on
>>> >> >>>> GUI
>>> >> >>>> To: "[email protected]"
>>> >><[email protected]>
>>> >> >>>>
>>> >> >>>>
>>> >> >>>> hi ,
>>> >> >>>>    please someone give me a hand to resolve this issue ,
>>>thanks.
>>> >> >>>>
>>> >> >>>> Regards,
>>> >> >>>> Santosh Akhilesh
>>> >> >>>> Bangalore R&D
>>> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD.
>>> >> >>>>
>>> >> >>>> www.huawei.com
>>> >> >>>>
>>> >> >>>>
>>> >> >>>>--------------------------------------------------------
>>> >> ---------------
>>> >> >>>>-
>>> >> >>>>-
>>> >> >>>>------------------------------------------------------------
>>> >> >>>> This e-mail and its attachments contain confidential
>>>information
>>> >>from
>>> >> >>>> HUAWEI, which
>>> >> >>>> is intended only for the person or entity whose address is
>>>listed
>>> >> >>>>above.
>>> >> >>>> Any use of the
>>> >> >>>> information contained herein in any way (including, but not
>>> limited
>>> >> >>>>to,
>>> >> >>>> total or partial
>>> >> >>>> disclosure, reproduction, or dissemination) by persons other
>>>than
>>> >>the
>>> >> >>>> intended
>>> >> >>>> recipient(s) is prohibited. If you receive this e-mail in
>>>error,
>>> >> >>>>please
>>> >> >>>> notify the sender by
>>> >> >>>> phone or email immediately and delete it!
>>> >> >>>>
>>> >> >>>> ________________________________________
>>> >> >>>> From: Santoshakhilesh [[email protected]]
>>> >> >>>> Sent: Tuesday, February 24, 2015 12:55 PM
>>> >> >>>> To: [email protected]
>>> >> >>>> Cc: Kulbhushan Rana
>>> >> >>>> Subject: FW: Error while making cube & Measure option is not
>>> >> >>>>responding
>>> >> >>>>on
>>> >> >>>> GUI
>>> >> >>>>
>>> >> >>>> 2. If  I ignore and continue and try to save the cube I get an
>>> >> >>>>exception
>>> >> >>>> in Kylin.log , I have checked the path is set correctly and
>>> >> >>>>HCatInputFormat
>>> >> >>>> this file is present in hive-hcatalog-core-0.14.0.jar . Please
>>>let
>>> >>me
>>> >> >>>>know
>>> >> >>>> what can I do to resolve this ?
>>> >> >>>>
>>> >> >>>>  -- This was path issue , now no more exception in kylin.log
>>> >> >>>>
>>> >> >>>> But saveing cube still fails with error. And still can't add
>>> >>measures.
>>> >> >>>>
>>> >> >>>> Error Message
>>> >> >>>> Failed to take action.
>>> >> >>>>
>>> >> >>>> In log I can find no exception. Following is the last log in
>>> >>kylin.log
>>> >> >>>>
>>> >> >>>> [pool-3-thread-1]:[2015-02-24
>>> >> >>>>
>>> >> >>>>20:47:15,613][INFO][org.apache.kylin.job.impl.threadpool
>>> >> .DefaultSchedul
>>> >> >>>>e
>>> >> >>>>r
>>> >> >>>>$FetcherRunner.run(DefaultScheduler.java:117)]
>>> >> >>>> - Job Fetcher: 0 running, 0 actual running, 0 ready, 6 others
>>> >> >>>> [http-bio-7070-exec-2]:[2015-02-24
>>> >> >>>>
>>> >> >>>>20:47:51,610][DEBUG][org.apache.kylin.rest.controller.Cu
>>> >> beController.de
>>> >> >>>>s
>>> >> >>>>e
>>> >> >>>>rializeDataModelDesc(CubeController.java:459)]
>>> >> >>>> - Saving cube {
>>> >> >>>>   "name": "",
>>> >> >>>>   "fact_table": "RETAIL.FACT_SALES",
>>> >> >>>>   "lookups": [],
>>> >> >>>>   "filter_condition": "",
>>> >> >>>>   "capacity": "SMALL",
>>> >> >>>>   "partition_desc": {
>>> >> >>>>     "partition_date_column": "",
>>> >> >>>>     "partition_date_start": 0,
>>> >> >>>>     "partition_type": "APPEND"
>>> >> >>>>   },
>>> >> >>>>   "last_modified": 0
>>> >> >>>> }
>>> >> >>>>
>>> >> >>>>
>>> >> >>>> local access logs all with 200 , so seems ok.
>>> >> >>>>
>>> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:46:56 +0800] "GET
>>> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:07 +0800] "GET
>>> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:27 +0800] "GET
>>> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:28 +0800] "GET
>>> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:34 +0800] "GET
>>> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:48 +0800] "GET
>>> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:51 +0800] "POST
>>> >>/kylin/api/cubes
>>> >> >>>> HTTP/1.1" 200 701
>>> >> >>>>
>>> >> >>>>
>>> >> >>>> Regards,
>>> >> >>>> Santosh Akhilesh
>>> >> >>>> Bangalore R&D
>>> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD.
>>> >> >>>>
>>> >> >>>> www.huawei.com
>>> >> >>>>
>>> >> >>>>
>>> >> >>>>--------------------------------------------------------
>>> >> ---------------
>>> >> >>>>-
>>> >> >>>>-
>>> >> >>>>------------------------------------------------------------
>>> >> >>>> This e-mail and its attachments contain confidential
>>>information
>>> >>from
>>> >> >>>> HUAWEI, which
>>> >> >>>> is intended only for the person or entity whose address is
>>>listed
>>> >> >>>>above.
>>> >> >>>> Any use of the
>>> >> >>>> information contained herein in any way (including, but not
>>> limited
>>> >> >>>>to,
>>> >> >>>> total or partial
>>> >> >>>> disclosure, reproduction, or dissemination) by persons other
>>>than
>>> >>the
>>> >> >>>> intended
>>> >> >>>> recipient(s) is prohibited. If you receive this e-mail in
>>>error,
>>> >> >>>>please
>>> >> >>>> notify the sender by
>>> >> >>>> phone or email immediately and delete it!
>>> >> >>>>
>>> >> >>>> ________________________________________
>>> >> >>>> From: Santoshakhilesh [[email protected]]
>>> >> >>>> Sent: Tuesday, February 24, 2015 12:09 PM
>>> >> >>>> To: [email protected]
>>> >> >>>> Cc: Kulbhushan Rana
>>> >> >>>> Subject: Error while making cube & Measure option is not
>>> >>responding on
>>> >> >>>>GUI
>>> >> >>>>
>>> >> >>>> Hi All ,
>>> >> >>>>
>>> >> >>>>     I am building a simple cube for test and using the binary
>>> build
>>> >> >>>>0.7.1
>>> >> >>>> . I have following hive tables with columns.
>>> >> >>>>
>>> >> >>>>
>>> >> >>>>
>>> >> >>>> fact_sales:
>>> >> >>>>
>>> >> >>>> storeid                 int
>>> >> >>>> itemid                  int
>>> >> >>>> custid                  int
>>> >> >>>> qty                     int
>>> >> >>>> price                   double
>>> >> >>>>
>>> >> >>>> dim_customer
>>> >> >>>> custid                  int
>>> >> >>>> name                    string
>>> >> >>>>
>>> >> >>>> dim_item
>>> >> >>>>
>>> >> >>>> itemid                  int
>>> >> >>>> category                string
>>> >> >>>> brand                   string
>>> >> >>>> color                   string
>>> >> >>>>
>>> >> >>>> dim_store
>>> >> >>>>
>>> >> >>>> storeid                 int
>>> >> >>>> city                    string
>>> >> >>>> state                   string
>>> >> >>>>
>>> >> >>>> Please help me to answer following issues;
>>> >> >>>>
>>> >> >>>>
>>> >> >>>>
>>> >> >>>> 1. When I go to measure section and click on measure option ,
>>> >>there is
>>> >> >>>>no
>>> >> >>>> response , I want add measure on qty and price with sum
>>> >> >>>>
>>> >> >>>> 2. If  I ignore and continue and try to save the cube I get an
>>> >> >>>>exception
>>> >> >>>> in Kylin.log , I have checked the path is set correctly and
>>> >> >>>>HCatInputFormat
>>> >> >>>> this file is present in hive-hcatalog-core-0.14.0.jar . Please
>>>let
>>> >>me
>>> >> >>>>know
>>> >> >>>> what can I do to resolve this ?
>>> >> >>>>
>>> >> >>>> 3. Also I have another question since this is a test and data
>>>is
>>> >>small
>>> >> >>>>I
>>> >> >>>> have not partitioned the fact table , is it ok to skip
>>>partition
>>> >>stage
>>> >> >>>> while cube build ?
>>> >> >>>>
>>> >> >>>>
>>> >> >>>>
>>> >> >>>> Exception
>>> >> >>>>
>>> >> >>>> pool-4-thread-4]:[2015-02-24
>>> >> >>>>
>>> >> >>>>19:26:32,577][ERROR][org.apache.kylin.job.impl.threadpoo
>>> >> l.DefaultSchedu
>>> >> >>>>l
>>> >> >>>>e
>>> >> >>>>r$JobRunner.run(DefaultScheduler.java:134)]
>>> >> >>>> - ExecuteException job:c3532a6f-97ea-474a-b36a-218dd517cedb
>>> >> >>>> org.apache.kylin.job.exception.ExecuteException:
>>> >> >>>> org.apache.kylin.job.exception.ExecuteException:
>>> >> >>>> java.lang.NoClassDefFoundError:
>>> >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat
>>> >> >>>>  at
>>> >> >>>>
>>> >> >>>>org.apache.kylin.job.execution.AbstractExecutable.execut
>>> >> e(AbstractExecu
>>> >> >>>>t
>>> >> >>>>a
>>> >> >>>>ble.java:102)
>>> >> >>>>  at
>>> >> >>>>
>>> >> >>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$Jo
>>> >> bRunner.run(Def
>>> >> >>>>a
>>> >> >>>>u
>>> >> >>>>ltScheduler.java:132)
>>> >> >>>>  at
>>> >> >>>>
>>> >> >>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread
>>> >> PoolExecutor.ja
>>> >> >>>>v
>>> >> >>>>a
>>> >> >>>>:1145)
>>> >> >>>>  at
>>> >> >>>>
>>> >> >>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa
>>> >> dPoolExecutor.j
>>> >> >>>>a
>>> >> >>>>v
>>> >> >>>>a:615)
>>> >> >>>>  at java.lang.Thread.run(Thread.java:745)
>>> >> >>>> Caused by: org.apache.kylin.job.exception.ExecuteException:
>>> >> >>>> java.lang.NoClassDefFoundError:
>>> >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat
>>> >> >>>>
>>> >> >>>>
>>> >> >>>>
>>> >> >>>> The JSON is as below.
>>> >> >>>>
>>> >> >>>>
>>> >> >>>>
>>> >> >>>> {
>>> >> >>>>   "name": "Retail_Cube",
>>> >> >>>>   "description": "",
>>> >> >>>>   "dimensions": [
>>> >> >>>>     {
>>> >> >>>>       "name": "RETAIL.FACT_SALES.STOREID",
>>> >> >>>>       "table": "RETAIL.FACT_SALES",
>>> >> >>>>       "hierarchy": false,
>>> >> >>>>       "derived": null,
>>> >> >>>>       "column": [
>>> >> >>>>         "STOREID"
>>> >> >>>>       ],
>>> >> >>>>       "id": 1
>>> >> >>>>     },
>>> >> >>>>     {
>>> >> >>>>       "name": "RETAIL.FACT_SALES.ITEMID",
>>> >> >>>>       "table": "RETAIL.FACT_SALES",
>>> >> >>>>       "hierarchy": false,
>>> >> >>>>       "derived": null,
>>> >> >>>>       "column": [
>>> >> >>>>         "ITEMID"
>>> >> >>>>       ],
>>> >> >>>>       "id": 2
>>> >> >>>>     },
>>> >> >>>>     {
>>> >> >>>>       "name": "RETAIL.FACT_SALES.CUSTID",
>>> >> >>>>       "table": "RETAIL.FACT_SALES",
>>> >> >>>>       "hierarchy": false,
>>> >> >>>>       "derived": null,
>>> >> >>>>       "column": [
>>> >> >>>>         "CUSTID"
>>> >> >>>>       ],
>>> >> >>>>       "id": 3
>>> >> >>>>     }
>>> >> >>>>   ],
>>> >> >>>>   "measures": [
>>> >> >>>>     {
>>> >> >>>>       "id": 1,
>>> >> >>>>       "name": "_COUNT_",
>>> >> >>>>       "function": {
>>> >> >>>>         "expression": "COUNT",
>>> >> >>>>         "returntype": "bigint",
>>> >> >>>>         "parameter": {
>>> >> >>>>           "type": "constant",
>>> >> >>>>           "value": 1
>>> >> >>>>         }
>>> >> >>>>       }
>>> >> >>>>     }
>>> >> >>>>   ],
>>> >> >>>>   "rowkey": {
>>> >> >>>>     "rowkey_columns": [
>>> >> >>>>       {
>>> >> >>>>         "column": "STOREID",
>>> >> >>>>         "length": 0,
>>> >> >>>>         "dictionary": "true",
>>> >> >>>>         "mandatory": false
>>> >> >>>>       },
>>> >> >>>>       {
>>> >> >>>>         "column": "ITEMID",
>>> >> >>>>         "length": 0,
>>> >> >>>>         "dictionary": "true",
>>> >> >>>>         "mandatory": false
>>> >> >>>>       },
>>> >> >>>>       {
>>> >> >>>>         "column": "CUSTID",
>>> >> >>>>         "length": 0,
>>> >> >>>>         "dictionary": "true",
>>> >> >>>>         "mandatory": false
>>> >> >>>>       }
>>> >> >>>>     ],
>>> >> >>>>     "aggregation_groups": [
>>> >> >>>>       [
>>> >> >>>>         "STOREID",
>>> >> >>>>         "ITEMID",
>>> >> >>>>         "CUSTID"
>>> >> >>>>       ]
>>> >> >>>>     ]
>>> >> >>>>   },
>>> >> >>>>   "notify_list": [],
>>> >> >>>>   "capacity": "",
>>> >> >>>>   "hbase_mapping": {
>>> >> >>>>     "column_family": [
>>> >> >>>>       {
>>> >> >>>>         "name": "f1",
>>> >> >>>>         "columns": [
>>> >> >>>>           {
>>> >> >>>>             "qualifier": "m",
>>> >> >>>>             "measure_refs": [
>>> >> >>>>               "_COUNT_"
>>> >> >>>>             ]
>>> >> >>>>           }
>>> >> >>>>         ]
>>> >> >>>>       }
>>> >> >>>>     ]
>>> >> >>>>   },
>>> >> >>>>   "project": "RetailProject",
>>> >> >>>>   "model_name": "Retail_Cube"
>>> >> >>>> }
>>> >> >>>>
>>> >> >>>>
>>> >> >>>>
>>> >> >>>>
>>> >> >>>>
>>> >> >>>> Regards,
>>> >> >>>> Santosh Akhilesh
>>> >> >>>> Bangalore R&D
>>> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD.
>>> >> >>>>
>>> >> >>>> www.huawei.com
>>> >> >>>>
>>> >> >>>>
>>> >> >>>>--------------------------------------------------------
>>> >> ---------------
>>> >> >>>>-
>>> >> >>>>-
>>> >> >>>>------------------------------------------------------------
>>> >> >>>> This e-mail and its attachments contain confidential 
>>>information
>>> >>from
>>> >> >>>> HUAWEI, which
>>> >> >>>> is intended only for the person or entity whose address is 
>>>listed
>>> >> >>>>above.
>>> >> >>>> Any use of the
>>> >> >>>> information contained herein in any way (including, but not
>>> limited
>>> >> >>>>to,
>>> >> >>>> total or partial
>>> >> >>>> disclosure, reproduction, or dissemination) by persons other 
>>>than
>>> >>the
>>> >> >>>> intended
>>> >> >>>> recipient(s) is prohibited. If you receive this e-mail in 
>>>error,
>>> >> >>>>please
>>> >> >>>> notify the sender by
>>> >> >>>> phone or email immediately and delete it!
>>> >> >>>>
>>> >> >>>>
>>> >> >>>>
>>> >> >>>> --
>>> >> >>>> Regards,
>>> >> >>>> Santosh Akhilesh
>>> >> >>>> +91-0-9845482201
>>> >> >>>>
>>> >> >>>
>>> >> >>>
>>> >> >>>
>>> >> >>>--
>>> >> >>>Regards,
>>> >> >>>Santosh Akhilesh
>>> >> >>>+91-0-9845482201
>>> >>
>>> >>
>>>
>>>

Reply via email to