The 0.7.1 is test version, its package contains the “snapshot” suffix; we
will upload a new package there; Luke will also add a message there to
avoid this confusion;

Regarding the problem that you encountered, could you please open a JIRA
ticket for tracking? Here is link of Apache JIRA:

https://issues.apache.org/jira/secure/Dashboard.jspa


Thanks for the feedback!

On 2/26/15, 10:21 PM, "Santosh Akhilesh" <[email protected]> wrote:

>Actually I see this being published on kylin webpage.
>http://kylin.incubator.apache.org/download/
>I am using 0.7.1 inverted index branch binary distribution.
>If this is not stable please give me the link of stable branch I would try
>building and testing tonight.
>On Thu, 26 Feb 2015 at 7:30 pm, Shi, Shaofeng <[email protected]> wrote:
>
>> Hi Santosh, it is not recommended to use the dev code branch (actually I
>> don’t know how you get the v0.7.x build and what’s the detail version of
>> that; each day we submit many changes to that);
>>
>> The options are 1) switch back to latest release v0.6.5; or 2) wait for
>> the formal release of 0.7, that should be in March; Otherwise, we
>>couldn’t
>> ensure there is no new problems come out in your next steps;
>>
>> On 2/26/15, 5:39 PM, "Santosh Akhilesh" <[email protected]>
>>wrote:
>>
>> >Hi Shaofeng
>> >So what do you suggest , how should I proceed further with this
>>release?
>> >Will there be a patch? Any alternate way I can create cube?
>> >Please suggest.
>> >Regards
>> >Santosh
>> >On Thu, 26 Feb 2015 at 3:04 pm, Shi, Shaofeng <[email protected]> wrote:
>> >
>> >> Hi Santosh,
>> >>
>> >> 0.7.1 hasn’t been formally released; from 0.6.x to 0.7.x we have
>> >>metadata
>> >> structure change; While, the web UI (cube wizard) for this change
>>hasn’t
>> >> been stabilized; So it is not strange that you got trouble when
>>saving
>> >>the
>> >> cube;
>> >>
>> >> @Jason, any idea about the JS error?
>> >>
>> >> On 2/26/15, 5:08 PM, "Santosh Akhilesh" <[email protected]>
>> >>wrote:
>> >>
>> >> >Hi Shaofeng,
>> >> >
>> >> >I am using the binary distribution 0.7.1. I have not been able to
>>save
>> >> >cube
>> >> >even once. I have tried creating new project and from local machine
>>and
>> >> >server machine. But I am always stuck with this error. I am never
>> >>allowed
>> >> >to add measures and never been able to save the cube. I also see the
>> >> >kylin.log and it always tries to save cube with append mode. One
>>thing
>> >>I
>> >> >need to tell that at partition stage since I don't have a big fact
>> >>table
>> >> >now I have not partititioned the fact table and I skip this step.
>>Does
>> >> >this
>> >> >have affect in saving the cube. Is this because some metadata is
>> >>available
>> >> >and it tries to modify the cube? I am using latest Hadoop 2.6.6. Yes
>> >>kylin
>> >> >propert I have not added the jar. I will add them and check. But
>>cube
>> >> >creation failure is really puzzling me. I could see no error logs in
>> >> >kylin.log.
>> >> >Regards
>> >> >Santosh
>> >> >On Thu, 26 Feb 2015 at 1:40 pm, Shi, Shaofeng <[email protected]>
>> wrote:
>> >> >
>> >> >> Which version or code branch are you using? I assume you’re using
>>the
>> >> >> stable version from master; Seems you’re trying to edit an
>>existing
>> >>cube
>> >> >> to add new measurement, try refresh your browser's cache; if it
>>still
>> >> >> couldn’t be saved, try to create a new cube;
>> >> >>
>> >> >> The two error traces in tomcat need be taken care:
>> >> >>
>> >> >> 1) java.lang.NoClassDefFoundError:
>> >> >>org/apache/kylin/common/mr/KylinMapper
>> >> >>         Please check kylin.properties file, making sure the
>> >> >>“kylin.job.jar”
>> >> >> points to a right jar file; It will be loaded in Map-reduce;
>> >> >>
>> >> >> 2) java.lang.IllegalArgumentException: No enum constant
>> >> >> org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS
>> >> >>         This indicates your hadoop version might be old; Please
>>check
>> >> >>and
>> >> >> ensure
>> >> >> hadoop version is 2.2 or above.
>> >> >>
>> >> >> On 2/26/15, 3:21 PM, "Santoshakhilesh"
>><[email protected]>
>> >> >> wrote:
>> >> >>
>> >> >> >Hi Shaofeng ,
>> >> >> >
>> >> >> >   I am using chrome , When I click on button to add measures ,
>> >> >>following
>> >> >> >is error on chrome console. When I try to save the cube there is
>>no
>> >> >>error
>> >> >> >in console. I just get a error dialog saying failed to take
>>action
>> >>and
>> >> >> >gives me the JSON cube schema.
>> >> >> >
>> >> >> >Error on chrome debug console is as below;
>> >> >> >
>> >> >> > ReferenceError: CubeDescModel is not defined
>> >> >> >    at h.$scope.addNewMeasure (scripts.min.0.js:15984)
>> >> >> >    at scripts.min.0.js:180
>> >> >> >    at scripts.min.0.js:197
>> >> >> >    at h.$eval (scripts.min.0.js:119)
>> >> >> >    at h.$apply (scripts.min.0.js:119)
>> >> >> >    at HTMLButtonElement.<anonymous> (scripts.min.0.js:197)
>> >> >> >    at HTMLButtonElement.m.event.dispatch (scripts.min.0.js:3)
>> >> >> >    at HTMLButtonElement.r.handle
>> >> >> >(scripts.min.0.js:3)scripts.min.0.js:100 (anonymous
>> >> >> >function)scripts.min.0.js:77 (anonymous
>> >>function)scripts.min.0.js:119
>> >> >> >h.$applyscripts.min.0.js:197 (anonymous
>>function)scripts.min.0.js:3
>> >> >> >m.event.dispatchscripts.min.0.js:3 r.handle
>> >> >> >
>> >> >> >   About the hive table import , I got pass the run shell command
>> >> >> >exception but it still fails the hadoop log is;
>> >> >> >2015-02-26 20:46:48,332 INFO [main] org.apache.hadoop.mapred.
>> >> YarnChild:
>> >> >> >mapreduce.cluster.local.dir for child:
>> >> >> >/tmp/hadoop-root/nm-local-dir/usercache/root/appcache/appli
>> >> >> cation_14249530
>> >> >> >91340_0002
>> >> >> >2015-02-26 20:46:48,776 INFO [main]
>> >> >> >org.apache.hadoop.conf.Configuration.deprecation: session.id is
>> >> >> >deprecated. Instead, use dfs.metrics.session-id
>> >> >> >2015-02-26 20:46:49,310 INFO [main]
>>org.apache.hadoop.mapred.Task:
>> >> >>Using
>> >> >> >ResourceCalculatorProcessTree : [ ]
>> >> >> >2015-02-26 20:46:49,386 FATAL [main]
>> >> >>org.apache.hadoop.mapred.YarnChild:
>> >> >> >Error running child : java.lang.NoClassDefFoundError:
>> >> >> >org/apache/kylin/common/mr/KylinMapper
>> >> >> > at java.lang.ClassLoader.defineClass1(Native Method)
>> >> >> > at java.lang.ClassLoader.defineClass(ClassLoader.java:800)
>> >> >> > at
>> >> >> >java.security.SecureClassLoader.defineClass(
>> >> SecureClassLoader.java:142)
>> >> >> > at java.net.URLClassLoader.defineClass(URLClassLoader.java:449)
>> >> >> > at java.net.URLClassLoader.access$100(URLClassLoader.java:71)
>> >> >> > at java.net.URLClassLoader$1.run(URLClassLoader.java:361)
>> >> >> > at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
>> >> >> > at java.security.AccessController.doPrivileged(Native Method)
>> >> >> > at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
>> >> >> > at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
>> >> >> > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
>> >> >> >
>> >> >> >tomcat logs:
>> >> >> >usage: HiveColumnCardinalityJob
>> >> >> > -output <path>        Output path
>> >> >> > -table <table name>   The hive table name
>> >> >> >[pool-4-thread-2]:[2015-02-26
>> >> >> >20:47:49,936][ERROR][org.apache.kylin.job.common.HadoopShel
>> >> >> lExecutable.doW
>> >> >> >ork(HadoopShellExecutable.java:64)] - error execute
>> >> >> >HadoopShellExecutable{id=d4730d26-7fe6-412e-9841-3288ab362c5b-00,
>> >> >> >name=null, state=RUNNING}
>> >> >> >java.lang.IllegalArgumentException: No enum constant
>> >> >> >org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS
>> >> >> > at java.lang.Enum.valueOf(Enum.java:236)
>> >> >> > at
>> >> >> 
>>>org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.valueOf(
>> >> >> Framewo
>> >> >> >rkCounterGroup.java:148)
>> >> >> > at
>> >> >> >org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.
>> >> >> findCounter(Fra
>> >> >> >meworkCounterGroup.java:182)
>> >> >> > at
>> >> >> >org.apache.hadoop.mapreduce.counters.AbstractCounters.findC
>> >> >> ounter(Abstract
>> >> >> >Counters.java:154)
>> >> >> > at
>> >> >> >org.apache.hadoop.mapreduce.TypeConverter.fromYarn(TypeConv
>> >> >> erter.java:240)
>> >> >> > at
>> >> >> >org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounte
>> >> >> rs(ClientServic
>> >> >> >eDelegate.java:370)
>> >> >> > at
>> >> >> >org.apache.hadoop.mapred.YARNRunner.getJobCounters(
>> >> YARNRunner.java:511)
>> >> >> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756)
>> >> >> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753)
>> >> >> > at java.security.AccessController.doPrivileged(Native Method)
>> >> >> > at javax.security.auth.Subject.doAs(Subject.java:415)
>> >> >> > at
>> >> >> >org.apache.hadoop.security.UserGroupInformation.doAs(UserGr
>> >> >> oupInformation.
>> >> >> >java:1491)
>> >> >> > at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753)
>> >> >> > at
>> >>org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1361)
>> >> >> > at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.
>> java:1289)
>> >> >> > at
>> >> >> >org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForComple
>> >> >> tion(AbstractHa
>> >> >> >doopJob.java:134)
>> >> >> > at
>> >> >>
>> >>>org.apache.kylin.job.hadoop.cardinality.HiveColumnCardinalityJob.run(
>> >> >> HiveC
>> >> >> >olumnCardinalityJob.java:114)
>> >> >> > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>> >> >> > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
>> >> >> > at
>> >> >> >org.apache.kylin.job.common.HadoopShellExecutable.doWork(Ha
>> >> >> doopShellExecut
>> >> >> >able.java:62)
>> >> >> > at
>> >> >> >org.apache.kylin.job.execution.AbstractExecutable.execute(A
>> >> >> bstractExecutab
>> >> >> >le.java:99)
>> >> >> > at
>> >> >> >org.apache.kylin.job.execution.DefaultChainedExecutable.doW
>> >> >> ork(DefaultChai
>> >> >> >nedExecutable.java:50)
>> >> >> > at
>> >> >> >org.apache.kylin.job.execution.AbstractExecutable.execute(A
>> >> >> bstractExecutab
>> >> >> >le.java:99)
>> >> >> > at
>> >> >> >org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRu
>> >> >> nner.run(Defaul
>> >> >> >tScheduler.java:132)
>> >> >> > at
>> >> >> >java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoo
>> >> >> lExecutor.java:
>> >> >> >1145)
>> >> >> > at
>> >> >> >java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPo
>> >> >> olExecutor.java
>> >> >> >:615)
>> >> >> > at java.lang.Thread.run(Thread.java:745)
>> >> >> >
>> >> >> >Regards,
>> >> >> >Santosh Akhilesh
>> >> >> >Bangalore R&D
>> >> >> >HUAWEI TECHNOLOGIES CO.,LTD.
>> >> >> >
>> >> >> >www.huawei.com
>> >> >> >-----------------------------------------------------------
>> >> >> ---------------
>> >> >> >-----------------------------------------------------------
>> >> >> >This e-mail and its attachments contain confidential information
>> >>from
>> >> >> >HUAWEI, which
>> >> >> >is intended only for the person or entity whose address is listed
>> >> >>above.
>> >> >> >Any use of the
>> >> >> >information contained herein in any way (including, but not
>>limited
>> >>to,
>> >> >> >total or partial
>> >> >> >disclosure, reproduction, or dissemination) by persons other than
>> >>the
>> >> >> >intended
>> >> >> >recipient(s) is prohibited. If you receive this e-mail in error,
>> >>please
>> >> >> >notify the sender by
>> >> >> >phone or email immediately and delete it!
>> >> >> >
>> >> >> >________________________________________
>> >> >> >From: Shi, Shaofeng [[email protected]]
>> >> >> >Sent: Thursday, February 26, 2015 11:32 AM
>> >> >> >To: [email protected]
>> >> >> >Cc: Kulbhushan Rana
>> >> >> >Subject: Re: Error while making cube & Measure option is not
>> >>responding
>> >> >> >on GUI
>> >> >> >
>> >> >> >Hi Santosh, hive table importing issue should not impact on cube
>> >> >>saving.
>> >> >> >
>> >> >> >If you couldn’t save the cube, firstly please check whether
>>there is
>> >> >>error
>> >> >> >in the tomcat’s log; If not please check your web browser; We
>> >>suggest
>> >> >>use
>> >> >> >Firefox (with firebug add-on) or Chrome, open the JS console
>>(press
>> >> >>F12)
>> >> >> >and then operate web UI, check whether there is any error
>>reported
>> >>in
>> >> >> >browser.
>> >> >> >
>> >> >> >
>> >> >> >On 2/26/15, 1:08 PM, "Santoshakhilesh"
>><[email protected]
>> >
>> >> >> >wrote:
>> >> >> >
>> >> >> >>Hi Shaofeng ,
>> >> >> >>   Thanks for replying.
>> >> >> >>   Yes I am checking the yarn exception, But I find that this
>>error
>> >> >>comes
>> >> >> >>while importing the hive table to kylin.
>> >> >> >>   Even if this error comes , hive tables is exported
>>successfully
>> >>in
>> >> >> >>kylin. Is this the reason why cube saving has failed ?
>> >> >> >>
>> >> >> >>   Next step when I go on creating the cube for following
>>schema  I
>> >> >>get
>> >> >> >>error at last step while saving and  I am unable to add any
>> >>measures ,
>> >> >> >>clicking on measure option just dont pop up any dialog,
>> >> >> >>
>> >> >> >>I am using star schema with fact_sales as fact table and dim_*
>>as
>> >> >> >>dimension tables.
>> >> >> >>
>> >> >> >> fact_sales:
>> >> >> >> storeid                 int
>> >> >> >> itemid                  int
>> >> >> >> custid                  int
>> >> >> >> qty                     int
>> >> >> >> price                   double
>> >> >> >>
>> >> >> >>dim_customer
>> >> >> >> custid                  int
>> >> >> >> name                    string
>> >> >> >>
>> >> >> >> dim_item
>> >> >> >> itemid                  int
>> >> >> >> category                string
>> >> >> >> brand                   string
>> >> >> >> color                   string
>> >> >> >>
>> >> >> >>dim_store
>> >> >> >> storeid                 int
>> >> >> >> city                    string
>> >> >> >> state                   string
>> >> >> >>
>> >> >> >>JSON is as below.
>> >> >> >> The JSON is as below.
>> >> >> >>
>> >> >> >> {
>> >> >> >>   "name": "Retail_Cube",
>> >> >> >>   "description": "",
>> >> >> >>   "dimensions": [
>> >> >> >>     {
>> >> >> >>       "name": "RETAIL.FACT_SALES.STOREID",
>> >> >> >>       "table": "RETAIL.FACT_SALES",
>> >> >> >>       "hierarchy": false,
>> >> >> >>       "derived": null,
>> >> >> >>       "column": [
>> >> >> >>         "STOREID"
>> >> >> >>       ],
>> >> >> >>       "id": 1
>> >> >> >>     },
>> >> >> >>     {
>> >> >> >>       "name": "RETAIL.FACT_SALES.ITEMID",
>> >> >> >>       "table": "RETAIL.FACT_SALES",
>> >> >> >>       "hierarchy": false,
>> >> >> >>       "derived": null,
>> >> >> >>       "column": [
>> >> >> >>         "ITEMID"
>> >> >> >>       ],
>> >> >> >>       "id": 2
>> >> >> >>     },
>> >> >> >>     {
>> >> >> >>       "name": "RETAIL.FACT_SALES.CUSTID",
>> >> >> >>       "table": "RETAIL.FACT_SALES",
>> >> >> >>       "hierarchy": false,
>> >> >> >>       "derived": null,
>> >> >> >>       "column": [
>> >> >> >>         "CUSTID"
>> >> >> >>       ],
>> >> >> >>       "id": 3
>> >> >> >>     }
>> >> >> >>   ],
>> >> >> >>   "measures": [
>> >> >> >>     {
>> >> >> >>       "id": 1,
>> >> >> >>       "name": "_COUNT_",
>> >> >> >>       "function": {
>> >> >> >>         "expression": "COUNT",
>> >> >> >>         "returntype": "bigint",
>> >> >> >>         "parameter": {
>> >> >> >>           "type": "constant",
>> >> >> >>           "value": 1
>> >> >> >>         }
>> >> >> >>       }
>> >> >> >>     }
>> >> >> >>   ],
>> >> >> >>   "rowkey": {
>> >> >> >>     "rowkey_columns": [
>> >> >> >>       {
>> >> >> >>         "column": "STOREID",
>> >> >> >>         "length": 0,
>> >> >> >>         "dictionary": "true",
>> >> >> >>         "mandatory": false
>> >> >> >>       },
>> >> >> >>       {
>> >> >> >>         "column": "ITEMID",
>> >> >> >>         "length": 0,
>> >> >> >>         "dictionary": "true",
>> >> >> >>         "mandatory": false
>> >> >> >>       },
>> >> >> >>       {
>> >> >> >>         "column": "CUSTID",
>> >> >> >>         "length": 0,
>> >> >> >>         "dictionary": "true",
>> >> >> >>         "mandatory": false
>> >> >> >>       }
>> >> >> >>     ],
>> >> >> >>     "aggregation_groups": [
>> >> >> >>       [
>> >> >> >>         "STOREID",
>> >> >> >>         "ITEMID",
>> >> >> >>         "CUSTID"
>> >> >> >>       ]
>> >> >> >>     ]
>> >> >> >>   },
>> >> >> >>   "notify_list": [],
>> >> >> >>   "capacity": "",
>> >> >> >>   "hbase_mapping": {
>> >> >> >>     "column_family": [
>> >> >> >>       {
>> >> >> >>         "name": "f1",
>> >> >> >>         "columns": [
>> >> >> >>           {
>> >> >> >>             "qualifier": "m",
>> >> >> >>             "measure_refs": [
>> >> >> >>               "_COUNT_"
>> >> >> >>             ]
>> >> >> >>           }
>> >> >> >>         ]
>> >> >> >>       }
>> >> >> >>     ]
>> >> >> >>   },
>> >> >> >>   "project": "RetailProject",
>> >> >> >>   "model_name": "Retail_Cube"
>> >> >> >> }
>> >> >> >>
>> >> >> >>Regards,
>> >> >> >>Santosh Akhilesh
>> >> >> >>Bangalore R&D
>> >> >> >>HUAWEI TECHNOLOGIES CO.,LTD.
>> >> >> >>
>> >> >> >>www.huawei.com
>> >> >> >>----------------------------------------------------------
>> >> >> ---------------
>> >> >> >>-
>> >> >> >>-----------------------------------------------------------
>> >> >> >>This e-mail and its attachments contain confidential information
>> >>from
>> >> >> >>HUAWEI, which
>> >> >> >>is intended only for the person or entity whose address is
>>listed
>> >> >>above.
>> >> >> >>Any use of the
>> >> >> >>information contained herein in any way (including, but not
>>limited
>> >> >>to,
>> >> >> >>total or partial
>> >> >> >>disclosure, reproduction, or dissemination) by persons other
>>than
>> >>the
>> >> >> >>intended
>> >> >> >>recipient(s) is prohibited. If you receive this e-mail in error,
>> >> >>please
>> >> >> >>notify the sender by
>> >> >> >>phone or email immediately and delete it!
>> >> >> >>
>> >> >> >>________________________________________
>> >> >> >>From: Shi, Shaofeng [[email protected]]
>> >> >> >>Sent: Thursday, February 26, 2015 7:01 AM
>> >> >> >>To: [email protected]
>> >> >> >>Subject: Re: Error while making cube & Measure option is not
>> >> >>responding
>> >> >> >>on GUI
>> >> >> >>
>> >> >> >>Hi Santosh,
>> >> >> >>
>> >> >> >>It looks like hadoop failed to execute some shell command in the
>> >> >> >>container; You need dive into hadoop to see what¹s the concrete
>> >>error.
>> >> >> >>You
>> >> >> >>can use yarn logs command to fetch all logs:
>> >> >> >>
>> >> >> >>yarn logs -applicationId <app_id>
>> >> >> >>
>> >> >> >>
>> >> >> >>On 2/25/15, 7:39 PM, "Santosh Akhilesh"
>><[email protected]
>> >
>> >> >> >>wrote:
>> >> >> >>
>> >> >> >>>Hi Luke / Shaofeng ,
>> >> >> >>>           Can you please help me to check this issue.
>> >> >> >>>Regards,
>> >> >> >>>Santosh Akhilesh
>> >> >> >>>
>> >> >> >>>On Tue, Feb 24, 2015 at 10:41 PM, Santosh Akhilesh <
>> >> >> >>>[email protected]> wrote:
>> >> >> >>>
>> >> >> >>>> Hi All ,
>> >> >> >>>>         is it because of following error in map reduce job ?
>> >>what
>> >> >> >>>>could
>> >> >> >>>>be
>> >> >> >>>> way to resolve this , a google search says that its issue of
>> >>Yarn
>> >> >> >>>>class
>> >> >> >>>> path , but I am not sure what it is ?
>> >> >> >>>>
>> >> >> >>>> Kylin Hive Column Cardinality Job table=RETAIL.FACT_SALES
>> >> >> >>>> output=/tmp/cardinality/RETAIL.FACT_SALES
>> >> >> >>>>
>> >> >> >>>> Application application_1424791969399_0008 failed 2 times due
>> >>to AM
>> >> >> >>>> Container for appattempt_1424791969399_0008_000002 exited
>>with
>> >> >> >>>>exitCode: 1
>> >> >> >>>> For more detailed output, check application tracking page:
>> >> >> >>>> 
>>http://santosh:8088/proxy/application_1424791969399_0008/Then,
>> >> >>click
>> >> >> >>>>on
>> >> >> >>>> links to logs of each attempt.
>> >> >> >>>> Diagnostics: Exception from container-launch.
>> >> >> >>>> Container id: container_1424791969399_0008_02_000001
>> >> >> >>>> Exit code: 1
>> >> >> >>>> Stack trace: ExitCodeException exitCode=1:
>> >> >> >>>> at org.apache.hadoop.util.Shell.runCommand(Shell.java:538)
>> >> >> >>>> at org.apache.hadoop.util.Shell.run(Shell.java:455)
>> >> >> >>>> at
>> >> >> >>>>
>> >> >> >>>>org.apache.hadoop.util.Shell$ShellCommandExecutor.execut
>> >> >> e(Shell.java:71
>> >> >> >>>>5
>> >> >> >>>>)
>> >> >> >>>> at
>> >> >> >>>>
>> >> >> >>>>org.apache.hadoop.yarn.server.nodemanager.DefaultContain
>> >> >> erExecutor.laun
>> >> >> >>>>c
>> >> >> >>>>h
>> >> >> >>>>Container(DefaultContainerExecutor.java:211)
>> >> >> >>>> at
>> >> >> >>>>
>> >> >> >>>>org.apache.hadoop.yarn.server.nodemanager.containermanag
>> >> >> er.launcher.Con
>> >> >> >>>>t
>> >> >> >>>>a
>> >> >> >>>>inerLaunch.call(ContainerLaunch.java:302)
>> >> >> >>>> at
>> >> >> >>>>
>> >> >> >>>>org.apache.hadoop.yarn.server.nodemanager.containermanag
>> >> >> er.launcher.Con
>> >> >> >>>>t
>> >> >> >>>>a
>> >> >> >>>>inerLaunch.call(ContainerLaunch.java:82)
>> >> >> >>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>> >> >> >>>> at
>> >> >> >>>>
>> >> >> >>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread
>> >> >> PoolExecutor.ja
>> >> >> >>>>v
>> >> >> >>>>a
>> >> >> >>>>:1145)
>> >> >> >>>> at
>> >> >> >>>>
>> >> >> >>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa
>> >> >> dPoolExecutor.j
>> >> >> >>>>a
>> >> >> >>>>v
>> >> >> >>>>a:615)
>> >> >> >>>> at java.lang.Thread.run(Thread.java:745)
>> >> >> >>>> Container exited with a non-zero exit code 1
>> >> >> >>>> Failing this attempt. Failing the application.
>> >> >> >>>>
>> >> >> >>>> ---------- Forwarded message ----------
>> >> >> >>>> From: Santoshakhilesh <[email protected]>
>> >> >> >>>> Date: Tue, Feb 24, 2015 at 7:41 PM
>> >> >> >>>> Subject: FW: Error while making cube & Measure option is not
>> >> >> >>>>responding
>> >> >> >>>>on
>> >> >> >>>> GUI
>> >> >> >>>> To: "[email protected]"
>> >> >><[email protected]>
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>> hi ,
>> >> >> >>>>    please someone give me a hand to resolve this issue ,
>>thanks.
>> >> >> >>>>
>> >> >> >>>> Regards,
>> >> >> >>>> Santosh Akhilesh
>> >> >> >>>> Bangalore R&D
>> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD.
>> >> >> >>>>
>> >> >> >>>> www.huawei.com
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>>--------------------------------------------------------
>> >> >> ---------------
>> >> >> >>>>-
>> >> >> >>>>-
>> >> >> >>>>------------------------------------------------------------
>> >> >> >>>> This e-mail and its attachments contain confidential
>>information
>> >> >>from
>> >> >> >>>> HUAWEI, which
>> >> >> >>>> is intended only for the person or entity whose address is
>> >>listed
>> >> >> >>>>above.
>> >> >> >>>> Any use of the
>> >> >> >>>> information contained herein in any way (including, but not
>> >>limited
>> >> >> >>>>to,
>> >> >> >>>> total or partial
>> >> >> >>>> disclosure, reproduction, or dissemination) by persons other
>> >>than
>> >> >>the
>> >> >> >>>> intended
>> >> >> >>>> recipient(s) is prohibited. If you receive this e-mail in
>>error,
>> >> >> >>>>please
>> >> >> >>>> notify the sender by
>> >> >> >>>> phone or email immediately and delete it!
>> >> >> >>>>
>> >> >> >>>> ________________________________________
>> >> >> >>>> From: Santoshakhilesh [[email protected]]
>> >> >> >>>> Sent: Tuesday, February 24, 2015 12:55 PM
>> >> >> >>>> To: [email protected]
>> >> >> >>>> Cc: Kulbhushan Rana
>> >> >> >>>> Subject: FW: Error while making cube & Measure option is not
>> >> >> >>>>responding
>> >> >> >>>>on
>> >> >> >>>> GUI
>> >> >> >>>>
>> >> >> >>>> 2. If  I ignore and continue and try to save the cube I get
>>an
>> >> >> >>>>exception
>> >> >> >>>> in Kylin.log , I have checked the path is set correctly and
>> >> >> >>>>HCatInputFormat
>> >> >> >>>> this file is present in hive-hcatalog-core-0.14.0.jar .
>>Please
>> >>let
>> >> >>me
>> >> >> >>>>know
>> >> >> >>>> what can I do to resolve this ?
>> >> >> >>>>
>> >> >> >>>>  -- This was path issue , now no more exception in kylin.log
>> >> >> >>>>
>> >> >> >>>> But saveing cube still fails with error. And still can't add
>> >> >>measures.
>> >> >> >>>>
>> >> >> >>>> Error Message
>> >> >> >>>> Failed to take action.
>> >> >> >>>>
>> >> >> >>>> In log I can find no exception. Following is the last log in
>> >> >>kylin.log
>> >> >> >>>>
>> >> >> >>>> [pool-3-thread-1]:[2015-02-24
>> >> >> >>>>
>> >> >> >>>>20:47:15,613][INFO][org.apache.kylin.job.impl.threadpool
>> >> >> .DefaultSchedul
>> >> >> >>>>e
>> >> >> >>>>r
>> >> >> >>>>$FetcherRunner.run(DefaultScheduler.java:117)]
>> >> >> >>>> - Job Fetcher: 0 running, 0 actual running, 0 ready, 6 others
>> >> >> >>>> [http-bio-7070-exec-2]:[2015-02-24
>> >> >> >>>>
>> >> >> >>>>20:47:51,610][DEBUG][org.apache.kylin.rest.controller.Cu
>> >> >> beController.de
>> >> >> >>>>s
>> >> >> >>>>e
>> >> >> >>>>rializeDataModelDesc(CubeController.java:459)]
>> >> >> >>>> - Saving cube {
>> >> >> >>>>   "name": "",
>> >> >> >>>>   "fact_table": "RETAIL.FACT_SALES",
>> >> >> >>>>   "lookups": [],
>> >> >> >>>>   "filter_condition": "",
>> >> >> >>>>   "capacity": "SMALL",
>> >> >> >>>>   "partition_desc": {
>> >> >> >>>>     "partition_date_column": "",
>> >> >> >>>>     "partition_date_start": 0,
>> >> >> >>>>     "partition_type": "APPEND"
>> >> >> >>>>   },
>> >> >> >>>>   "last_modified": 0
>> >> >> >>>> }
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>> local access logs all with 200 , so seems ok.
>> >> >> >>>>
>> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:46:56 +0800] "GET
>> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:07 +0800] "GET
>> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:27 +0800] "GET
>> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:28 +0800] "GET
>> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:34 +0800] "GET
>> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:48 +0800] "GET
>> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:51 +0800] "POST
>> >> >>/kylin/api/cubes
>> >> >> >>>> HTTP/1.1" 200 701
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>> Regards,
>> >> >> >>>> Santosh Akhilesh
>> >> >> >>>> Bangalore R&D
>> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD.
>> >> >> >>>>
>> >> >> >>>> www.huawei.com
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>>--------------------------------------------------------
>> >> >> ---------------
>> >> >> >>>>-
>> >> >> >>>>-
>> >> >> >>>>------------------------------------------------------------
>> >> >> >>>> This e-mail and its attachments contain confidential
>>information
>> >> >>from
>> >> >> >>>> HUAWEI, which
>> >> >> >>>> is intended only for the person or entity whose address is
>> >>listed
>> >> >> >>>>above.
>> >> >> >>>> Any use of the
>> >> >> >>>> information contained herein in any way (including, but not
>> >>limited
>> >> >> >>>>to,
>> >> >> >>>> total or partial
>> >> >> >>>> disclosure, reproduction, or dissemination) by persons other
>> >>than
>> >> >>the
>> >> >> >>>> intended
>> >> >> >>>> recipient(s) is prohibited. If you receive this e-mail in
>>error,
>> >> >> >>>>please
>> >> >> >>>> notify the sender by
>> >> >> >>>> phone or email immediately and delete it!
>> >> >> >>>>
>> >> >> >>>> ________________________________________
>> >> >> >>>> From: Santoshakhilesh [[email protected]]
>> >> >> >>>> Sent: Tuesday, February 24, 2015 12:09 PM
>> >> >> >>>> To: [email protected]
>> >> >> >>>> Cc: Kulbhushan Rana
>> >> >> >>>> Subject: Error while making cube & Measure option is not
>> >> >>responding on
>> >> >> >>>>GUI
>> >> >> >>>>
>> >> >> >>>> Hi All ,
>> >> >> >>>>
>> >> >> >>>>     I am building a simple cube for test and using the binary
>> >>build
>> >> >> >>>>0.7.1
>> >> >> >>>> . I have following hive tables with columns.
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>> fact_sales:
>> >> >> >>>>
>> >> >> >>>> storeid                 int
>> >> >> >>>> itemid                  int
>> >> >> >>>> custid                  int
>> >> >> >>>> qty                     int
>> >> >> >>>> price                   double
>> >> >> >>>>
>> >> >> >>>> dim_customer
>> >> >> >>>> custid                  int
>> >> >> >>>> name                    string
>> >> >> >>>>
>> >> >> >>>> dim_item
>> >> >> >>>>
>> >> >> >>>> itemid                  int
>> >> >> >>>> category                string
>> >> >> >>>> brand                   string
>> >> >> >>>> color                   string
>> >> >> >>>>
>> >> >> >>>> dim_store
>> >> >> >>>>
>> >> >> >>>> storeid                 int
>> >> >> >>>> city                    string
>> >> >> >>>> state                   string
>> >> >> >>>>
>> >> >> >>>> Please help me to answer following issues;
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>> 1. When I go to measure section and click on measure option ,
>> >> >>there is
>> >> >> >>>>no
>> >> >> >>>> response , I want add measure on qty and price with sum
>> >> >> >>>>
>> >> >> >>>> 2. If  I ignore and continue and try to save the cube I get
>>an
>> >> >> >>>>exception
>> >> >> >>>> in Kylin.log , I have checked the path is set correctly and
>> >> >> >>>>HCatInputFormat
>> >> >> >>>> this file is present in hive-hcatalog-core-0.14.0.jar .
>>Please
>> >>let
>> >> >>me
>> >> >> >>>>know
>> >> >> >>>> what can I do to resolve this ?
>> >> >> >>>>
>> >> >> >>>> 3. Also I have another question since this is a test and
>>data is
>> >> >>small
>> >> >> >>>>I
>> >> >> >>>> have not partitioned the fact table , is it ok to skip
>>partition
>> >> >>stage
>> >> >> >>>> while cube build ?
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>> Exception
>> >> >> >>>>
>> >> >> >>>> pool-4-thread-4]:[2015-02-24
>> >> >> >>>>
>> >> >> >>>>19:26:32,577][ERROR][org.apache.kylin.job.impl.threadpoo
>> >> >> l.DefaultSchedu
>> >> >> >>>>l
>> >> >> >>>>e
>> >> >> >>>>r$JobRunner.run(DefaultScheduler.java:134)]
>> >> >> >>>> - ExecuteException job:c3532a6f-97ea-474a-b36a-218dd517cedb
>> >> >> >>>> org.apache.kylin.job.exception.ExecuteException:
>> >> >> >>>> org.apache.kylin.job.exception.ExecuteException:
>> >> >> >>>> java.lang.NoClassDefFoundError:
>> >> >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat
>> >> >> >>>>  at
>> >> >> >>>>
>> >> >> >>>>org.apache.kylin.job.execution.AbstractExecutable.execut
>> >> >> e(AbstractExecu
>> >> >> >>>>t
>> >> >> >>>>a
>> >> >> >>>>ble.java:102)
>> >> >> >>>>  at
>> >> >> >>>>
>> >> >> >>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$Jo
>> >> >> bRunner.run(Def
>> >> >> >>>>a
>> >> >> >>>>u
>> >> >> >>>>ltScheduler.java:132)
>> >> >> >>>>  at
>> >> >> >>>>
>> >> >> >>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread
>> >> >> PoolExecutor.ja
>> >> >> >>>>v
>> >> >> >>>>a
>> >> >> >>>>:1145)
>> >> >> >>>>  at
>> >> >> >>>>
>> >> >> >>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa
>> >> >> dPoolExecutor.j
>> >> >> >>>>a
>> >> >> >>>>v
>> >> >> >>>>a:615)
>> >> >> >>>>  at java.lang.Thread.run(Thread.java:745)
>> >> >> >>>> Caused by: org.apache.kylin.job.exception.ExecuteException:
>> >> >> >>>> java.lang.NoClassDefFoundError:
>> >> >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>> The JSON is as below.
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>> {
>> >> >> >>>>   "name": "Retail_Cube",
>> >> >> >>>>   "description": "",
>> >> >> >>>>   "dimensions": [
>> >> >> >>>>     {
>> >> >> >>>>       "name": "RETAIL.FACT_SALES.STOREID",
>> >> >> >>>>       "table": "RETAIL.FACT_SALES",
>> >> >> >>>>       "hierarchy": false,
>> >> >> >>>>       "derived": null,
>> >> >> >>>>       "column": [
>> >> >> >>>>         "STOREID"
>> >> >> >>>>       ],
>> >> >> >>>>       "id": 1
>> >> >> >>>>     },
>> >> >> >>>>     {
>> >> >> >>>>       "name": "RETAIL.FACT_SALES.ITEMID",
>> >> >> >>>>       "table": "RETAIL.FACT_SALES",
>> >> >> >>>>       "hierarchy": false,
>> >> >> >>>>       "derived": null,
>> >> >> >>>>       "column": [
>> >> >> >>>>         "ITEMID"
>> >> >> >>>>       ],
>> >> >> >>>>       "id": 2
>> >> >> >>>>     },
>> >> >> >>>>     {
>> >> >> >>>>       "name": "RETAIL.FACT_SALES.CUSTID",
>> >> >> >>>>       "table": "RETAIL.FACT_SALES",
>> >> >> >>>>       "hierarchy": false,
>> >> >> >>>>       "derived": null,
>> >> >> >>>>       "column": [
>> >> >> >>>>         "CUSTID"
>> >> >> >>>>       ],
>> >> >> >>>>       "id": 3
>> >> >> >>>>     }
>> >> >> >>>>   ],
>> >> >> >>>>   "measures": [
>> >> >> >>>>     {
>> >> >> >>>>       "id": 1,
>> >> >> >>>>       "name": "_COUNT_",
>> >> >> >>>>       "function": {
>> >> >> >>>>         "expression": "COUNT",
>> >> >> >>>>         "returntype": "bigint",
>> >> >> >>>>         "parameter": {
>> >> >> >>>>           "type": "constant",
>> >> >> >>>>           "value": 1
>> >> >> >>>>         }
>> >> >> >>>>       }
>> >> >> >>>>     }
>> >> >> >>>>   ],
>> >> >> >>>>   "rowkey": {
>> >> >> >>>>     "rowkey_columns": [
>> >> >> >>>>       {
>> >> >> >>>>         "column": "STOREID",
>> >> >> >>>>         "length": 0,
>> >> >> >>>>         "dictionary": "true",
>> >> >> >>>>         "mandatory": false
>> >> >> >>>>       },
>> >> >> >>>>       {
>> >> >> >>>>         "column": "ITEMID",
>> >> >> >>>>         "length": 0,
>> >> >> >>>>         "dictionary": "true",
>> >> >> >>>>         "mandatory": false
>> >> >> >>>>       },
>> >> >> >>>>       {
>> >> >> >>>>         "column": "CUSTID",
>> >> >> >>>>         "length": 0,
>> >> >> >>>>         "dictionary": "true",
>> >> >> >>>>         "mandatory": false
>> >> >> >>>>       }
>> >> >> >>>>     ],
>> >> >> >>>>     "aggregation_groups": [
>> >> >> >>>>       [
>> >> >> >>>>         "STOREID",
>> >> >> >>>>         "ITEMID",
>> >> >> >>>>         "CUSTID"
>> >> >> >>>>       ]
>> >> >> >>>>     ]
>> >> >> >>>>   },
>> >> >> >>>>   "notify_list": [],
>> >> >> >>>>   "capacity": "",
>> >> >> >>>>   "hbase_mapping": {
>> >> >> >>>>     "column_family": [
>> >> >> >>>>       {
>> >> >> >>>>         "name": "f1",
>> >> >> >>>>         "columns": [
>> >> >> >>>>           {
>> >> >> >>>>             "qualifier": "m",
>> >> >> >>>>             "measure_refs": [
>> >> >> >>>>               "_COUNT_"
>> >> >> >>>>             ]
>> >> >> >>>>           }
>> >> >> >>>>         ]
>> >> >> >>>>       }
>> >> >> >>>>     ]
>> >> >> >>>>   },
>> >> >> >>>>   "project": "RetailProject",
>> >> >> >>>>   "model_name": "Retail_Cube"
>> >> >> >>>> }
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>> Regards,
>> >> >> >>>> Santosh Akhilesh
>> >> >> >>>> Bangalore R&D
>> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD.
>> >> >> >>>>
>> >> >> >>>> www.huawei.com
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>>--------------------------------------------------------
>> >> >> ---------------
>> >> >> >>>>-
>> >> >> >>>>-
>> >> >> >>>>------------------------------------------------------------
>> >> >> >>>> This e-mail and its attachments contain confidential 
>>information
>> >> >>from
>> >> >> >>>> HUAWEI, which
>> >> >> >>>> is intended only for the person or entity whose address is
>> >>listed
>> >> >> >>>>above.
>> >> >> >>>> Any use of the
>> >> >> >>>> information contained herein in any way (including, but not
>> >>limited
>> >> >> >>>>to,
>> >> >> >>>> total or partial
>> >> >> >>>> disclosure, reproduction, or dissemination) by persons other
>> >>than
>> >> >>the
>> >> >> >>>> intended
>> >> >> >>>> recipient(s) is prohibited. If you receive this e-mail in 
>>error,
>> >> >> >>>>please
>> >> >> >>>> notify the sender by
>> >> >> >>>> phone or email immediately and delete it!
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>> --
>> >> >> >>>> Regards,
>> >> >> >>>> Santosh Akhilesh
>> >> >> >>>> +91-0-9845482201
>> >> >> >>>>
>> >> >> >>>
>> >> >> >>>
>> >> >> >>>
>> >> >> >>>--
>> >> >> >>>Regards,
>> >> >> >>>Santosh Akhilesh
>> >> >> >>>+91-0-9845482201
>> >> >>
>> >> >>
>> >>
>> >>
>>
>>

Reply via email to