Yes I did
On Fri, 27 Feb 2015 at 8:30 am, Shi, Shaofeng <[email protected]> wrote:

> Did you append the database name before table name when importing it to
> Kylin?
>
> On 2/27/15, 10:51 AM, "Santosh Akhilesh" <[email protected]>
> wrote:
>
> >Ok , but it's strange that this issue is not reported while importing hive
> >table.
> >If this is limitation of this version , I guess it's better to detect and
> >prompt user during hive table import. This is just my suggestion.
> >On Fri, 27 Feb 2015 at 8:02 am, Shi, Shaofeng <[email protected]> wrote:
> >
> >> In 0.6.x it only support tables in default database, this is a
> >>limitation;
> >> The support for non-default tables will be released in 0.7;
> >>
> >> To bypass this issue for now, please copy the table to default database
> >>as
> >> a workaround;
> >>
> >> On 2/27/15, 10:16 AM, "Santosh Akhilesh" <[email protected]>
> >> wrote:
> >>
> >> >@Jason
> >> >thanks , but now as suggested by Saofeng I m not using the inverted
> >>index
> >> >brach as its not stable.
> >> >I have switched back to 0.6 branch , in this branch yesterday night I
> >> >could
> >> >crete the cube successfully but there is issue while building it. I
> >>feel
> >> >that at step 1 of cube build  while creating flat table when command is
> >> >issued to hive if the tables are not under default datbase flat table
> >> >creation is failed and cube build fails. my fact and dimension tables
> >>are
> >> >under a database called retail.
> >> >
> >> >@Saofeng - Can you please confirm this behavior ? Do I need to create
> >>the
> >> >hive tables under default database?
> >> >
> >> >On Fri, Feb 27, 2015 at 7:32 AM, jason zhong <[email protected]>
> >> >wrote:
> >> >
> >> >> @Santoshakhilesh
> >> >>
> >> >> 1. When I go to measure section and click on measure option , there
> >>is
> >> >>no
> >> >> response , I want add measure on qty and price with sum
> >> >>          --bug fixed on inverted-index branch
> >> >>
> >> >>
> >> >> On Fri, Feb 27, 2015 at 3:03 AM, Santosh Akhilesh <
> >> >> [email protected]
> >> >> > wrote:
> >> >>
> >> >> > Hi Shaofeng ,
> >> >> >      I have build the 0.6 version and now able to create the cube
> >> >> > successfully.
> >> >> >      While building the cube , it fails at step1 with following
> >>error.
> >> >> > Table not found 'DIM_ITEM'
> >> >> >      the table exists , but its under retail data base and not
> >>under
> >> >> > default database.
> >> >> >      does kylin require hive taables to be under default database
> >>? I
> >> >>see
> >> >> > the flat table being created under default database.
> >> >> >
> >> >> > Logging initialized using configuration in
> >> >> >
> >> >> >
> >> >>
> >> >>jar:file:/home/santosh/work/frameworks/apache-hive-1.0.0/
> >> lib/hive-common-
> >> >>1.0.0.jar!/hive-log4j.properties
> >> >> > SLF4J: Class path contains multiple SLF4J bindings.
> >> >> > SLF4J: Found binding in
> >> >> >
> >> >> >
> >> >>
> >> >>[jar:file:/home/santosh/work/frameworks/hadoop-2.6.0/
> >> share/hadoop/common/
> >> >>lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> >> >> > SLF4J: Found binding in
> >> >> >
> >> >> >
> >> >>
> >> >>[jar:file:/home/santosh/work/frameworks/apache-hive-1.
> >> 0.0/lib/hive-jdbc-1
> >> >>.0.0-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> >> >> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for
> an
> >> >> > explanation.
> >> >> > SLF4J: Actual binding is of type
> >>[org.slf4j.impl.Log4jLoggerFactory]
> >> >> > OK
> >> >> > Time taken: 0.964 seconds
> >> >> > OK
> >> >> > Time taken: 0.948 seconds
> >> >> > FAILED: SemanticException [Error 10001]: Line 12:11 Table not found
> >> >> > 'DIM_ITEM'
> >> >> >
> >> >> >
> >> >> >
> >> >> > Command is as below.
> >> >> >
> >> >> > hive -e "DROP TABLE IF EXISTS
> >> >> >
> >>
> >>>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_
> 8c0f_07d1f559dd44
> >>>>;
> >> >> > CREATE EXTERNAL TABLE IF NOT EXISTS
> >> >> >
> >>
> >>>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_
> 8c0f_07d1f559dd44
> >> >> > (
> >> >> > STOREID int
> >> >> > ,ITEMID int
> >> >> > ,CUSTID int
> >> >> > ,QTY int
> >> >> > ,AMOUNT double
> >> >> > )
> >> >> > ROW FORMAT DELIMITED FIELDS TERMINATED BY '\177'
> >> >> > STORED AS SEQUENCEFILE
> >> >> > LOCATION
> >> >> >
> >> >>
> >> >>'/tmp/kylin-8b30b29b-5f2c-4b63-8c0f-07d1f559dd44/kylin_
> >> intermediate_test_
> >> >>FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd44';
> >> >> > SET hive.exec.compress.output=true;
> >> >> > SET hive.auto.convert.join.noconditionaltask = true;
> >> >> > SET hive.auto.convert.join.noconditionaltask.size = 300000000;
> >> >> > INSERT OVERWRITE TABLE
> >> >> >
> >>
> >>>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_
> 8c0f_07d1f559dd44
> >> >> > SELECT
> >> >> > FACT_SALES.STOREID
> >> >> > ,FACT_SALES.ITEMID
> >> >> > ,FACT_SALES.CUSTID
> >> >> > ,FACT_SALES.QTY
> >> >> > ,FACT_SALES.AMOUNT
> >> >> > FROM FACT_SALES
> >> >> > INNER JOIN DIM_STORE
> >> >> > ON FACT_SALES.STOREID = DIM_STORE.SROREID
> >> >> > INNER JOIN DIM_ITEM
> >> >> > ON FACT_SALES.ITEMID = DIM_ITEM.ITEMID
> >> >> > INNER JOIN DIM_CUSTOMER
> >> >> > ON FACT_SALES.CUSTID = DIM_CUSTOMER.CUSTID
> >> >> > ;
> >> >> > "
> >> >> >
> >> >> >
> >> >> >
> >> >> > On Thu, Feb 26, 2015 at 8:11 PM, Shi, Shaofeng <[email protected]>
> >> >>wrote:
> >> >> >
> >> >> > > The 0.7.1 is test version, its package contains the “snapshot”
> >> >>suffix;
> >> >> we
> >> >> > > will upload a new package there; Luke will also add a message
> >>there
> >> >>to
> >> >> > > avoid this confusion;
> >> >> > >
> >> >> > > Regarding the problem that you encountered, could you please
> >>open a
> >> >> JIRA
> >> >> > > ticket for tracking? Here is link of Apache JIRA:
> >> >> > >
> >> >> > > https://issues.apache.org/jira/secure/Dashboard.jspa
> >> >> > >
> >> >> > >
> >> >> > > Thanks for the feedback!
> >> >> > >
> >> >> > > On 2/26/15, 10:21 PM, "Santosh Akhilesh"
> >><[email protected]
> >> >
> >> >> > > wrote:
> >> >> > >
> >> >> > > >Actually I see this being published on kylin webpage.
> >> >> > > >http://kylin.incubator.apache.org/download/
> >> >> > > >I am using 0.7.1 inverted index branch binary distribution.
> >> >> > > >If this is not stable please give me the link of stable branch I
> >> >>would
> >> >> > try
> >> >> > > >building and testing tonight.
> >> >> > > >On Thu, 26 Feb 2015 at 7:30 pm, Shi, Shaofeng <[email protected]
> >
> >> >> wrote:
> >> >> > > >
> >> >> > > >> Hi Santosh, it is not recommended to use the dev code branch
> >> >> > (actually I
> >> >> > > >> don’t know how you get the v0.7.x build and what’s the detail
> >> >> version
> >> >> > of
> >> >> > > >> that; each day we submit many changes to that);
> >> >> > > >>
> >> >> > > >> The options are 1) switch back to latest release v0.6.5; or 2)
> >> >>wait
> >> >> > for
> >> >> > > >> the formal release of 0.7, that should be in March;
> >>Otherwise, we
> >> >> > > >>couldn’t
> >> >> > > >> ensure there is no new problems come out in your next steps;
> >> >> > > >>
> >> >> > > >> On 2/26/15, 5:39 PM, "Santosh Akhilesh"
> >> >><[email protected]>
> >> >> > > >>wrote:
> >> >> > > >>
> >> >> > > >> >Hi Shaofeng
> >> >> > > >> >So what do you suggest , how should I proceed further with
> >>this
> >> >> > > >>release?
> >> >> > > >> >Will there be a patch? Any alternate way I can create cube?
> >> >> > > >> >Please suggest.
> >> >> > > >> >Regards
> >> >> > > >> >Santosh
> >> >> > > >> >On Thu, 26 Feb 2015 at 3:04 pm, Shi, Shaofeng
> >><[email protected]
> >> >
> >> >> > > wrote:
> >> >> > > >> >
> >> >> > > >> >> Hi Santosh,
> >> >> > > >> >>
> >> >> > > >> >> 0.7.1 hasn’t been formally released; from 0.6.x to 0.7.x we
> >> >>have
> >> >> > > >> >>metadata
> >> >> > > >> >> structure change; While, the web UI (cube wizard) for this
> >> >>change
> >> >> > > >>hasn’t
> >> >> > > >> >> been stabilized; So it is not strange that you got trouble
> >> >>when
> >> >> > > >>saving
> >> >> > > >> >>the
> >> >> > > >> >> cube;
> >> >> > > >> >>
> >> >> > > >> >> @Jason, any idea about the JS error?
> >> >> > > >> >>
> >> >> > > >> >> On 2/26/15, 5:08 PM, "Santosh Akhilesh" <
> >> >> [email protected]
> >> >> > >
> >> >> > > >> >>wrote:
> >> >> > > >> >>
> >> >> > > >> >> >Hi Shaofeng,
> >> >> > > >> >> >
> >> >> > > >> >> >I am using the binary distribution 0.7.1. I have not been
> >> >>able
> >> >> to
> >> >> > > >>save
> >> >> > > >> >> >cube
> >> >> > > >> >> >even once. I have tried creating new project and from
> >>local
> >> >> > machine
> >> >> > > >>and
> >> >> > > >> >> >server machine. But I am always stuck with this error. I
> >>am
> >> >> never
> >> >> > > >> >>allowed
> >> >> > > >> >> >to add measures and never been able to save the cube. I
> >>also
> >> >>see
> >> >> > the
> >> >> > > >> >> >kylin.log and it always tries to save cube with append
> >>mode.
> >> >>One
> >> >> > > >>thing
> >> >> > > >> >>I
> >> >> > > >> >> >need to tell that at partition stage since I don't have a
> >>big
> >> >> fact
> >> >> > > >> >>table
> >> >> > > >> >> >now I have not partititioned the fact table and I skip
> >>this
> >> >> step.
> >> >> > > >>Does
> >> >> > > >> >> >this
> >> >> > > >> >> >have affect in saving the cube. Is this because some
> >> >>metadata is
> >> >> > > >> >>available
> >> >> > > >> >> >and it tries to modify the cube? I am using latest Hadoop
> >> >>2.6.6.
> >> >> > Yes
> >> >> > > >> >>kylin
> >> >> > > >> >> >propert I have not added the jar. I will add them and
> >>check.
> >> >>But
> >> >> > > >>cube
> >> >> > > >> >> >creation failure is really puzzling me. I could see no
> >>error
> >> >> logs
> >> >> > in
> >> >> > > >> >> >kylin.log.
> >> >> > > >> >> >Regards
> >> >> > > >> >> >Santosh
> >> >> > > >> >> >On Thu, 26 Feb 2015 at 1:40 pm, Shi, Shaofeng
> >> >><[email protected]
> >> >> >
> >> >> > > >> wrote:
> >> >> > > >> >> >
> >> >> > > >> >> >> Which version or code branch are you using? I assume
> >>you’re
> >> >> > using
> >> >> > > >>the
> >> >> > > >> >> >> stable version from master; Seems you’re trying to edit
> >>an
> >> >> > > >>existing
> >> >> > > >> >>cube
> >> >> > > >> >> >> to add new measurement, try refresh your browser's
> >>cache;
> >> >>if
> >> >> it
> >> >> > > >>still
> >> >> > > >> >> >> couldn’t be saved, try to create a new cube;
> >> >> > > >> >> >>
> >> >> > > >> >> >> The two error traces in tomcat need be taken care:
> >> >> > > >> >> >>
> >> >> > > >> >> >> 1) java.lang.NoClassDefFoundError:
> >> >> > > >> >> >>org/apache/kylin/common/mr/KylinMapper
> >> >> > > >> >> >>         Please check kylin.properties file, making sure
> >>the
> >> >> > > >> >> >>“kylin.job.jar”
> >> >> > > >> >> >> points to a right jar file; It will be loaded in
> >> >>Map-reduce;
> >> >> > > >> >> >>
> >> >> > > >> >> >> 2) java.lang.IllegalArgumentException: No enum constant
> >> >> > > >> >> >> org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS
> >> >> > > >> >> >>         This indicates your hadoop version might be old;
> >> >> Please
> >> >> > > >>check
> >> >> > > >> >> >>and
> >> >> > > >> >> >> ensure
> >> >> > > >> >> >> hadoop version is 2.2 or above.
> >> >> > > >> >> >>
> >> >> > > >> >> >> On 2/26/15, 3:21 PM, "Santoshakhilesh"
> >> >> > > >><[email protected]>
> >> >> > > >> >> >> wrote:
> >> >> > > >> >> >>
> >> >> > > >> >> >> >Hi Shaofeng ,
> >> >> > > >> >> >> >
> >> >> > > >> >> >> >   I am using chrome , When I click on button to add
> >> >> measures ,
> >> >> > > >> >> >>following
> >> >> > > >> >> >> >is error on chrome console. When I try to save the cube
> >> >>there
> >> >> > is
> >> >> > > >>no
> >> >> > > >> >> >>error
> >> >> > > >> >> >> >in console. I just get a error dialog saying failed to
> >> >>take
> >> >> > > >>action
> >> >> > > >> >>and
> >> >> > > >> >> >> >gives me the JSON cube schema.
> >> >> > > >> >> >> >
> >> >> > > >> >> >> >Error on chrome debug console is as below;
> >> >> > > >> >> >> >
> >> >> > > >> >> >> > ReferenceError: CubeDescModel is not defined
> >> >> > > >> >> >> >    at h.$scope.addNewMeasure (scripts.min.0.js:15984)
> >> >> > > >> >> >> >    at scripts.min.0.js:180
> >> >> > > >> >> >> >    at scripts.min.0.js:197
> >> >> > > >> >> >> >    at h.$eval (scripts.min.0.js:119)
> >> >> > > >> >> >> >    at h.$apply (scripts.min.0.js:119)
> >> >> > > >> >> >> >    at HTMLButtonElement.<anonymous>
> >> >>(scripts.min.0.js:197)
> >> >> > > >> >> >> >    at HTMLButtonElement.m.event.dispatch
> >> >> (scripts.min.0.js:3)
> >> >> > > >> >> >> >    at HTMLButtonElement.r.handle
> >> >> > > >> >> >> >(scripts.min.0.js:3)scripts.min.0.js:100 (anonymous
> >> >> > > >> >> >> >function)scripts.min.0.js:77 (anonymous
> >> >> > > >> >>function)scripts.min.0.js:119
> >> >> > > >> >> >> >h.$applyscripts.min.0.js:197 (anonymous
> >> >> > > >>function)scripts.min.0.js:3
> >> >> > > >> >> >> >m.event.dispatchscripts.min.0.js:3 r.handle
> >> >> > > >> >> >> >
> >> >> > > >> >> >> >   About the hive table import , I got pass the run
> >>shell
> >> >> > command
> >> >> > > >> >> >> >exception but it still fails the hadoop log is;
> >> >> > > >> >> >> >2015-02-26 20:46:48,332 INFO [main]
> >> >>org.apache.hadoop.mapred.
> >> >> > > >> >> YarnChild:
> >> >> > > >> >> >> >mapreduce.cluster.local.dir for child:
> >> >> > > >> >> >>
> >> >>>/tmp/hadoop-root/nm-local-dir/usercache/root/appcache/appli
> >> >> > > >> >> >> cation_14249530
> >> >> > > >> >> >> >91340_0002
> >> >> > > >> >> >> >2015-02-26 20:46:48,776 INFO [main]
> >> >> > > >> >> >> >org.apache.hadoop.conf.Configuration.deprecation:
> >> >>session.id
> >> >> > is
> >> >> > > >> >> >> >deprecated. Instead, use dfs.metrics.session-id
> >> >> > > >> >> >> >2015-02-26 20:46:49,310 INFO [main]
> >> >> > > >>org.apache.hadoop.mapred.Task:
> >> >> > > >> >> >>Using
> >> >> > > >> >> >> >ResourceCalculatorProcessTree : [ ]
> >> >> > > >> >> >> >2015-02-26 20:46:49,386 FATAL [main]
> >> >> > > >> >> >>org.apache.hadoop.mapred.YarnChild:
> >> >> > > >> >> >> >Error running child : java.lang.NoClassDefFoundError:
> >> >> > > >> >> >> >org/apache/kylin/common/mr/KylinMapper
> >> >> > > >> >> >> > at java.lang.ClassLoader.defineClass1(Native Method)
> >> >> > > >> >> >> > at
> >> >>java.lang.ClassLoader.defineClass(ClassLoader.java:800)
> >> >> > > >> >> >> > at
> >> >> > > >> >> >> >java.security.SecureClassLoader.defineClass(
> >> >> > > >> >> SecureClassLoader.java:142)
> >> >> > > >> >> >> > at
> >> >> > java.net.URLClassLoader.defineClass(URLClassLoader.java:449)
> >> >> > > >> >> >> > at
> >> >> java.net.URLClassLoader.access$100(URLClassLoader.java:71)
> >> >> > > >> >> >> > at java.net.URLClassLoader$1.run(
> >> URLClassLoader.java:361)
> >> >> > > >> >> >> > at java.net.URLClassLoader$1.run(
> >> URLClassLoader.java:355)
> >> >> > > >> >> >> > at java.security.AccessController.doPrivileged(Native
> >> >> Method)
> >> >> > > >> >> >> > at
> >> >> java.net.URLClassLoader.findClass(URLClassLoader.java:354)
> >> >> > > >> >> >> > at
> >>java.lang.ClassLoader.loadClass(ClassLoader.java:425)
> >> >> > > >> >> >> > at
> >> >> > sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
> >> >> > > >> >> >> >
> >> >> > > >> >> >> >tomcat logs:
> >> >> > > >> >> >> >usage: HiveColumnCardinalityJob
> >> >> > > >> >> >> > -output <path>        Output path
> >> >> > > >> >> >> > -table <table name>   The hive table name
> >> >> > > >> >> >> >[pool-4-thread-2]:[2015-02-26
> >> >> > > >> >> >>
> >> >>>20:47:49,936][ERROR][org.apache.kylin.job.common.HadoopShel
> >> >> > > >> >> >> lExecutable.doW
> >> >> > > >> >> >> >ork(HadoopShellExecutable.java:64)] - error execute
> >> >> > > >> >> >>
> >> >> > >HadoopShellExecutable{id=d4730d26-7fe6-412e-9841-3288ab362c5b-00,
> >> >> > > >> >> >> >name=null, state=RUNNING}
> >> >> > > >> >> >> >java.lang.IllegalArgumentException: No enum constant
> >> >> > > >> >> >> >org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS
> >> >> > > >> >> >> > at java.lang.Enum.valueOf(Enum.java:236)
> >> >> > > >> >> >> > at
> >> >> > > >> >> >>
> >> >> > >
> >> >>>>>org.apache.hadoop.mapreduce.counters.
> FrameworkCounterGroup.valueOf(
> >> >> > > >> >> >> Framewo
> >> >> > > >> >> >> >rkCounterGroup.java:148)
> >> >> > > >> >> >> > at
> >> >> > > >> >> >>
> >> >>>org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.
> >> >> > > >> >> >> findCounter(Fra
> >> >> > > >> >> >> >meworkCounterGroup.java:182)
> >> >> > > >> >> >> > at
> >> >> > > >> >> >>
> >> >>>org.apache.hadoop.mapreduce.counters.AbstractCounters.findC
> >> >> > > >> >> >> ounter(Abstract
> >> >> > > >> >> >> >Counters.java:154)
> >> >> > > >> >> >> > at
> >> >> > > >> >> >>
> >> >>>org.apache.hadoop.mapreduce.TypeConverter.fromYarn(TypeConv
> >> >> > > >> >> >> erter.java:240)
> >> >> > > >> >> >> > at
> >> >> > > >> >> >>
> >> >>>org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounte
> >> >> > > >> >> >> rs(ClientServic
> >> >> > > >> >> >> >eDelegate.java:370)
> >> >> > > >> >> >> > at
> >> >> > > >> >> >> >org.apache.hadoop.mapred.YARNRunner.getJobCounters(
> >> >> > > >> >> YARNRunner.java:511)
> >> >> > > >> >> >> > at org.apache.hadoop.mapreduce.
> Job$7.run(Job.java:756)
> >> >> > > >> >> >> > at org.apache.hadoop.mapreduce.
> Job$7.run(Job.java:753)
> >> >> > > >> >> >> > at java.security.AccessController.doPrivileged(Native
> >> >> Method)
> >> >> > > >> >> >> > at javax.security.auth.Subject.doAs(Subject.java:415)
> >> >> > > >> >> >> > at
> >> >> > > >> >> >>
> >> >>>org.apache.hadoop.security.UserGroupInformation.doAs(UserGr
> >> >> > > >> >> >> oupInformation.
> >> >> > > >> >> >> >java:1491)
> >> >> > > >> >> >> > at
> >> >>org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753)
> >> >> > > >> >> >> > at
> >> >> > > >>
> >> >>>>org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1361)
> >> >> > > >> >> >> > at org.apache.hadoop.mapreduce.
> >> Job.waitForCompletion(Job.
> >> >> > > >> java:1289)
> >> >> > > >> >> >> > at
> >> >> > > >> >> >>
> >> >>>org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForComple
> >> >> > > >> >> >> tion(AbstractHa
> >> >> > > >> >> >> >doopJob.java:134)
> >> >> > > >> >> >> > at
> >> >> > > >> >> >>
> >> >> > > >>
> >> >> >
> >>
> >>>>>>>org.apache.kylin.job.hadoop.cardinality.
> HiveColumnCardinalityJob.run
> >>>>>>>(
> >> >> > > >> >> >> HiveC
> >> >> > > >> >> >> >olumnCardinalityJob.java:114)
> >> >> > > >> >> >> > at
> >> >>org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> >> >> > > >> >> >> > at
> >> >>org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
> >> >> > > >> >> >> > at
> >> >> > > >> >> >>
> >> >>>org.apache.kylin.job.common.HadoopShellExecutable.doWork(Ha
> >> >> > > >> >> >> doopShellExecut
> >> >> > > >> >> >> >able.java:62)
> >> >> > > >> >> >> > at
> >> >> > > >> >> >>
> >> >>>org.apache.kylin.job.execution.AbstractExecutable.execute(A
> >> >> > > >> >> >> bstractExecutab
> >> >> > > >> >> >> >le.java:99)
> >> >> > > >> >> >> > at
> >> >> > > >> >> >>
> >> >>>org.apache.kylin.job.execution.DefaultChainedExecutable.doW
> >> >> > > >> >> >> ork(DefaultChai
> >> >> > > >> >> >> >nedExecutable.java:50)
> >> >> > > >> >> >> > at
> >> >> > > >> >> >>
> >> >>>org.apache.kylin.job.execution.AbstractExecutable.execute(A
> >> >> > > >> >> >> bstractExecutab
> >> >> > > >> >> >> >le.java:99)
> >> >> > > >> >> >> > at
> >> >> > > >> >> >>
> >> >>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRu
> >> >> > > >> >> >> nner.run(Defaul
> >> >> > > >> >> >> >tScheduler.java:132)
> >> >> > > >> >> >> > at
> >> >> > > >> >> >>
> >> >>>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoo
> >> >> > > >> >> >> lExecutor.java:
> >> >> > > >> >> >> >1145)
> >> >> > > >> >> >> > at
> >> >> > > >> >> >>
> >> >>>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPo
> >> >> > > >> >> >> olExecutor.java
> >> >> > > >> >> >> >:615)
> >> >> > > >> >> >> > at java.lang.Thread.run(Thread.java:745)
> >> >> > > >> >> >> >
> >> >> > > >> >> >> >Regards,
> >> >> > > >> >> >> >Santosh Akhilesh
> >> >> > > >> >> >> >Bangalore R&D
> >> >> > > >> >> >> >HUAWEI TECHNOLOGIES CO.,LTD.
> >> >> > > >> >> >> >
> >> >> > > >> >> >> >www.huawei.com
> >> >> > > >> >> >>
> >> >>>-----------------------------------------------------------
> >> >> > > >> >> >> ---------------
> >> >> > > >> >> >>
> >> >>>-----------------------------------------------------------
> >> >> > > >> >> >> >This e-mail and its attachments contain confidential
> >> >> > information
> >> >> > > >> >>from
> >> >> > > >> >> >> >HUAWEI, which
> >> >> > > >> >> >> >is intended only for the person or entity whose
> >>address is
> >> >> > listed
> >> >> > > >> >> >>above.
> >> >> > > >> >> >> >Any use of the
> >> >> > > >> >> >> >information contained herein in any way (including, but
> >> >>not
> >> >> > > >>limited
> >> >> > > >> >>to,
> >> >> > > >> >> >> >total or partial
> >> >> > > >> >> >> >disclosure, reproduction, or dissemination) by persons
> >> >>other
> >> >> > than
> >> >> > > >> >>the
> >> >> > > >> >> >> >intended
> >> >> > > >> >> >> >recipient(s) is prohibited. If you receive this e-mail
> >>in
> >> >> > error,
> >> >> > > >> >>please
> >> >> > > >> >> >> >notify the sender by
> >> >> > > >> >> >> >phone or email immediately and delete it!
> >> >> > > >> >> >> >
> >> >> > > >> >> >> >________________________________________
> >> >> > > >> >> >> >From: Shi, Shaofeng [[email protected]]
> >> >> > > >> >> >> >Sent: Thursday, February 26, 2015 11:32 AM
> >> >> > > >> >> >> >To: [email protected]
> >> >> > > >> >> >> >Cc: Kulbhushan Rana
> >> >> > > >> >> >> >Subject: Re: Error while making cube & Measure option
> >>is
> >> >>not
> >> >> > > >> >>responding
> >> >> > > >> >> >> >on GUI
> >> >> > > >> >> >> >
> >> >> > > >> >> >> >Hi Santosh, hive table importing issue should not
> >>impact
> >> >>on
> >> >> > cube
> >> >> > > >> >> >>saving.
> >> >> > > >> >> >> >
> >> >> > > >> >> >> >If you couldn’t save the cube, firstly please check
> >> >>whether
> >> >> > > >>there is
> >> >> > > >> >> >>error
> >> >> > > >> >> >> >in the tomcat’s log; If not please check your web
> >> >>browser; We
> >> >> > > >> >>suggest
> >> >> > > >> >> >>use
> >> >> > > >> >> >> >Firefox (with firebug add-on) or Chrome, open the JS
> >> >>console
> >> >> > > >>(press
> >> >> > > >> >> >>F12)
> >> >> > > >> >> >> >and then operate web UI, check whether there is any
> >>error
> >> >> > > >>reported
> >> >> > > >> >>in
> >> >> > > >> >> >> >browser.
> >> >> > > >> >> >> >
> >> >> > > >> >> >> >
> >> >> > > >> >> >> >On 2/26/15, 1:08 PM, "Santoshakhilesh"
> >> >> > > >><[email protected]
> >> >> > > >> >
> >> >> > > >> >> >> >wrote:
> >> >> > > >> >> >> >
> >> >> > > >> >> >> >>Hi Shaofeng ,
> >> >> > > >> >> >> >>   Thanks for replying.
> >> >> > > >> >> >> >>   Yes I am checking the yarn exception, But I find
> >>that
> >> >> this
> >> >> > > >>error
> >> >> > > >> >> >>comes
> >> >> > > >> >> >> >>while importing the hive table to kylin.
> >> >> > > >> >> >> >>   Even if this error comes , hive tables is exported
> >> >> > > >>successfully
> >> >> > > >> >>in
> >> >> > > >> >> >> >>kylin. Is this the reason why cube saving has failed ?
> >> >> > > >> >> >> >>
> >> >> > > >> >> >> >>   Next step when I go on creating the cube for
> >>following
> >> >> > > >>schema  I
> >> >> > > >> >> >>get
> >> >> > > >> >> >> >>error at last step while saving and  I am unable to
> >>add
> >> >>any
> >> >> > > >> >>measures ,
> >> >> > > >> >> >> >>clicking on measure option just dont pop up any
> >>dialog,
> >> >> > > >> >> >> >>
> >> >> > > >> >> >> >>I am using star schema with fact_sales as fact table
> >>and
> >> >> dim_*
> >> >> > > >>as
> >> >> > > >> >> >> >>dimension tables.
> >> >> > > >> >> >> >>
> >> >> > > >> >> >> >> fact_sales:
> >> >> > > >> >> >> >> storeid                 int
> >> >> > > >> >> >> >> itemid                  int
> >> >> > > >> >> >> >> custid                  int
> >> >> > > >> >> >> >> qty                     int
> >> >> > > >> >> >> >> price                   double
> >> >> > > >> >> >> >>
> >> >> > > >> >> >> >>dim_customer
> >> >> > > >> >> >> >> custid                  int
> >> >> > > >> >> >> >> name                    string
> >> >> > > >> >> >> >>
> >> >> > > >> >> >> >> dim_item
> >> >> > > >> >> >> >> itemid                  int
> >> >> > > >> >> >> >> category                string
> >> >> > > >> >> >> >> brand                   string
> >> >> > > >> >> >> >> color                   string
> >> >> > > >> >> >> >>
> >> >> > > >> >> >> >>dim_store
> >> >> > > >> >> >> >> storeid                 int
> >> >> > > >> >> >> >> city                    string
> >> >> > > >> >> >> >> state                   string
> >> >> > > >> >> >> >>
> >> >> > > >> >> >> >>JSON is as below.
> >> >> > > >> >> >> >> The JSON is as below.
> >> >> > > >> >> >> >>
> >> >> > > >> >> >> >> {
> >> >> > > >> >> >> >>   "name": "Retail_Cube",
> >> >> > > >> >> >> >>   "description": "",
> >> >> > > >> >> >> >>   "dimensions": [
> >> >> > > >> >> >> >>     {
> >> >> > > >> >> >> >>       "name": "RETAIL.FACT_SALES.STOREID",
> >> >> > > >> >> >> >>       "table": "RETAIL.FACT_SALES",
> >> >> > > >> >> >> >>       "hierarchy": false,
> >> >> > > >> >> >> >>       "derived": null,
> >> >> > > >> >> >> >>       "column": [
> >> >> > > >> >> >> >>         "STOREID"
> >> >> > > >> >> >> >>       ],
> >> >> > > >> >> >> >>       "id": 1
> >> >> > > >> >> >> >>     },
> >> >> > > >> >> >> >>     {
> >> >> > > >> >> >> >>       "name": "RETAIL.FACT_SALES.ITEMID",
> >> >> > > >> >> >> >>       "table": "RETAIL.FACT_SALES",
> >> >> > > >> >> >> >>       "hierarchy": false,
> >> >> > > >> >> >> >>       "derived": null,
> >> >> > > >> >> >> >>       "column": [
> >> >> > > >> >> >> >>         "ITEMID"
> >> >> > > >> >> >> >>       ],
> >> >> > > >> >> >> >>       "id": 2
> >> >> > > >> >> >> >>     },
> >> >> > > >> >> >> >>     {
> >> >> > > >> >> >> >>       "name": "RETAIL.FACT_SALES.CUSTID",
> >> >> > > >> >> >> >>       "table": "RETAIL.FACT_SALES",
> >> >> > > >> >> >> >>       "hierarchy": false,
> >> >> > > >> >> >> >>       "derived": null,
> >> >> > > >> >> >> >>       "column": [
> >> >> > > >> >> >> >>         "CUSTID"
> >> >> > > >> >> >> >>       ],
> >> >> > > >> >> >> >>       "id": 3
> >> >> > > >> >> >> >>     }
> >> >> > > >> >> >> >>   ],
> >> >> > > >> >> >> >>   "measures": [
> >> >> > > >> >> >> >>     {
> >> >> > > >> >> >> >>       "id": 1,
> >> >> > > >> >> >> >>       "name": "_COUNT_",
> >> >> > > >> >> >> >>       "function": {
> >> >> > > >> >> >> >>         "expression": "COUNT",
> >> >> > > >> >> >> >>         "returntype": "bigint",
> >> >> > > >> >> >> >>         "parameter": {
> >> >> > > >> >> >> >>           "type": "constant",
> >> >> > > >> >> >> >>           "value": 1
> >> >> > > >> >> >> >>         }
> >> >> > > >> >> >> >>       }
> >> >> > > >> >> >> >>     }
> >> >> > > >> >> >> >>   ],
> >> >> > > >> >> >> >>   "rowkey": {
> >> >> > > >> >> >> >>     "rowkey_columns": [
> >> >> > > >> >> >> >>       {
> >> >> > > >> >> >> >>         "column": "STOREID",
> >> >> > > >> >> >> >>         "length": 0,
> >> >> > > >> >> >> >>         "dictionary": "true",
> >> >> > > >> >> >> >>         "mandatory": false
> >> >> > > >> >> >> >>       },
> >> >> > > >> >> >> >>       {
> >> >> > > >> >> >> >>         "column": "ITEMID",
> >> >> > > >> >> >> >>         "length": 0,
> >> >> > > >> >> >> >>         "dictionary": "true",
> >> >> > > >> >> >> >>         "mandatory": false
> >> >> > > >> >> >> >>       },
> >> >> > > >> >> >> >>       {
> >> >> > > >> >> >> >>         "column": "CUSTID",
> >> >> > > >> >> >> >>         "length": 0,
> >> >> > > >> >> >> >>         "dictionary": "true",
> >> >> > > >> >> >> >>         "mandatory": false
> >> >> > > >> >> >> >>       }
> >> >> > > >> >> >> >>     ],
> >> >> > > >> >> >> >>     "aggregation_groups": [
> >> >> > > >> >> >> >>       [
> >> >> > > >> >> >> >>         "STOREID",
> >> >> > > >> >> >> >>         "ITEMID",
> >> >> > > >> >> >> >>         "CUSTID"
> >> >> > > >> >> >> >>       ]
> >> >> > > >> >> >> >>     ]
> >> >> > > >> >> >> >>   },
> >> >> > > >> >> >> >>   "notify_list": [],
> >> >> > > >> >> >> >>   "capacity": "",
> >> >> > > >> >> >> >>   "hbase_mapping": {
> >> >> > > >> >> >> >>     "column_family": [
> >> >> > > >> >> >> >>       {
> >> >> > > >> >> >> >>         "name": "f1",
> >> >> > > >> >> >> >>         "columns": [
> >> >> > > >> >> >> >>           {
> >> >> > > >> >> >> >>             "qualifier": "m",
> >> >> > > >> >> >> >>             "measure_refs": [
> >> >> > > >> >> >> >>               "_COUNT_"
> >> >> > > >> >> >> >>             ]
> >> >> > > >> >> >> >>           }
> >> >> > > >> >> >> >>         ]
> >> >> > > >> >> >> >>       }
> >> >> > > >> >> >> >>     ]
> >> >> > > >> >> >> >>   },
> >> >> > > >> >> >> >>   "project": "RetailProject",
> >> >> > > >> >> >> >>   "model_name": "Retail_Cube"
> >> >> > > >> >> >> >> }
> >> >> > > >> >> >> >>
> >> >> > > >> >> >> >>Regards,
> >> >> > > >> >> >> >>Santosh Akhilesh
> >> >> > > >> >> >> >>Bangalore R&D
> >> >> > > >> >> >> >>HUAWEI TECHNOLOGIES CO.,LTD.
> >> >> > > >> >> >> >>
> >> >> > > >> >> >> >>www.huawei.com
> >> >> > > >> >> >>
> >> >>>>----------------------------------------------------------
> >> >> > > >> >> >> ---------------
> >> >> > > >> >> >> >>-
> >> >> > > >> >> >>
> >> >>>>-----------------------------------------------------------
> >> >> > > >> >> >> >>This e-mail and its attachments contain confidential
> >> >> > information
> >> >> > > >> >>from
> >> >> > > >> >> >> >>HUAWEI, which
> >> >> > > >> >> >> >>is intended only for the person or entity whose
> >>address
> >> >>is
> >> >> > > >>listed
> >> >> > > >> >> >>above.
> >> >> > > >> >> >> >>Any use of the
> >> >> > > >> >> >> >>information contained herein in any way (including,
> >>but
> >> >>not
> >> >> > > >>limited
> >> >> > > >> >> >>to,
> >> >> > > >> >> >> >>total or partial
> >> >> > > >> >> >> >>disclosure, reproduction, or dissemination) by persons
> >> >>other
> >> >> > > >>than
> >> >> > > >> >>the
> >> >> > > >> >> >> >>intended
> >> >> > > >> >> >> >>recipient(s) is prohibited. If you receive this
> >>e-mail in
> >> >> > error,
> >> >> > > >> >> >>please
> >> >> > > >> >> >> >>notify the sender by
> >> >> > > >> >> >> >>phone or email immediately and delete it!
> >> >> > > >> >> >> >>
> >> >> > > >> >> >> >>________________________________________
> >> >> > > >> >> >> >>From: Shi, Shaofeng [[email protected]]
> >> >> > > >> >> >> >>Sent: Thursday, February 26, 2015 7:01 AM
> >> >> > > >> >> >> >>To: [email protected]
> >> >> > > >> >> >> >>Subject: Re: Error while making cube & Measure option
> >>is
> >> >>not
> >> >> > > >> >> >>responding
> >> >> > > >> >> >> >>on GUI
> >> >> > > >> >> >> >>
> >> >> > > >> >> >> >>Hi Santosh,
> >> >> > > >> >> >> >>
> >> >> > > >> >> >> >>It looks like hadoop failed to execute some shell
> >> >>command in
> >> >> > the
> >> >> > > >> >> >> >>container; You need dive into hadoop to see what¹s the
> >> >> > concrete
> >> >> > > >> >>error.
> >> >> > > >> >> >> >>You
> >> >> > > >> >> >> >>can use yarn logs command to fetch all logs:
> >> >> > > >> >> >> >>
> >> >> > > >> >> >> >>yarn logs -applicationId <app_id>
> >> >> > > >> >> >> >>
> >> >> > > >> >> >> >>
> >> >> > > >> >> >> >>On 2/25/15, 7:39 PM, "Santosh Akhilesh"
> >> >> > > >><[email protected]
> >> >> > > >> >
> >> >> > > >> >> >> >>wrote:
> >> >> > > >> >> >> >>
> >> >> > > >> >> >> >>>Hi Luke / Shaofeng ,
> >> >> > > >> >> >> >>>           Can you please help me to check this
> >>issue.
> >> >> > > >> >> >> >>>Regards,
> >> >> > > >> >> >> >>>Santosh Akhilesh
> >> >> > > >> >> >> >>>
> >> >> > > >> >> >> >>>On Tue, Feb 24, 2015 at 10:41 PM, Santosh Akhilesh <
> >> >> > > >> >> >> >>>[email protected]> wrote:
> >> >> > > >> >> >> >>>
> >> >> > > >> >> >> >>>> Hi All ,
> >> >> > > >> >> >> >>>>         is it because of following error in map
> >>reduce
> >> >> job
> >> >> > ?
> >> >> > > >> >>what
> >> >> > > >> >> >> >>>>could
> >> >> > > >> >> >> >>>>be
> >> >> > > >> >> >> >>>> way to resolve this , a google search says that its
> >> >>issue
> >> >> > of
> >> >> > > >> >>Yarn
> >> >> > > >> >> >> >>>>class
> >> >> > > >> >> >> >>>> path , but I am not sure what it is ?
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> Kylin Hive Column Cardinality Job
> >> >>table=RETAIL.FACT_SALES
> >> >> > > >> >> >> >>>> output=/tmp/cardinality/RETAIL.FACT_SALES
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> Application application_1424791969399_0008 failed 2
> >> >>times
> >> >> > due
> >> >> > > >> >>to AM
> >> >> > > >> >> >> >>>> Container for appattempt_1424791969399_0008_000002
> >> >>exited
> >> >> > > >>with
> >> >> > > >> >> >> >>>>exitCode: 1
> >> >> > > >> >> >> >>>> For more detailed output, check application
> >>tracking
> >> >> page:
> >> >> > > >> >> >> >>>>
> >> >> > > >>http://santosh:8088/proxy/application_1424791969399_0008/Then,
> >> >> > > >> >> >>click
> >> >> > > >> >> >> >>>>on
> >> >> > > >> >> >> >>>> links to logs of each attempt.
> >> >> > > >> >> >> >>>> Diagnostics: Exception from container-launch.
> >> >> > > >> >> >> >>>> Container id:
> >>container_1424791969399_0008_02_000001
> >> >> > > >> >> >> >>>> Exit code: 1
> >> >> > > >> >> >> >>>> Stack trace: ExitCodeException exitCode=1:
> >> >> > > >> >> >> >>>> at
> >> >> org.apache.hadoop.util.Shell.runCommand(Shell.java:538)
> >> >> > > >> >> >> >>>> at org.apache.hadoop.util.Shell.
> run(Shell.java:455)
> >> >> > > >> >> >> >>>> at
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >>
> >> >>>>>>org.apache.hadoop.util.Shell$ShellCommandExecutor.execut
> >> >> > > >> >> >> e(Shell.java:71
> >> >> > > >> >> >> >>>>5
> >> >> > > >> >> >> >>>>)
> >> >> > > >> >> >> >>>> at
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >>
> >> >>>>>>org.apache.hadoop.yarn.server.nodemanager.DefaultContain
> >> >> > > >> >> >> erExecutor.laun
> >> >> > > >> >> >> >>>>c
> >> >> > > >> >> >> >>>>h
> >> >> > > >> >> >> >>>>Container(DefaultContainerExecutor.java:211)
> >> >> > > >> >> >> >>>> at
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >>
> >> >>>>>>org.apache.hadoop.yarn.server.nodemanager.containermanag
> >> >> > > >> >> >> er.launcher.Con
> >> >> > > >> >> >> >>>>t
> >> >> > > >> >> >> >>>>a
> >> >> > > >> >> >> >>>>inerLaunch.call(ContainerLaunch.java:302)
> >> >> > > >> >> >> >>>> at
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >>
> >> >>>>>>org.apache.hadoop.yarn.server.nodemanager.containermanag
> >> >> > > >> >> >> er.launcher.Con
> >> >> > > >> >> >> >>>>t
> >> >> > > >> >> >> >>>>a
> >> >> > > >> >> >> >>>>inerLaunch.call(ContainerLaunch.java:82)
> >> >> > > >> >> >> >>>> at
> >> >> java.util.concurrent.FutureTask.run(FutureTask.java:262)
> >> >> > > >> >> >> >>>> at
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >>
> >> >>>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread
> >> >> > > >> >> >> PoolExecutor.ja
> >> >> > > >> >> >> >>>>v
> >> >> > > >> >> >> >>>>a
> >> >> > > >> >> >> >>>>:1145)
> >> >> > > >> >> >> >>>> at
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >>
> >> >>>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa
> >> >> > > >> >> >> dPoolExecutor.j
> >> >> > > >> >> >> >>>>a
> >> >> > > >> >> >> >>>>v
> >> >> > > >> >> >> >>>>a:615)
> >> >> > > >> >> >> >>>> at java.lang.Thread.run(Thread.java:745)
> >> >> > > >> >> >> >>>> Container exited with a non-zero exit code 1
> >> >> > > >> >> >> >>>> Failing this attempt. Failing the application.
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> ---------- Forwarded message ----------
> >> >> > > >> >> >> >>>> From: Santoshakhilesh <[email protected]
> >
> >> >> > > >> >> >> >>>> Date: Tue, Feb 24, 2015 at 7:41 PM
> >> >> > > >> >> >> >>>> Subject: FW: Error while making cube & Measure
> >>option
> >> >>is
> >> >> > not
> >> >> > > >> >> >> >>>>responding
> >> >> > > >> >> >> >>>>on
> >> >> > > >> >> >> >>>> GUI
> >> >> > > >> >> >> >>>> To: "[email protected]"
> >> >> > > >> >> >><[email protected]>
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> hi ,
> >> >> > > >> >> >> >>>>    please someone give me a hand to resolve this
> >> >>issue ,
> >> >> > > >>thanks.
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> Regards,
> >> >> > > >> >> >> >>>> Santosh Akhilesh
> >> >> > > >> >> >> >>>> Bangalore R&D
> >> >> > > >> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD.
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> www.huawei.com
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >>
> >> >>>>>>--------------------------------------------------------
> >> >> > > >> >> >> ---------------
> >> >> > > >> >> >> >>>>-
> >> >> > > >> >> >> >>>>-
> >> >> > > >> >> >>
> >> >> >>>>------------------------------------------------------------
> >> >> > > >> >> >> >>>> This e-mail and its attachments contain
> >>confidential
> >> >> > > >>information
> >> >> > > >> >> >>from
> >> >> > > >> >> >> >>>> HUAWEI, which
> >> >> > > >> >> >> >>>> is intended only for the person or entity whose
> >> >>address
> >> >> is
> >> >> > > >> >>listed
> >> >> > > >> >> >> >>>>above.
> >> >> > > >> >> >> >>>> Any use of the
> >> >> > > >> >> >> >>>> information contained herein in any way (including,
> >> >>but
> >> >> not
> >> >> > > >> >>limited
> >> >> > > >> >> >> >>>>to,
> >> >> > > >> >> >> >>>> total or partial
> >> >> > > >> >> >> >>>> disclosure, reproduction, or dissemination) by
> >>persons
> >> >> > other
> >> >> > > >> >>than
> >> >> > > >> >> >>the
> >> >> > > >> >> >> >>>> intended
> >> >> > > >> >> >> >>>> recipient(s) is prohibited. If you receive this
> >> >>e-mail in
> >> >> > > >>error,
> >> >> > > >> >> >> >>>>please
> >> >> > > >> >> >> >>>> notify the sender by
> >> >> > > >> >> >> >>>> phone or email immediately and delete it!
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> ________________________________________
> >> >> > > >> >> >> >>>> From: Santoshakhilesh [[email protected]
> ]
> >> >> > > >> >> >> >>>> Sent: Tuesday, February 24, 2015 12:55 PM
> >> >> > > >> >> >> >>>> To: [email protected]
> >> >> > > >> >> >> >>>> Cc: Kulbhushan Rana
> >> >> > > >> >> >> >>>> Subject: FW: Error while making cube & Measure
> >>option
> >> >>is
> >> >> > not
> >> >> > > >> >> >> >>>>responding
> >> >> > > >> >> >> >>>>on
> >> >> > > >> >> >> >>>> GUI
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> 2. If  I ignore and continue and try to save the
> >>cube
> >> >>I
> >> >> get
> >> >> > > >>an
> >> >> > > >> >> >> >>>>exception
> >> >> > > >> >> >> >>>> in Kylin.log , I have checked the path is set
> >> >>correctly
> >> >> and
> >> >> > > >> >> >> >>>>HCatInputFormat
> >> >> > > >> >> >> >>>> this file is present in
> >>hive-hcatalog-core-0.14.0.jar
> >> >>.
> >> >> > > >>Please
> >> >> > > >> >>let
> >> >> > > >> >> >>me
> >> >> > > >> >> >> >>>>know
> >> >> > > >> >> >> >>>> what can I do to resolve this ?
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>>  -- This was path issue , now no more exception in
> >> >> > kylin.log
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> But saveing cube still fails with error. And still
> >> >>can't
> >> >> > add
> >> >> > > >> >> >>measures.
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> Error Message
> >> >> > > >> >> >> >>>> Failed to take action.
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> In log I can find no exception. Following is the
> >>last
> >> >>log
> >> >> > in
> >> >> > > >> >> >>kylin.log
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> [pool-3-thread-1]:[2015-02-24
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >>
> >> >>>>>>20:47:15,613][INFO][org.apache.kylin.job.impl.threadpool
> >> >> > > >> >> >> .DefaultSchedul
> >> >> > > >> >> >> >>>>e
> >> >> > > >> >> >> >>>>r
> >> >> > > >> >> >> >>>>$FetcherRunner.run(DefaultScheduler.java:117)]
> >> >> > > >> >> >> >>>> - Job Fetcher: 0 running, 0 actual running, 0
> >>ready, 6
> >> >> > others
> >> >> > > >> >> >> >>>> [http-bio-7070-exec-2]:[2015-02-24
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >>
> >> >>>>>>20:47:51,610][DEBUG][org.apache.kylin.rest.controller.Cu
> >> >> > > >> >> >> beController.de
> >> >> > > >> >> >> >>>>s
> >> >> > > >> >> >> >>>>e
> >> >> > > >> >> >> >>>>rializeDataModelDesc(CubeController.java:459)]
> >> >> > > >> >> >> >>>> - Saving cube {
> >> >> > > >> >> >> >>>>   "name": "",
> >> >> > > >> >> >> >>>>   "fact_table": "RETAIL.FACT_SALES",
> >> >> > > >> >> >> >>>>   "lookups": [],
> >> >> > > >> >> >> >>>>   "filter_condition": "",
> >> >> > > >> >> >> >>>>   "capacity": "SMALL",
> >> >> > > >> >> >> >>>>   "partition_desc": {
> >> >> > > >> >> >> >>>>     "partition_date_column": "",
> >> >> > > >> >> >> >>>>     "partition_date_start": 0,
> >> >> > > >> >> >> >>>>     "partition_type": "APPEND"
> >> >> > > >> >> >> >>>>   },
> >> >> > > >> >> >> >>>>   "last_modified": 0
> >> >> > > >> >> >> >>>> }
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> local access logs all with 200 , so seems ok.
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:46:56 +0800] "GET
> >> >> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
> >> >> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:07 +0800] "GET
> >> >> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
> >> >> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:27 +0800] "GET
> >> >> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
> >> >> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:28 +0800] "GET
> >> >> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
> >> >> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:34 +0800] "GET
> >> >> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
> >> >> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:48 +0800] "GET
> >> >> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246
> >> >> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:51 +0800]
> >>"POST
> >> >> > > >> >> >>/kylin/api/cubes
> >> >> > > >> >> >> >>>> HTTP/1.1" 200 701
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> Regards,
> >> >> > > >> >> >> >>>> Santosh Akhilesh
> >> >> > > >> >> >> >>>> Bangalore R&D
> >> >> > > >> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD.
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> www.huawei.com
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >>
> >> >>>>>>--------------------------------------------------------
> >> >> > > >> >> >> ---------------
> >> >> > > >> >> >> >>>>-
> >> >> > > >> >> >> >>>>-
> >> >> > > >> >> >>
> >> >> >>>>------------------------------------------------------------
> >> >> > > >> >> >> >>>> This e-mail and its attachments contain
> >>confidential
> >> >> > > >>information
> >> >> > > >> >> >>from
> >> >> > > >> >> >> >>>> HUAWEI, which
> >> >> > > >> >> >> >>>> is intended only for the person or entity whose
> >> >>address
> >> >> is
> >> >> > > >> >>listed
> >> >> > > >> >> >> >>>>above.
> >> >> > > >> >> >> >>>> Any use of the
> >> >> > > >> >> >> >>>> information contained herein in any way (including,
> >> >>but
> >> >> not
> >> >> > > >> >>limited
> >> >> > > >> >> >> >>>>to,
> >> >> > > >> >> >> >>>> total or partial
> >> >> > > >> >> >> >>>> disclosure, reproduction, or dissemination) by
> >>persons
> >> >> > other
> >> >> > > >> >>than
> >> >> > > >> >> >>the
> >> >> > > >> >> >> >>>> intended
> >> >> > > >> >> >> >>>> recipient(s) is prohibited. If you receive this
> >> >>e-mail in
> >> >> > > >>error,
> >> >> > > >> >> >> >>>>please
> >> >> > > >> >> >> >>>> notify the sender by
> >> >> > > >> >> >> >>>> phone or email immediately and delete it!
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> ________________________________________
> >> >> > > >> >> >> >>>> From: Santoshakhilesh [[email protected]
> ]
> >> >> > > >> >> >> >>>> Sent: Tuesday, February 24, 2015 12:09 PM
> >> >> > > >> >> >> >>>> To: [email protected]
> >> >> > > >> >> >> >>>> Cc: Kulbhushan Rana
> >> >> > > >> >> >> >>>> Subject: Error while making cube & Measure option
> >>is
> >> >>not
> >> >> > > >> >> >>responding on
> >> >> > > >> >> >> >>>>GUI
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> Hi All ,
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>>     I am building a simple cube for test and using
> >>the
> >> >> > binary
> >> >> > > >> >>build
> >> >> > > >> >> >> >>>>0.7.1
> >> >> > > >> >> >> >>>> . I have following hive tables with columns.
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> fact_sales:
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> storeid                 int
> >> >> > > >> >> >> >>>> itemid                  int
> >> >> > > >> >> >> >>>> custid                  int
> >> >> > > >> >> >> >>>> qty                     int
> >> >> > > >> >> >> >>>> price                   double
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> dim_customer
> >> >> > > >> >> >> >>>> custid                  int
> >> >> > > >> >> >> >>>> name                    string
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> dim_item
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> itemid                  int
> >> >> > > >> >> >> >>>> category                string
> >> >> > > >> >> >> >>>> brand                   string
> >> >> > > >> >> >> >>>> color                   string
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> dim_store
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> storeid                 int
> >> >> > > >> >> >> >>>> city                    string
> >> >> > > >> >> >> >>>> state                   string
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> Please help me to answer following issues;
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> 1. When I go to measure section and click on
> >>measure
> >> >> > option ,
> >> >> > > >> >> >>there is
> >> >> > > >> >> >> >>>>no
> >> >> > > >> >> >> >>>> response , I want add measure on qty and price with
> >> >>sum
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> 2. If  I ignore and continue and try to save the
> >>cube
> >> >>I
> >> >> get
> >> >> > > >>an
> >> >> > > >> >> >> >>>>exception
> >> >> > > >> >> >> >>>> in Kylin.log , I have checked the path is set
> >> >>correctly
> >> >> and
> >> >> > > >> >> >> >>>>HCatInputFormat
> >> >> > > >> >> >> >>>> this file is present in
> >>hive-hcatalog-core-0.14.0.jar
> >> >>.
> >> >> > > >>Please
> >> >> > > >> >>let
> >> >> > > >> >> >>me
> >> >> > > >> >> >> >>>>know
> >> >> > > >> >> >> >>>> what can I do to resolve this ?
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> 3. Also I have another question since this is a
> >>test
> >> >>and
> >> >> > > >>data is
> >> >> > > >> >> >>small
> >> >> > > >> >> >> >>>>I
> >> >> > > >> >> >> >>>> have not partitioned the fact table , is it ok to
> >>skip
> >> >> > > >>partition
> >> >> > > >> >> >>stage
> >> >> > > >> >> >> >>>> while cube build ?
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> Exception
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >> >>>> pool-4-thread-4]:[2015-02-24
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >>
> >> >>>>>>19:26:32,577][ERROR][org.apache.kylin.job.impl.threadpoo
> >> >> > > >> >> >> l.DefaultSchedu
> >> >> > > >> >> >> >>>>l
> >> >> > > >> >> >> >>>>e
> >> >> > > >> >> >> >>>>r$JobRunner.run(DefaultScheduler.java:134)]
> >> >> > > >> >> >> >>>> - ExecuteException
> >> >> job:c3532a6f-97ea-474a-b36a-218dd517cedb
> >> >> > > >> >> >> >>>> org.apache.kylin.job.exception.ExecuteException:
> >> >> > > >> >> >> >>>> org.apache.kylin.job.exception.ExecuteException:
> >> >> > > >> >> >> >>>> java.lang.NoClassDefFoundError:
> >> >> > > >> >> >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat
> >> >> > > >> >> >> >>>>  at
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >>
> >> >>>>>>org.apache.kylin.job.execution.AbstractExecutable.execut
> >> >> > > >> >> >> e(AbstractExecu
> >> >> > > >> >> >> >>>>t
> >> >> > > >> >> >> >>>>a
> >> >> > > >> >> >> >>>>ble.java:102)
> >> >> > > >> >> >> >>>>  at
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >>
> >> >>>>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$Jo
> >> >> > > >> >> >> bRunner.run(Def
> >> >> > > >> >> >> >>>>a
> >> >> > > >> >> >> >>>>u
> >> >> > > >> >> >> >>>>ltScheduler.java:132)
> >> >> > > >> >> >> >>>>  at
> >> >> > > >> >> >> >>>>
> >> >> > > >> >> >>
> >> >>>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread
> >> >> > > >> >> >> PoolExecutor.ja
> >> >> > > >

Reply via email to