In 0.6.x the packages are named with “com.kylinolap.xxx”, from 0.7 we renamed the package to “org.apache.kylin.xxx”; When you downgrade to 0.6, did you also replace the jar location with 0.6 ones in kylin.properties?
On 2/27/15, 3:13 PM, "Santoshakhilesh" <[email protected]> wrote: >Hi Shaofeng , > I have added my fact and dimension tables under default database >of hive. > Now stage 1 of Cube Build is ok. And there is failure at step2. > The map reduce job for the finding distinct columns of fact table >is error. Yarn log is as below. > Strangely this is class not found error. I have checked the >Kylin.properties and the jar is already set as below. >kylin. log has one exception connecting to linux/10.19.93.68 to >0.0.0.0:10020 > Please help me to give a clue , I am also trying to check meanwhile > >Thanks. >kylin property ># Temp folder in hdfs >kylin.hdfs.working.dir=/tmp ># Path to the local(relative to job engine) job jar, job engine will use >this jar >kylin.job.jar=/tmp/kylin/kylin-job-latest.jar > >Map Reduce error >---------------------------- >2015-02-27 20:24:25,262 FATAL [main] org.apache.hadoop.mapred.YarnChild: >Error running child : java.lang.NoClassDefFoundError: >com/kylinolap/common/mr/KylinMapper > at java.lang.ClassLoader.defineClass1(Native Method) > at java.lang.ClassLoader.defineClass(ClassLoader.java:800) > at >java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) > at java.net.URLClassLoader.defineClass(URLClassLoader.java:449) > at java.net.URLClassLoader.access$100(URLClassLoader.java:71) > at java.net.URLClassLoader$1.run(URLClassLoader.java:361) > at java.net.URLClassLoader$1.run(URLClassLoader.java:355) > at java.security.AccessController.doPrivileged(Native Method) > at java.net.URLClassLoader.findClass(URLClassLoader.java:354) > at java.lang.ClassLoader.loadClass(ClassLoader.java:425) > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) > at java.lang.ClassLoader.loadClass(ClassLoader.java:358) > at java.lang.Class.forName0(Native Method) > at java.lang.Class.forName(Class.java:274) > at >org.apache.hadoop.conf.Configuration.getClassByNameOrNull(Configuration.ja >va:2013) > >Kylin.log >QuartzScheduler_Worker-20]:[2015-02-27 >20:25:00,663][DEBUG][com.kylinolap.job.engine.JobFetcher.execute(JobFetche >r.java:60)] - 0 pending jobs >[QuartzScheduler_Worker-19]:[2015-02-27 >20:25:01,730][ERROR][com.kylinolap.job.cmd.JavaHadoopCmdOutput.updateJobCo >unter(JavaHadoopCmdOutput.java:176)] - java.io.IOException: >java.net.ConnectException: Call From linux/10.19.93.68 to 0.0.0.0:10020 >failed on connection exception: java.net.ConnectException: Connection >refused; For more details see: >http://wiki.apache.org/hadoop/ConnectionRefused >com.kylinolap.job.exception.JobException: java.io.IOException: >java.net.ConnectException: Call From linux/10.19.93.68 to 0.0.0.0:10020 >failed on connection exception: java.net.ConnectException: Connection >refused; For more details see: >http://wiki.apache.org/hadoop/ConnectionRefused > at >com.kylinolap.job.hadoop.AbstractHadoopJob.getCounters(AbstractHadoopJob.j >ava:289) > at >com.kylinolap.job.cmd.JavaHadoopCmdOutput.updateJobCounter(JavaHadoopCmdOu >tput.java:162) > at >com.kylinolap.job.cmd.JavaHadoopCmdOutput.getStatus(JavaHadoopCmdOutput.ja >va:85) > at >com.kylinolap.job.flow.AsyncJobFlowNode.execute(AsyncJobFlowNode.java:86) > at org.quartz.core.JobRunShell.run(JobRunShell.java:202) > at >org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java:5 >73) >Caused by: java.io.IOException: java.net.ConnectException: Call From >linux/10.19.93.68 to 0.0.0.0:10020 failed on connection exception: >java.net.ConnectException: Connection refused; For more details see: >http://wiki.apache.org/hadoop/ConnectionRefused > at >org.apache.hadoop.mapred.ClientServiceDelegate.invoke(ClientServiceDelegat >e.java:331) > at >org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounters(ClientServic >eDelegate.java:368) > at >org.apache.hadoop.mapred.YARNRunner.getJobCounters(YARNRunner.java:511) > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756) > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at >org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation. >java:1491) > at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753) > at >com.kylinolap.job.hadoop.AbstractHadoopJob.getCounters(AbstractHadoopJob.j >ava:287) > ... 5 more > >Regards, >Santosh Akhilesh >Bangalore R&D >HUAWEI TECHNOLOGIES CO.,LTD. > >www.huawei.com >-------------------------------------------------------------------------- >----------------------------------------------------------- >This e-mail and its attachments contain confidential information from >HUAWEI, which >is intended only for the person or entity whose address is listed above. >Any use of the >information contained herein in any way (including, but not limited to, >total or partial >disclosure, reproduction, or dissemination) by persons other than the >intended >recipient(s) is prohibited. If you receive this e-mail in error, please >notify the sender by >phone or email immediately and delete it! > >________________________________________ >From: Shi, Shaofeng [[email protected]] >Sent: Friday, February 27, 2015 8:01 AM >To: [email protected] >Subject: Re: Error while making cube & Measure option is not responding >on GUI > >In 0.6.x it only support tables in default database, this is a limitation; >The support for non-default tables will be released in 0.7; > >To bypass this issue for now, please copy the table to default database as >a workaround; > >On 2/27/15, 10:16 AM, "Santosh Akhilesh" <[email protected]> >wrote: > >>@Jason >>thanks , but now as suggested by Saofeng I m not using the inverted index >>brach as its not stable. >>I have switched back to 0.6 branch , in this branch yesterday night I >>could >>crete the cube successfully but there is issue while building it. I feel >>that at step 1 of cube build while creating flat table when command is >>issued to hive if the tables are not under default datbase flat table >>creation is failed and cube build fails. my fact and dimension tables are >>under a database called retail. >> >>@Saofeng - Can you please confirm this behavior ? Do I need to create the >>hive tables under default database? >> >>On Fri, Feb 27, 2015 at 7:32 AM, jason zhong <[email protected]> >>wrote: >> >>> @Santoshakhilesh >>> >>> 1. When I go to measure section and click on measure option , there is >>>no >>> response , I want add measure on qty and price with sum >>> --bug fixed on inverted-index branch >>> >>> >>> On Fri, Feb 27, 2015 at 3:03 AM, Santosh Akhilesh < >>> [email protected] >>> > wrote: >>> >>> > Hi Shaofeng , >>> > I have build the 0.6 version and now able to create the cube >>> > successfully. >>> > While building the cube , it fails at step1 with following >>>error. >>> > Table not found 'DIM_ITEM' >>> > the table exists , but its under retail data base and not under >>> > default database. >>> > does kylin require hive taables to be under default database ? I >>>see >>> > the flat table being created under default database. >>> > >>> > Logging initialized using configuration in >>> > >>> > >>> >>>jar:file:/home/santosh/work/frameworks/apache-hive-1.0.0/lib/hive-common >>>- >>>1.0.0.jar!/hive-log4j.properties >>> > SLF4J: Class path contains multiple SLF4J bindings. >>> > SLF4J: Found binding in >>> > >>> > >>> >>>[jar:file:/home/santosh/work/frameworks/hadoop-2.6.0/share/hadoop/common >>>/ >>>lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class] >>> > SLF4J: Found binding in >>> > >>> > >>> >>>[jar:file:/home/santosh/work/frameworks/apache-hive-1.0.0/lib/hive-jdbc- >>>1 >>>.0.0-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class] >>> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an >>> > explanation. >>> > SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] >>> > OK >>> > Time taken: 0.964 seconds >>> > OK >>> > Time taken: 0.948 seconds >>> > FAILED: SemanticException [Error 10001]: Line 12:11 Table not found >>> > 'DIM_ITEM' >>> > >>> > >>> > >>> > Command is as below. >>> > >>> > hive -e "DROP TABLE IF EXISTS >>> > >>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd44; >>> > CREATE EXTERNAL TABLE IF NOT EXISTS >>> > >>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd44 >>> > ( >>> > STOREID int >>> > ,ITEMID int >>> > ,CUSTID int >>> > ,QTY int >>> > ,AMOUNT double >>> > ) >>> > ROW FORMAT DELIMITED FIELDS TERMINATED BY '\177' >>> > STORED AS SEQUENCEFILE >>> > LOCATION >>> > >>> >>>'/tmp/kylin-8b30b29b-5f2c-4b63-8c0f-07d1f559dd44/kylin_intermediate_test >>>_ >>>FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd44'; >>> > SET hive.exec.compress.output=true; >>> > SET hive.auto.convert.join.noconditionaltask = true; >>> > SET hive.auto.convert.join.noconditionaltask.size = 300000000; >>> > INSERT OVERWRITE TABLE >>> > >>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd44 >>> > SELECT >>> > FACT_SALES.STOREID >>> > ,FACT_SALES.ITEMID >>> > ,FACT_SALES.CUSTID >>> > ,FACT_SALES.QTY >>> > ,FACT_SALES.AMOUNT >>> > FROM FACT_SALES >>> > INNER JOIN DIM_STORE >>> > ON FACT_SALES.STOREID = DIM_STORE.SROREID >>> > INNER JOIN DIM_ITEM >>> > ON FACT_SALES.ITEMID = DIM_ITEM.ITEMID >>> > INNER JOIN DIM_CUSTOMER >>> > ON FACT_SALES.CUSTID = DIM_CUSTOMER.CUSTID >>> > ; >>> > " >>> > >>> > >>> > >>> > On Thu, Feb 26, 2015 at 8:11 PM, Shi, Shaofeng <[email protected]> >>>wrote: >>> > >>> > > The 0.7.1 is test version, its package contains the “snapshot” >>>suffix; >>> we >>> > > will upload a new package there; Luke will also add a message there >>>to >>> > > avoid this confusion; >>> > > >>> > > Regarding the problem that you encountered, could you please open a >>> JIRA >>> > > ticket for tracking? Here is link of Apache JIRA: >>> > > >>> > > https://issues.apache.org/jira/secure/Dashboard.jspa >>> > > >>> > > >>> > > Thanks for the feedback! >>> > > >>> > > On 2/26/15, 10:21 PM, "Santosh Akhilesh" >>><[email protected]> >>> > > wrote: >>> > > >>> > > >Actually I see this being published on kylin webpage. >>> > > >http://kylin.incubator.apache.org/download/ >>> > > >I am using 0.7.1 inverted index branch binary distribution. >>> > > >If this is not stable please give me the link of stable branch I >>>would >>> > try >>> > > >building and testing tonight. >>> > > >On Thu, 26 Feb 2015 at 7:30 pm, Shi, Shaofeng <[email protected]> >>> wrote: >>> > > > >>> > > >> Hi Santosh, it is not recommended to use the dev code branch >>> > (actually I >>> > > >> don’t know how you get the v0.7.x build and what’s the detail >>> version >>> > of >>> > > >> that; each day we submit many changes to that); >>> > > >> >>> > > >> The options are 1) switch back to latest release v0.6.5; or 2) >>>wait >>> > for >>> > > >> the formal release of 0.7, that should be in March; Otherwise, >>>we >>> > > >>couldn’t >>> > > >> ensure there is no new problems come out in your next steps; >>> > > >> >>> > > >> On 2/26/15, 5:39 PM, "Santosh Akhilesh" >>><[email protected]> >>> > > >>wrote: >>> > > >> >>> > > >> >Hi Shaofeng >>> > > >> >So what do you suggest , how should I proceed further with this >>> > > >>release? >>> > > >> >Will there be a patch? Any alternate way I can create cube? >>> > > >> >Please suggest. >>> > > >> >Regards >>> > > >> >Santosh >>> > > >> >On Thu, 26 Feb 2015 at 3:04 pm, Shi, Shaofeng >>><[email protected]> >>> > > wrote: >>> > > >> > >>> > > >> >> Hi Santosh, >>> > > >> >> >>> > > >> >> 0.7.1 hasn’t been formally released; from 0.6.x to 0.7.x we >>>have >>> > > >> >>metadata >>> > > >> >> structure change; While, the web UI (cube wizard) for this >>>change >>> > > >>hasn’t >>> > > >> >> been stabilized; So it is not strange that you got trouble >>>when >>> > > >>saving >>> > > >> >>the >>> > > >> >> cube; >>> > > >> >> >>> > > >> >> @Jason, any idea about the JS error? >>> > > >> >> >>> > > >> >> On 2/26/15, 5:08 PM, "Santosh Akhilesh" < >>> [email protected] >>> > > >>> > > >> >>wrote: >>> > > >> >> >>> > > >> >> >Hi Shaofeng, >>> > > >> >> > >>> > > >> >> >I am using the binary distribution 0.7.1. I have not been >>>able >>> to >>> > > >>save >>> > > >> >> >cube >>> > > >> >> >even once. I have tried creating new project and from local >>> > machine >>> > > >>and >>> > > >> >> >server machine. But I am always stuck with this error. I am >>> never >>> > > >> >>allowed >>> > > >> >> >to add measures and never been able to save the cube. I also >>>see >>> > the >>> > > >> >> >kylin.log and it always tries to save cube with append mode. >>>One >>> > > >>thing >>> > > >> >>I >>> > > >> >> >need to tell that at partition stage since I don't have a >>>big >>> fact >>> > > >> >>table >>> > > >> >> >now I have not partititioned the fact table and I skip this >>> step. >>> > > >>Does >>> > > >> >> >this >>> > > >> >> >have affect in saving the cube. Is this because some >>>metadata is >>> > > >> >>available >>> > > >> >> >and it tries to modify the cube? I am using latest Hadoop >>>2.6.6. >>> > Yes >>> > > >> >>kylin >>> > > >> >> >propert I have not added the jar. I will add them and check. >>>But >>> > > >>cube >>> > > >> >> >creation failure is really puzzling me. I could see no error >>> logs >>> > in >>> > > >> >> >kylin.log. >>> > > >> >> >Regards >>> > > >> >> >Santosh >>> > > >> >> >On Thu, 26 Feb 2015 at 1:40 pm, Shi, Shaofeng >>><[email protected] >>> > >>> > > >> wrote: >>> > > >> >> > >>> > > >> >> >> Which version or code branch are you using? I assume >>>you’re >>> > using >>> > > >>the >>> > > >> >> >> stable version from master; Seems you’re trying to edit an >>> > > >>existing >>> > > >> >>cube >>> > > >> >> >> to add new measurement, try refresh your browser's cache; >>>if >>> it >>> > > >>still >>> > > >> >> >> couldn’t be saved, try to create a new cube; >>> > > >> >> >> >>> > > >> >> >> The two error traces in tomcat need be taken care: >>> > > >> >> >> >>> > > >> >> >> 1) java.lang.NoClassDefFoundError: >>> > > >> >> >>org/apache/kylin/common/mr/KylinMapper >>> > > >> >> >> Please check kylin.properties file, making sure >>>the >>> > > >> >> >>“kylin.job.jar” >>> > > >> >> >> points to a right jar file; It will be loaded in >>>Map-reduce; >>> > > >> >> >> >>> > > >> >> >> 2) java.lang.IllegalArgumentException: No enum constant >>> > > >> >> >> org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS >>> > > >> >> >> This indicates your hadoop version might be old; >>> Please >>> > > >>check >>> > > >> >> >>and >>> > > >> >> >> ensure >>> > > >> >> >> hadoop version is 2.2 or above. >>> > > >> >> >> >>> > > >> >> >> On 2/26/15, 3:21 PM, "Santoshakhilesh" >>> > > >><[email protected]> >>> > > >> >> >> wrote: >>> > > >> >> >> >>> > > >> >> >> >Hi Shaofeng , >>> > > >> >> >> > >>> > > >> >> >> > I am using chrome , When I click on button to add >>> measures , >>> > > >> >> >>following >>> > > >> >> >> >is error on chrome console. When I try to save the cube >>>there >>> > is >>> > > >>no >>> > > >> >> >>error >>> > > >> >> >> >in console. I just get a error dialog saying failed to >>>take >>> > > >>action >>> > > >> >>and >>> > > >> >> >> >gives me the JSON cube schema. >>> > > >> >> >> > >>> > > >> >> >> >Error on chrome debug console is as below; >>> > > >> >> >> > >>> > > >> >> >> > ReferenceError: CubeDescModel is not defined >>> > > >> >> >> > at h.$scope.addNewMeasure (scripts.min.0.js:15984) >>> > > >> >> >> > at scripts.min.0.js:180 >>> > > >> >> >> > at scripts.min.0.js:197 >>> > > >> >> >> > at h.$eval (scripts.min.0.js:119) >>> > > >> >> >> > at h.$apply (scripts.min.0.js:119) >>> > > >> >> >> > at HTMLButtonElement.<anonymous> >>>(scripts.min.0.js:197) >>> > > >> >> >> > at HTMLButtonElement.m.event.dispatch >>> (scripts.min.0.js:3) >>> > > >> >> >> > at HTMLButtonElement.r.handle >>> > > >> >> >> >(scripts.min.0.js:3)scripts.min.0.js:100 (anonymous >>> > > >> >> >> >function)scripts.min.0.js:77 (anonymous >>> > > >> >>function)scripts.min.0.js:119 >>> > > >> >> >> >h.$applyscripts.min.0.js:197 (anonymous >>> > > >>function)scripts.min.0.js:3 >>> > > >> >> >> >m.event.dispatchscripts.min.0.js:3 r.handle >>> > > >> >> >> > >>> > > >> >> >> > About the hive table import , I got pass the run shell >>> > command >>> > > >> >> >> >exception but it still fails the hadoop log is; >>> > > >> >> >> >2015-02-26 20:46:48,332 INFO [main] >>>org.apache.hadoop.mapred. >>> > > >> >> YarnChild: >>> > > >> >> >> >mapreduce.cluster.local.dir for child: >>> > > >> >> >> >>>>/tmp/hadoop-root/nm-local-dir/usercache/root/appcache/appli >>> > > >> >> >> cation_14249530 >>> > > >> >> >> >91340_0002 >>> > > >> >> >> >2015-02-26 20:46:48,776 INFO [main] >>> > > >> >> >> >org.apache.hadoop.conf.Configuration.deprecation: >>>session.id >>> > is >>> > > >> >> >> >deprecated. Instead, use dfs.metrics.session-id >>> > > >> >> >> >2015-02-26 20:46:49,310 INFO [main] >>> > > >>org.apache.hadoop.mapred.Task: >>> > > >> >> >>Using >>> > > >> >> >> >ResourceCalculatorProcessTree : [ ] >>> > > >> >> >> >2015-02-26 20:46:49,386 FATAL [main] >>> > > >> >> >>org.apache.hadoop.mapred.YarnChild: >>> > > >> >> >> >Error running child : java.lang.NoClassDefFoundError: >>> > > >> >> >> >org/apache/kylin/common/mr/KylinMapper >>> > > >> >> >> > at java.lang.ClassLoader.defineClass1(Native Method) >>> > > >> >> >> > at >>>java.lang.ClassLoader.defineClass(ClassLoader.java:800) >>> > > >> >> >> > at >>> > > >> >> >> >java.security.SecureClassLoader.defineClass( >>> > > >> >> SecureClassLoader.java:142) >>> > > >> >> >> > at >>> > java.net.URLClassLoader.defineClass(URLClassLoader.java:449) >>> > > >> >> >> > at >>> java.net.URLClassLoader.access$100(URLClassLoader.java:71) >>> > > >> >> >> > at >>>java.net.URLClassLoader$1.run(URLClassLoader.java:361) >>> > > >> >> >> > at >>>java.net.URLClassLoader$1.run(URLClassLoader.java:355) >>> > > >> >> >> > at java.security.AccessController.doPrivileged(Native >>> Method) >>> > > >> >> >> > at >>> java.net.URLClassLoader.findClass(URLClassLoader.java:354) >>> > > >> >> >> > at java.lang.ClassLoader.loadClass(ClassLoader.java:425) >>> > > >> >> >> > at >>> > sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) >>> > > >> >> >> > >>> > > >> >> >> >tomcat logs: >>> > > >> >> >> >usage: HiveColumnCardinalityJob >>> > > >> >> >> > -output <path> Output path >>> > > >> >> >> > -table <table name> The hive table name >>> > > >> >> >> >[pool-4-thread-2]:[2015-02-26 >>> > > >> >> >> >>>>20:47:49,936][ERROR][org.apache.kylin.job.common.HadoopShel >>> > > >> >> >> lExecutable.doW >>> > > >> >> >> >ork(HadoopShellExecutable.java:64)] - error execute >>> > > >> >> >> >>> > >HadoopShellExecutable{id=d4730d26-7fe6-412e-9841-3288ab362c5b-00, >>> > > >> >> >> >name=null, state=RUNNING} >>> > > >> >> >> >java.lang.IllegalArgumentException: No enum constant >>> > > >> >> >> >org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS >>> > > >> >> >> > at java.lang.Enum.valueOf(Enum.java:236) >>> > > >> >> >> > at >>> > > >> >> >> >>> > > >>>>>>org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.valueOf( >>> > > >> >> >> Framewo >>> > > >> >> >> >rkCounterGroup.java:148) >>> > > >> >> >> > at >>> > > >> >> >> >>>>org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup. >>> > > >> >> >> findCounter(Fra >>> > > >> >> >> >meworkCounterGroup.java:182) >>> > > >> >> >> > at >>> > > >> >> >> >>>>org.apache.hadoop.mapreduce.counters.AbstractCounters.findC >>> > > >> >> >> ounter(Abstract >>> > > >> >> >> >Counters.java:154) >>> > > >> >> >> > at >>> > > >> >> >> >>>>org.apache.hadoop.mapreduce.TypeConverter.fromYarn(TypeConv >>> > > >> >> >> erter.java:240) >>> > > >> >> >> > at >>> > > >> >> >> >>>>org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounte >>> > > >> >> >> rs(ClientServic >>> > > >> >> >> >eDelegate.java:370) >>> > > >> >> >> > at >>> > > >> >> >> >org.apache.hadoop.mapred.YARNRunner.getJobCounters( >>> > > >> >> YARNRunner.java:511) >>> > > >> >> >> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756) >>> > > >> >> >> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753) >>> > > >> >> >> > at java.security.AccessController.doPrivileged(Native >>> Method) >>> > > >> >> >> > at javax.security.auth.Subject.doAs(Subject.java:415) >>> > > >> >> >> > at >>> > > >> >> >> >>>>org.apache.hadoop.security.UserGroupInformation.doAs(UserGr >>> > > >> >> >> oupInformation. >>> > > >> >> >> >java:1491) >>> > > >> >> >> > at >>>org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753) >>> > > >> >> >> > at >>> > > >> >>>>>org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1361) >>> > > >> >> >> > at >>>org.apache.hadoop.mapreduce.Job.waitForCompletion(Job. >>> > > >> java:1289) >>> > > >> >> >> > at >>> > > >> >> >> >>>>org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForComple >>> > > >> >> >> tion(AbstractHa >>> > > >> >> >> >doopJob.java:134) >>> > > >> >> >> > at >>> > > >> >> >> >>> > > >> >>> > >>>>>>org.apache.kylin.job.hadoop.cardinality.HiveColumnCardinalityJob.run( >>> > > >> >> >> HiveC >>> > > >> >> >> >olumnCardinalityJob.java:114) >>> > > >> >> >> > at >>>org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) >>> > > >> >> >> > at >>>org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84) >>> > > >> >> >> > at >>> > > >> >> >> >>>>org.apache.kylin.job.common.HadoopShellExecutable.doWork(Ha >>> > > >> >> >> doopShellExecut >>> > > >> >> >> >able.java:62) >>> > > >> >> >> > at >>> > > >> >> >> >>>>org.apache.kylin.job.execution.AbstractExecutable.execute(A >>> > > >> >> >> bstractExecutab >>> > > >> >> >> >le.java:99) >>> > > >> >> >> > at >>> > > >> >> >> >>>>org.apache.kylin.job.execution.DefaultChainedExecutable.doW >>> > > >> >> >> ork(DefaultChai >>> > > >> >> >> >nedExecutable.java:50) >>> > > >> >> >> > at >>> > > >> >> >> >>>>org.apache.kylin.job.execution.AbstractExecutable.execute(A >>> > > >> >> >> bstractExecutab >>> > > >> >> >> >le.java:99) >>> > > >> >> >> > at >>> > > >> >> >> >>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRu >>> > > >> >> >> nner.run(Defaul >>> > > >> >> >> >tScheduler.java:132) >>> > > >> >> >> > at >>> > > >> >> >> >>>>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoo >>> > > >> >> >> lExecutor.java: >>> > > >> >> >> >1145) >>> > > >> >> >> > at >>> > > >> >> >> >>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPo >>> > > >> >> >> olExecutor.java >>> > > >> >> >> >:615) >>> > > >> >> >> > at java.lang.Thread.run(Thread.java:745) >>> > > >> >> >> > >>> > > >> >> >> >Regards, >>> > > >> >> >> >Santosh Akhilesh >>> > > >> >> >> >Bangalore R&D >>> > > >> >> >> >HUAWEI TECHNOLOGIES CO.,LTD. >>> > > >> >> >> > >>> > > >> >> >> >www.huawei.com >>> > > >> >> >> >>>>----------------------------------------------------------- >>> > > >> >> >> --------------- >>> > > >> >> >> >>>>----------------------------------------------------------- >>> > > >> >> >> >This e-mail and its attachments contain confidential >>> > information >>> > > >> >>from >>> > > >> >> >> >HUAWEI, which >>> > > >> >> >> >is intended only for the person or entity whose address >>>is >>> > listed >>> > > >> >> >>above. >>> > > >> >> >> >Any use of the >>> > > >> >> >> >information contained herein in any way (including, but >>>not >>> > > >>limited >>> > > >> >>to, >>> > > >> >> >> >total or partial >>> > > >> >> >> >disclosure, reproduction, or dissemination) by persons >>>other >>> > than >>> > > >> >>the >>> > > >> >> >> >intended >>> > > >> >> >> >recipient(s) is prohibited. If you receive this e-mail in >>> > error, >>> > > >> >>please >>> > > >> >> >> >notify the sender by >>> > > >> >> >> >phone or email immediately and delete it! >>> > > >> >> >> > >>> > > >> >> >> >________________________________________ >>> > > >> >> >> >From: Shi, Shaofeng [[email protected]] >>> > > >> >> >> >Sent: Thursday, February 26, 2015 11:32 AM >>> > > >> >> >> >To: [email protected] >>> > > >> >> >> >Cc: Kulbhushan Rana >>> > > >> >> >> >Subject: Re: Error while making cube & Measure option is >>>not >>> > > >> >>responding >>> > > >> >> >> >on GUI >>> > > >> >> >> > >>> > > >> >> >> >Hi Santosh, hive table importing issue should not impact >>>on >>> > cube >>> > > >> >> >>saving. >>> > > >> >> >> > >>> > > >> >> >> >If you couldn’t save the cube, firstly please check >>>whether >>> > > >>there is >>> > > >> >> >>error >>> > > >> >> >> >in the tomcat’s log; If not please check your web >>>browser; We >>> > > >> >>suggest >>> > > >> >> >>use >>> > > >> >> >> >Firefox (with firebug add-on) or Chrome, open the JS >>>console >>> > > >>(press >>> > > >> >> >>F12) >>> > > >> >> >> >and then operate web UI, check whether there is any error >>> > > >>reported >>> > > >> >>in >>> > > >> >> >> >browser. >>> > > >> >> >> > >>> > > >> >> >> > >>> > > >> >> >> >On 2/26/15, 1:08 PM, "Santoshakhilesh" >>> > > >><[email protected] >>> > > >> > >>> > > >> >> >> >wrote: >>> > > >> >> >> > >>> > > >> >> >> >>Hi Shaofeng , >>> > > >> >> >> >> Thanks for replying. >>> > > >> >> >> >> Yes I am checking the yarn exception, But I find that >>> this >>> > > >>error >>> > > >> >> >>comes >>> > > >> >> >> >>while importing the hive table to kylin. >>> > > >> >> >> >> Even if this error comes , hive tables is exported >>> > > >>successfully >>> > > >> >>in >>> > > >> >> >> >>kylin. Is this the reason why cube saving has failed ? >>> > > >> >> >> >> >>> > > >> >> >> >> Next step when I go on creating the cube for >>>following >>> > > >>schema I >>> > > >> >> >>get >>> > > >> >> >> >>error at last step while saving and I am unable to add >>>any >>> > > >> >>measures , >>> > > >> >> >> >>clicking on measure option just dont pop up any dialog, >>> > > >> >> >> >> >>> > > >> >> >> >>I am using star schema with fact_sales as fact table and >>> dim_* >>> > > >>as >>> > > >> >> >> >>dimension tables. >>> > > >> >> >> >> >>> > > >> >> >> >> fact_sales: >>> > > >> >> >> >> storeid int >>> > > >> >> >> >> itemid int >>> > > >> >> >> >> custid int >>> > > >> >> >> >> qty int >>> > > >> >> >> >> price double >>> > > >> >> >> >> >>> > > >> >> >> >>dim_customer >>> > > >> >> >> >> custid int >>> > > >> >> >> >> name string >>> > > >> >> >> >> >>> > > >> >> >> >> dim_item >>> > > >> >> >> >> itemid int >>> > > >> >> >> >> category string >>> > > >> >> >> >> brand string >>> > > >> >> >> >> color string >>> > > >> >> >> >> >>> > > >> >> >> >>dim_store >>> > > >> >> >> >> storeid int >>> > > >> >> >> >> city string >>> > > >> >> >> >> state string >>> > > >> >> >> >> >>> > > >> >> >> >>JSON is as below. >>> > > >> >> >> >> The JSON is as below. >>> > > >> >> >> >> >>> > > >> >> >> >> { >>> > > >> >> >> >> "name": "Retail_Cube", >>> > > >> >> >> >> "description": "", >>> > > >> >> >> >> "dimensions": [ >>> > > >> >> >> >> { >>> > > >> >> >> >> "name": "RETAIL.FACT_SALES.STOREID", >>> > > >> >> >> >> "table": "RETAIL.FACT_SALES", >>> > > >> >> >> >> "hierarchy": false, >>> > > >> >> >> >> "derived": null, >>> > > >> >> >> >> "column": [ >>> > > >> >> >> >> "STOREID" >>> > > >> >> >> >> ], >>> > > >> >> >> >> "id": 1 >>> > > >> >> >> >> }, >>> > > >> >> >> >> { >>> > > >> >> >> >> "name": "RETAIL.FACT_SALES.ITEMID", >>> > > >> >> >> >> "table": "RETAIL.FACT_SALES", >>> > > >> >> >> >> "hierarchy": false, >>> > > >> >> >> >> "derived": null, >>> > > >> >> >> >> "column": [ >>> > > >> >> >> >> "ITEMID" >>> > > >> >> >> >> ], >>> > > >> >> >> >> "id": 2 >>> > > >> >> >> >> }, >>> > > >> >> >> >> { >>> > > >> >> >> >> "name": "RETAIL.FACT_SALES.CUSTID", >>> > > >> >> >> >> "table": "RETAIL.FACT_SALES", >>> > > >> >> >> >> "hierarchy": false, >>> > > >> >> >> >> "derived": null, >>> > > >> >> >> >> "column": [ >>> > > >> >> >> >> "CUSTID" >>> > > >> >> >> >> ], >>> > > >> >> >> >> "id": 3 >>> > > >> >> >> >> } >>> > > >> >> >> >> ], >>> > > >> >> >> >> "measures": [ >>> > > >> >> >> >> { >>> > > >> >> >> >> "id": 1, >>> > > >> >> >> >> "name": "_COUNT_", >>> > > >> >> >> >> "function": { >>> > > >> >> >> >> "expression": "COUNT", >>> > > >> >> >> >> "returntype": "bigint", >>> > > >> >> >> >> "parameter": { >>> > > >> >> >> >> "type": "constant", >>> > > >> >> >> >> "value": 1 >>> > > >> >> >> >> } >>> > > >> >> >> >> } >>> > > >> >> >> >> } >>> > > >> >> >> >> ], >>> > > >> >> >> >> "rowkey": { >>> > > >> >> >> >> "rowkey_columns": [ >>> > > >> >> >> >> { >>> > > >> >> >> >> "column": "STOREID", >>> > > >> >> >> >> "length": 0, >>> > > >> >> >> >> "dictionary": "true", >>> > > >> >> >> >> "mandatory": false >>> > > >> >> >> >> }, >>> > > >> >> >> >> { >>> > > >> >> >> >> "column": "ITEMID", >>> > > >> >> >> >> "length": 0, >>> > > >> >> >> >> "dictionary": "true", >>> > > >> >> >> >> "mandatory": false >>> > > >> >> >> >> }, >>> > > >> >> >> >> { >>> > > >> >> >> >> "column": "CUSTID", >>> > > >> >> >> >> "length": 0, >>> > > >> >> >> >> "dictionary": "true", >>> > > >> >> >> >> "mandatory": false >>> > > >> >> >> >> } >>> > > >> >> >> >> ], >>> > > >> >> >> >> "aggregation_groups": [ >>> > > >> >> >> >> [ >>> > > >> >> >> >> "STOREID", >>> > > >> >> >> >> "ITEMID", >>> > > >> >> >> >> "CUSTID" >>> > > >> >> >> >> ] >>> > > >> >> >> >> ] >>> > > >> >> >> >> }, >>> > > >> >> >> >> "notify_list": [], >>> > > >> >> >> >> "capacity": "", >>> > > >> >> >> >> "hbase_mapping": { >>> > > >> >> >> >> "column_family": [ >>> > > >> >> >> >> { >>> > > >> >> >> >> "name": "f1", >>> > > >> >> >> >> "columns": [ >>> > > >> >> >> >> { >>> > > >> >> >> >> "qualifier": "m", >>> > > >> >> >> >> "measure_refs": [ >>> > > >> >> >> >> "_COUNT_" >>> > > >> >> >> >> ] >>> > > >> >> >> >> } >>> > > >> >> >> >> ] >>> > > >> >> >> >> } >>> > > >> >> >> >> ] >>> > > >> >> >> >> }, >>> > > >> >> >> >> "project": "RetailProject", >>> > > >> >> >> >> "model_name": "Retail_Cube" >>> > > >> >> >> >> } >>> > > >> >> >> >> >>> > > >> >> >> >>Regards, >>> > > >> >> >> >>Santosh Akhilesh >>> > > >> >> >> >>Bangalore R&D >>> > > >> >> >> >>HUAWEI TECHNOLOGIES CO.,LTD. >>> > > >> >> >> >> >>> > > >> >> >> >>www.huawei.com >>> > > >> >> >> >>>>>---------------------------------------------------------- >>> > > >> >> >> --------------- >>> > > >> >> >> >>- >>> > > >> >> >> >>>>>----------------------------------------------------------- >>> > > >> >> >> >>This e-mail and its attachments contain confidential >>> > information >>> > > >> >>from >>> > > >> >> >> >>HUAWEI, which >>> > > >> >> >> >>is intended only for the person or entity whose address >>>is >>> > > >>listed >>> > > >> >> >>above. >>> > > >> >> >> >>Any use of the >>> > > >> >> >> >>information contained herein in any way (including, but >>>not >>> > > >>limited >>> > > >> >> >>to, >>> > > >> >> >> >>total or partial >>> > > >> >> >> >>disclosure, reproduction, or dissemination) by persons >>>other >>> > > >>than >>> > > >> >>the >>> > > >> >> >> >>intended >>> > > >> >> >> >>recipient(s) is prohibited. If you receive this e-mail >>>in >>> > error, >>> > > >> >> >>please >>> > > >> >> >> >>notify the sender by >>> > > >> >> >> >>phone or email immediately and delete it! >>> > > >> >> >> >> >>> > > >> >> >> >>________________________________________ >>> > > >> >> >> >>From: Shi, Shaofeng [[email protected]] >>> > > >> >> >> >>Sent: Thursday, February 26, 2015 7:01 AM >>> > > >> >> >> >>To: [email protected] >>> > > >> >> >> >>Subject: Re: Error while making cube & Measure option is >>>not >>> > > >> >> >>responding >>> > > >> >> >> >>on GUI >>> > > >> >> >> >> >>> > > >> >> >> >>Hi Santosh, >>> > > >> >> >> >> >>> > > >> >> >> >>It looks like hadoop failed to execute some shell >>>command in >>> > the >>> > > >> >> >> >>container; You need dive into hadoop to see what¹s the >>> > concrete >>> > > >> >>error. >>> > > >> >> >> >>You >>> > > >> >> >> >>can use yarn logs command to fetch all logs: >>> > > >> >> >> >> >>> > > >> >> >> >>yarn logs -applicationId <app_id> >>> > > >> >> >> >> >>> > > >> >> >> >> >>> > > >> >> >> >>On 2/25/15, 7:39 PM, "Santosh Akhilesh" >>> > > >><[email protected] >>> > > >> > >>> > > >> >> >> >>wrote: >>> > > >> >> >> >> >>> > > >> >> >> >>>Hi Luke / Shaofeng , >>> > > >> >> >> >>> Can you please help me to check this issue. >>> > > >> >> >> >>>Regards, >>> > > >> >> >> >>>Santosh Akhilesh >>> > > >> >> >> >>> >>> > > >> >> >> >>>On Tue, Feb 24, 2015 at 10:41 PM, Santosh Akhilesh < >>> > > >> >> >> >>>[email protected]> wrote: >>> > > >> >> >> >>> >>> > > >> >> >> >>>> Hi All , >>> > > >> >> >> >>>> is it because of following error in map >>>reduce >>> job >>> > ? >>> > > >> >>what >>> > > >> >> >> >>>>could >>> > > >> >> >> >>>>be >>> > > >> >> >> >>>> way to resolve this , a google search says that its >>>issue >>> > of >>> > > >> >>Yarn >>> > > >> >> >> >>>>class >>> > > >> >> >> >>>> path , but I am not sure what it is ? >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> Kylin Hive Column Cardinality Job >>>table=RETAIL.FACT_SALES >>> > > >> >> >> >>>> output=/tmp/cardinality/RETAIL.FACT_SALES >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> Application application_1424791969399_0008 failed 2 >>>times >>> > due >>> > > >> >>to AM >>> > > >> >> >> >>>> Container for appattempt_1424791969399_0008_000002 >>>exited >>> > > >>with >>> > > >> >> >> >>>>exitCode: 1 >>> > > >> >> >> >>>> For more detailed output, check application tracking >>> page: >>> > > >> >> >> >>>> >>> > > >>http://santosh:8088/proxy/application_1424791969399_0008/Then, >>> > > >> >> >>click >>> > > >> >> >> >>>>on >>> > > >> >> >> >>>> links to logs of each attempt. >>> > > >> >> >> >>>> Diagnostics: Exception from container-launch. >>> > > >> >> >> >>>> Container id: container_1424791969399_0008_02_000001 >>> > > >> >> >> >>>> Exit code: 1 >>> > > >> >> >> >>>> Stack trace: ExitCodeException exitCode=1: >>> > > >> >> >> >>>> at >>> org.apache.hadoop.util.Shell.runCommand(Shell.java:538) >>> > > >> >> >> >>>> at org.apache.hadoop.util.Shell.run(Shell.java:455) >>> > > >> >> >> >>>> at >>> > > >> >> >> >>>> >>> > > >> >> >> >>>>>>>org.apache.hadoop.util.Shell$ShellCommandExecutor.execut >>> > > >> >> >> e(Shell.java:71 >>> > > >> >> >> >>>>5 >>> > > >> >> >> >>>>) >>> > > >> >> >> >>>> at >>> > > >> >> >> >>>> >>> > > >> >> >> >>>>>>>org.apache.hadoop.yarn.server.nodemanager.DefaultContain >>> > > >> >> >> erExecutor.laun >>> > > >> >> >> >>>>c >>> > > >> >> >> >>>>h >>> > > >> >> >> >>>>Container(DefaultContainerExecutor.java:211) >>> > > >> >> >> >>>> at >>> > > >> >> >> >>>> >>> > > >> >> >> >>>>>>>org.apache.hadoop.yarn.server.nodemanager.containermanag >>> > > >> >> >> er.launcher.Con >>> > > >> >> >> >>>>t >>> > > >> >> >> >>>>a >>> > > >> >> >> >>>>inerLaunch.call(ContainerLaunch.java:302) >>> > > >> >> >> >>>> at >>> > > >> >> >> >>>> >>> > > >> >> >> >>>>>>>org.apache.hadoop.yarn.server.nodemanager.containermanag >>> > > >> >> >> er.launcher.Con >>> > > >> >> >> >>>>t >>> > > >> >> >> >>>>a >>> > > >> >> >> >>>>inerLaunch.call(ContainerLaunch.java:82) >>> > > >> >> >> >>>> at >>> java.util.concurrent.FutureTask.run(FutureTask.java:262) >>> > > >> >> >> >>>> at >>> > > >> >> >> >>>> >>> > > >> >> >> >>>>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread >>> > > >> >> >> PoolExecutor.ja >>> > > >> >> >> >>>>v >>> > > >> >> >> >>>>a >>> > > >> >> >> >>>>:1145) >>> > > >> >> >> >>>> at >>> > > >> >> >> >>>> >>> > > >> >> >> >>>>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa >>> > > >> >> >> dPoolExecutor.j >>> > > >> >> >> >>>>a >>> > > >> >> >> >>>>v >>> > > >> >> >> >>>>a:615) >>> > > >> >> >> >>>> at java.lang.Thread.run(Thread.java:745) >>> > > >> >> >> >>>> Container exited with a non-zero exit code 1 >>> > > >> >> >> >>>> Failing this attempt. Failing the application. >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> ---------- Forwarded message ---------- >>> > > >> >> >> >>>> From: Santoshakhilesh <[email protected]> >>> > > >> >> >> >>>> Date: Tue, Feb 24, 2015 at 7:41 PM >>> > > >> >> >> >>>> Subject: FW: Error while making cube & Measure option >>>is >>> > not >>> > > >> >> >> >>>>responding >>> > > >> >> >> >>>>on >>> > > >> >> >> >>>> GUI >>> > > >> >> >> >>>> To: "[email protected]" >>> > > >> >> >><[email protected]> >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> hi , >>> > > >> >> >> >>>> please someone give me a hand to resolve this >>>issue , >>> > > >>thanks. >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> Regards, >>> > > >> >> >> >>>> Santosh Akhilesh >>> > > >> >> >> >>>> Bangalore R&D >>> > > >> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD. >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> www.huawei.com >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> >>> > > >> >> >> >>>>>>>-------------------------------------------------------- >>> > > >> >> >> --------------- >>> > > >> >> >> >>>>- >>> > > >> >> >> >>>>- >>> > > >> >> >> >>> >>>>------------------------------------------------------------ >>> > > >> >> >> >>>> This e-mail and its attachments contain confidential >>> > > >>information >>> > > >> >> >>from >>> > > >> >> >> >>>> HUAWEI, which >>> > > >> >> >> >>>> is intended only for the person or entity whose >>>address >>> is >>> > > >> >>listed >>> > > >> >> >> >>>>above. >>> > > >> >> >> >>>> Any use of the >>> > > >> >> >> >>>> information contained herein in any way (including, >>>but >>> not >>> > > >> >>limited >>> > > >> >> >> >>>>to, >>> > > >> >> >> >>>> total or partial >>> > > >> >> >> >>>> disclosure, reproduction, or dissemination) by >>>persons >>> > other >>> > > >> >>than >>> > > >> >> >>the >>> > > >> >> >> >>>> intended >>> > > >> >> >> >>>> recipient(s) is prohibited. If you receive this >>>e-mail in >>> > > >>error, >>> > > >> >> >> >>>>please >>> > > >> >> >> >>>> notify the sender by >>> > > >> >> >> >>>> phone or email immediately and delete it! >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> ________________________________________ >>> > > >> >> >> >>>> From: Santoshakhilesh [[email protected]] >>> > > >> >> >> >>>> Sent: Tuesday, February 24, 2015 12:55 PM >>> > > >> >> >> >>>> To: [email protected] >>> > > >> >> >> >>>> Cc: Kulbhushan Rana >>> > > >> >> >> >>>> Subject: FW: Error while making cube & Measure option >>>is >>> > not >>> > > >> >> >> >>>>responding >>> > > >> >> >> >>>>on >>> > > >> >> >> >>>> GUI >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> 2. If I ignore and continue and try to save the cube >>>I >>> get >>> > > >>an >>> > > >> >> >> >>>>exception >>> > > >> >> >> >>>> in Kylin.log , I have checked the path is set >>>correctly >>> and >>> > > >> >> >> >>>>HCatInputFormat >>> > > >> >> >> >>>> this file is present in hive-hcatalog-core-0.14.0.jar >>>. >>> > > >>Please >>> > > >> >>let >>> > > >> >> >>me >>> > > >> >> >> >>>>know >>> > > >> >> >> >>>> what can I do to resolve this ? >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> -- This was path issue , now no more exception in >>> > kylin.log >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> But saveing cube still fails with error. And still >>>can't >>> > add >>> > > >> >> >>measures. >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> Error Message >>> > > >> >> >> >>>> Failed to take action. >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> In log I can find no exception. Following is the last >>>log >>> > in >>> > > >> >> >>kylin.log >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> [pool-3-thread-1]:[2015-02-24 >>> > > >> >> >> >>>> >>> > > >> >> >> >>>>>>>20:47:15,613][INFO][org.apache.kylin.job.impl.threadpool >>> > > >> >> >> .DefaultSchedul >>> > > >> >> >> >>>>e >>> > > >> >> >> >>>>r >>> > > >> >> >> >>>>$FetcherRunner.run(DefaultScheduler.java:117)] >>> > > >> >> >> >>>> - Job Fetcher: 0 running, 0 actual running, 0 ready, >>>6 >>> > others >>> > > >> >> >> >>>> [http-bio-7070-exec-2]:[2015-02-24 >>> > > >> >> >> >>>> >>> > > >> >> >> >>>>>>>20:47:51,610][DEBUG][org.apache.kylin.rest.controller.Cu >>> > > >> >> >> beController.de >>> > > >> >> >> >>>>s >>> > > >> >> >> >>>>e >>> > > >> >> >> >>>>rializeDataModelDesc(CubeController.java:459)] >>> > > >> >> >> >>>> - Saving cube { >>> > > >> >> >> >>>> "name": "", >>> > > >> >> >> >>>> "fact_table": "RETAIL.FACT_SALES", >>> > > >> >> >> >>>> "lookups": [], >>> > > >> >> >> >>>> "filter_condition": "", >>> > > >> >> >> >>>> "capacity": "SMALL", >>> > > >> >> >> >>>> "partition_desc": { >>> > > >> >> >> >>>> "partition_date_column": "", >>> > > >> >> >> >>>> "partition_date_start": 0, >>> > > >> >> >> >>>> "partition_type": "APPEND" >>> > > >> >> >> >>>> }, >>> > > >> >> >> >>>> "last_modified": 0 >>> > > >> >> >> >>>> } >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> local access logs all with 200 , so seems ok. >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:46:56 +0800] "GET >>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:07 +0800] "GET >>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:27 +0800] "GET >>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:28 +0800] "GET >>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:34 +0800] "GET >>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:48 +0800] "GET >>> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >>> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:51 +0800] "POST >>> > > >> >> >>/kylin/api/cubes >>> > > >> >> >> >>>> HTTP/1.1" 200 701 >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> Regards, >>> > > >> >> >> >>>> Santosh Akhilesh >>> > > >> >> >> >>>> Bangalore R&D >>> > > >> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD. >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> www.huawei.com >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> >>> > > >> >> >> >>>>>>>-------------------------------------------------------- >>> > > >> >> >> --------------- >>> > > >> >> >> >>>>- >>> > > >> >> >> >>>>- >>> > > >> >> >> >>> >>>>------------------------------------------------------------ >>> > > >> >> >> >>>> This e-mail and its attachments contain confidential >>> > > >>information >>> > > >> >> >>from >>> > > >> >> >> >>>> HUAWEI, which >>> > > >> >> >> >>>> is intended only for the person or entity whose >>>address >>> is >>> > > >> >>listed >>> > > >> >> >> >>>>above. >>> > > >> >> >> >>>> Any use of the >>> > > >> >> >> >>>> information contained herein in any way (including, >>>but >>> not >>> > > >> >>limited >>> > > >> >> >> >>>>to, >>> > > >> >> >> >>>> total or partial >>> > > >> >> >> >>>> disclosure, reproduction, or dissemination) by >>>persons >>> > other >>> > > >> >>than >>> > > >> >> >>the >>> > > >> >> >> >>>> intended >>> > > >> >> >> >>>> recipient(s) is prohibited. If you receive this >>>e-mail in >>> > > >>error, >>> > > >> >> >> >>>>please >>> > > >> >> >> >>>> notify the sender by >>> > > >> >> >> >>>> phone or email immediately and delete it! >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> ________________________________________ >>> > > >> >> >> >>>> From: Santoshakhilesh [[email protected]] >>> > > >> >> >> >>>> Sent: Tuesday, February 24, 2015 12:09 PM >>> > > >> >> >> >>>> To: [email protected] >>> > > >> >> >> >>>> Cc: Kulbhushan Rana >>> > > >> >> >> >>>> Subject: Error while making cube & Measure option is >>>not >>> > > >> >> >>responding on >>> > > >> >> >> >>>>GUI >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> Hi All , >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> I am building a simple cube for test and using >>>the >>> > binary >>> > > >> >>build >>> > > >> >> >> >>>>0.7.1 >>> > > >> >> >> >>>> . I have following hive tables with columns. >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> fact_sales: >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> storeid int >>> > > >> >> >> >>>> itemid int >>> > > >> >> >> >>>> custid int >>> > > >> >> >> >>>> qty int >>> > > >> >> >> >>>> price double >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> dim_customer >>> > > >> >> >> >>>> custid int >>> > > >> >> >> >>>> name string >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> dim_item >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> itemid int >>> > > >> >> >> >>>> category string >>> > > >> >> >> >>>> brand string >>> > > >> >> >> >>>> color string >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> dim_store >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> storeid int >>> > > >> >> >> >>>> city string >>> > > >> >> >> >>>> state string >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> Please help me to answer following issues; >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> 1. When I go to measure section and click on measure >>> > option , >>> > > >> >> >>there is >>> > > >> >> >> >>>>no >>> > > >> >> >> >>>> response , I want add measure on qty and price with >>>sum >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> 2. If I ignore and continue and try to save the cube >>>I >>> get >>> > > >>an >>> > > >> >> >> >>>>exception >>> > > >> >> >> >>>> in Kylin.log , I have checked the path is set >>>correctly >>> and >>> > > >> >> >> >>>>HCatInputFormat >>> > > >> >> >> >>>> this file is present in hive-hcatalog-core-0.14.0.jar >>>. >>> > > >>Please >>> > > >> >>let >>> > > >> >> >>me >>> > > >> >> >> >>>>know >>> > > >> >> >> >>>> what can I do to resolve this ? >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> 3. Also I have another question since this is a test >>>and >>> > > >>data is >>> > > >> >> >>small >>> > > >> >> >> >>>>I >>> > > >> >> >> >>>> have not partitioned the fact table , is it ok to >>>skip >>> > > >>partition >>> > > >> >> >>stage >>> > > >> >> >> >>>> while cube build ? >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> Exception >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> pool-4-thread-4]:[2015-02-24 >>> > > >> >> >> >>>> >>> > > >> >> >> >>>>>>>19:26:32,577][ERROR][org.apache.kylin.job.impl.threadpoo >>> > > >> >> >> l.DefaultSchedu >>> > > >> >> >> >>>>l >>> > > >> >> >> >>>>e >>> > > >> >> >> >>>>r$JobRunner.run(DefaultScheduler.java:134)] >>> > > >> >> >> >>>> - ExecuteException >>> job:c3532a6f-97ea-474a-b36a-218dd517cedb >>> > > >> >> >> >>>> org.apache.kylin.job.exception.ExecuteException: >>> > > >> >> >> >>>> org.apache.kylin.job.exception.ExecuteException: >>> > > >> >> >> >>>> java.lang.NoClassDefFoundError: >>> > > >> >> >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat >>> > > >> >> >> >>>> at >>> > > >> >> >> >>>> >>> > > >> >> >> >>>>>>>org.apache.kylin.job.execution.AbstractExecutable.execut >>> > > >> >> >> e(AbstractExecu >>> > > >> >> >> >>>>t >>> > > >> >> >> >>>>a >>> > > >> >> >> >>>>ble.java:102) >>> > > >> >> >> >>>> at >>> > > >> >> >> >>>> >>> > > >> >> >> >>>>>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$Jo >>> > > >> >> >> bRunner.run(Def >>> > > >> >> >> >>>>a >>> > > >> >> >> >>>>u >>> > > >> >> >> >>>>ltScheduler.java:132) >>> > > >> >> >> >>>> at >>> > > >> >> >> >>>> >>> > > >> >> >> >>>>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread >>> > > >> >> >> PoolExecutor.ja >>> > > >> >> >> >>>>v >>> > > >> >> >> >>>>a >>> > > >> >> >> >>>>:1145) >>> > > >> >> >> >>>> at >>> > > >> >> >> >>>> >>> > > >> >> >> >>>>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa >>> > > >> >> >> dPoolExecutor.j >>> > > >> >> >> >>>>a >>> > > >> >> >> >>>>v >>> > > >> >> >> >>>>a:615) >>> > > >> >> >> >>>> at java.lang.Thread.run(Thread.java:745) >>> > > >> >> >> >>>> Caused by: >>> org.apache.kylin.job.exception.ExecuteException: >>> > > >> >> >> >>>> java.lang.NoClassDefFoundError: >>> > > >> >> >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> The JSON is as below. >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> { >>> > > >> >> >> >>>> "name": "Retail_Cube", >>> > > >> >> >> >>>> "description": "", >>> > > >> >> >> >>>> "dimensions": [ >>> > > >> >> >> >>>> { >>> > > >> >> >> >>>> "name": "RETAIL.FACT_SALES.STOREID", >>> > > >> >> >> >>>> "table": "RETAIL.FACT_SALES", >>> > > >> >> >> >>>> "hierarchy": false, >>> > > >> >> >> >>>> "derived": null, >>> > > >> >> >> >>>> "column": [ >>> > > >> >> >> >>>> "STOREID" >>> > > >> >> >> >>>> ], >>> > > >> >> >> >>>> "id": 1 >>> > > >> >> >> >>>> }, >>> > > >> >> >> >>>> { >>> > > >> >> >> >>>> "name": "RETAIL.FACT_SALES.ITEMID", >>> > > >> >> >> >>>> "table": "RETAIL.FACT_SALES", >>> > > >> >> >> >>>> "hierarchy": false, >>> > > >> >> >> >>>> "derived": null, >>> > > >> >> >> >>>> "column": [ >>> > > >> >> >> >>>> "ITEMID" >>> > > >> >> >> >>>> ], >>> > > >> >> >> >>>> "id": 2 >>> > > >> >> >> >>>> }, >>> > > >> >> >> >>>> { >>> > > >> >> >> >>>> "name": "RETAIL.FACT_SALES.CUSTID", >>> > > >> >> >> >>>> "table": "RETAIL.FACT_SALES", >>> > > >> >> >> >>>> "hierarchy": false, >>> > > >> >> >> >>>> "derived": null, >>> > > >> >> >> >>>> "column": [ >>> > > >> >> >> >>>> "CUSTID" >>> > > >> >> >> >>>> ], >>> > > >> >> >> >>>> "id": 3 >>> > > >> >> >> >>>> } >>> > > >> >> >> >>>> ], >>> > > >> >> >> >>>> "measures": [ >>> > > >> >> >> >>>> { >>> > > >> >> >> >>>> "id": 1, >>> > > >> >> >> >>>> "name": "_COUNT_", >>> > > >> >> >> >>>> "function": { >>> > > >> >> >> >>>> "expression": "COUNT", >>> > > >> >> >> >>>> "returntype": "bigint", >>> > > >> >> >> >>>> "parameter": { >>> > > >> >> >> >>>> "type": "constant", >>> > > >> >> >> >>>> "value": 1 >>> > > >> >> >> >>>> } >>> > > >> >> >> >>>> } >>> > > >> >> >> >>>> } >>> > > >> >> >> >>>> ], >>> > > >> >> >> >>>> "rowkey": { >>> > > >> >> >> >>>> "rowkey_columns": [ >>> > > >> >> >> >>>> { >>> > > >> >> >> >>>> "column": "STOREID", >>> > > >> >> >> >>>> "length": 0, >>> > > >> >> >> >>>> "dictionary": "true", >>> > > >> >> >> >>>> "mandatory": false >>> > > >> >> >> >>>> }, >>> > > >> >> >> >>>> { >>> > > >> >> >> >>>> "column": "ITEMID", >>> > > >> >> >> >>>> "length": 0, >>> > > >> >> >> >>>> "dictionary": "true", >>> > > >> >> >> >>>> "mandatory": false >>> > > >> >> >> >>>> }, >>> > > >> >> >> >>>> { >>> > > >> >> >> >>>> "column": "CUSTID", >>> > > >> >> >> >>>> "length": 0, >>> > > >> >> >> >>>> "dictionary": "true", >>> > > >> >> >> >>>> "mandatory": false >>> > > >> >> >> >>>> } >>> > > >> >> >> >>>> ], >>> > > >> >> >> >>>> "aggregation_groups": [ >>> > > >> >> >> >>>> [ >>> > > >> >> >> >>>> "STOREID", >>> > > >> >> >> >>>> "ITEMID", >>> > > >> >> >> >>>> "CUSTID" >>> > > >> >> >> >>>> ] >>> > > >> >> >> >>>> ] >>> > > >> >> >> >>>> }, >>> > > >> >> >> >>>> "notify_list": [], >>> > > >> >> >> >>>> "capacity": "", >>> > > >> >> >> >>>> "hbase_mapping": { >>> > > >> >> >> >>>> "column_family": [ >>> > > >> >> >> >>>> { >>> > > >> >> >> >>>> "name": "f1", >>> > > >> >> >> >>>> "columns": [ >>> > > >> >> >> >>>> { >>> > > >> >> >> >>>> "qualifier": "m", >>> > > >> >> >> >>>> "measure_refs": [ >>> > > >> >> >> >>>> "_COUNT_" >>> > > >> >> >> >>>> ] >>> > > >> >> >> >>>> } >>> > > >> >> >> >>>> ] >>> > > >> >> >> >>>> } >>> > > >> >> >> >>>> ] >>> > > >> >> >> >>>> }, >>> > > >> >> >> >>>> "project": "RetailProject", >>> > > >> >> >> >>>> "model_name": "Retail_Cube" >>> > > >> >> >> >>>> } >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> Regards, >>> > > >> >> >> >>>> Santosh Akhilesh >>> > > >> >> >> >>>> Bangalore R&D >>> > > >> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD. >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> www.huawei.com >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> >>> > > >> >> >> >>>>>>>-------------------------------------------------------- >>> > > >> >> >> --------------- >>> > > >> >> >> >>>>- >>> > > >> >> >> >>>>- >>> > > >> >> >> >>> >>>>------------------------------------------------------------ >>> > > >> >> >> >>>> This e-mail and its attachments contain confidential >>> > > >>information >>> > > >> >> >>from >>> > > >> >> >> >>>> HUAWEI, which >>> > > >> >> >> >>>> is intended only for the person or entity whose >>>address >>> is >>> > > >> >>listed >>> > > >> >> >> >>>>above. >>> > > >> >> >> >>>> Any use of the >>> > > >> >> >> >>>> information contained herein in any way (including, >>>but >>> not >>> > > >> >>limited >>> > > >> >> >> >>>>to, >>> > > >> >> >> >>>> total or partial >>> > > >> >> >> >>>> disclosure, reproduction, or dissemination) by >>>persons >>> > other >>> > > >> >>than >>> > > >> >> >>the >>> > > >> >> >> >>>> intended >>> > > >> >> >> >>>> recipient(s) is prohibited. If you receive this >>>e-mail in >>> > > >>error, >>> > > >> >> >> >>>>please >>> > > >> >> >> >>>> notify the sender by >>> > > >> >> >> >>>> phone or email immediately and delete it! >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> >>> > > >> >> >> >>>> -- >>> > > >> >> >> >>>> Regards, >>> > > >> >> >> >>>> Santosh Akhilesh >>> > > >> >> >> >>>> +91-0-9845482201 >>> > > >> >> >> >>>> >>> > > >> >> >> >>> >>> > > >> >> >> >>> >>> > > >> >> >> >>> >>> > > >> >> >> >>>-- >>> > > >> >> >> >>>Regards, >>> > > >> >> >> >>>Santosh Akhilesh >>> > > >> >> >> >>>+91-0-9845482201 >>> > > >> >> >> >>> > > >> >> >> >>> > > >> >> >>> > > >> >> >>> > > >> >>> > > >> >>> > > >>> > > >>> > >>> > >>> > -- >>> > Regards, >>> > Santosh Akhilesh >>> > +91-0-9845482201 >>> > >>> >> >> >> >>-- >>Regards, >>Santosh Akhilesh >>+91-0-9845482201
