Did you append the database name before table name when importing it to Kylin?
On 2/27/15, 10:51 AM, "Santosh Akhilesh" <[email protected]> wrote: >Ok , but it's strange that this issue is not reported while importing hive >table. >If this is limitation of this version , I guess it's better to detect and >prompt user during hive table import. This is just my suggestion. >On Fri, 27 Feb 2015 at 8:02 am, Shi, Shaofeng <[email protected]> wrote: > >> In 0.6.x it only support tables in default database, this is a >>limitation; >> The support for non-default tables will be released in 0.7; >> >> To bypass this issue for now, please copy the table to default database >>as >> a workaround; >> >> On 2/27/15, 10:16 AM, "Santosh Akhilesh" <[email protected]> >> wrote: >> >> >@Jason >> >thanks , but now as suggested by Saofeng I m not using the inverted >>index >> >brach as its not stable. >> >I have switched back to 0.6 branch , in this branch yesterday night I >> >could >> >crete the cube successfully but there is issue while building it. I >>feel >> >that at step 1 of cube build while creating flat table when command is >> >issued to hive if the tables are not under default datbase flat table >> >creation is failed and cube build fails. my fact and dimension tables >>are >> >under a database called retail. >> > >> >@Saofeng - Can you please confirm this behavior ? Do I need to create >>the >> >hive tables under default database? >> > >> >On Fri, Feb 27, 2015 at 7:32 AM, jason zhong <[email protected]> >> >wrote: >> > >> >> @Santoshakhilesh >> >> >> >> 1. When I go to measure section and click on measure option , there >>is >> >>no >> >> response , I want add measure on qty and price with sum >> >> --bug fixed on inverted-index branch >> >> >> >> >> >> On Fri, Feb 27, 2015 at 3:03 AM, Santosh Akhilesh < >> >> [email protected] >> >> > wrote: >> >> >> >> > Hi Shaofeng , >> >> > I have build the 0.6 version and now able to create the cube >> >> > successfully. >> >> > While building the cube , it fails at step1 with following >>error. >> >> > Table not found 'DIM_ITEM' >> >> > the table exists , but its under retail data base and not >>under >> >> > default database. >> >> > does kylin require hive taables to be under default database >>? I >> >>see >> >> > the flat table being created under default database. >> >> > >> >> > Logging initialized using configuration in >> >> > >> >> > >> >> >> >>jar:file:/home/santosh/work/frameworks/apache-hive-1.0.0/ >> lib/hive-common- >> >>1.0.0.jar!/hive-log4j.properties >> >> > SLF4J: Class path contains multiple SLF4J bindings. >> >> > SLF4J: Found binding in >> >> > >> >> > >> >> >> >>[jar:file:/home/santosh/work/frameworks/hadoop-2.6.0/ >> share/hadoop/common/ >> >>lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class] >> >> > SLF4J: Found binding in >> >> > >> >> > >> >> >> >>[jar:file:/home/santosh/work/frameworks/apache-hive-1. >> 0.0/lib/hive-jdbc-1 >> >>.0.0-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class] >> >> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an >> >> > explanation. >> >> > SLF4J: Actual binding is of type >>[org.slf4j.impl.Log4jLoggerFactory] >> >> > OK >> >> > Time taken: 0.964 seconds >> >> > OK >> >> > Time taken: 0.948 seconds >> >> > FAILED: SemanticException [Error 10001]: Line 12:11 Table not found >> >> > 'DIM_ITEM' >> >> > >> >> > >> >> > >> >> > Command is as below. >> >> > >> >> > hive -e "DROP TABLE IF EXISTS >> >> > >> >>>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd44 >>>>; >> >> > CREATE EXTERNAL TABLE IF NOT EXISTS >> >> > >> >>>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd44 >> >> > ( >> >> > STOREID int >> >> > ,ITEMID int >> >> > ,CUSTID int >> >> > ,QTY int >> >> > ,AMOUNT double >> >> > ) >> >> > ROW FORMAT DELIMITED FIELDS TERMINATED BY '\177' >> >> > STORED AS SEQUENCEFILE >> >> > LOCATION >> >> > >> >> >> >>'/tmp/kylin-8b30b29b-5f2c-4b63-8c0f-07d1f559dd44/kylin_ >> intermediate_test_ >> >>FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd44'; >> >> > SET hive.exec.compress.output=true; >> >> > SET hive.auto.convert.join.noconditionaltask = true; >> >> > SET hive.auto.convert.join.noconditionaltask.size = 300000000; >> >> > INSERT OVERWRITE TABLE >> >> > >> >>>>kylin_intermediate_test_FULL_BUILD_8b30b29b_5f2c_4b63_8c0f_07d1f559dd44 >> >> > SELECT >> >> > FACT_SALES.STOREID >> >> > ,FACT_SALES.ITEMID >> >> > ,FACT_SALES.CUSTID >> >> > ,FACT_SALES.QTY >> >> > ,FACT_SALES.AMOUNT >> >> > FROM FACT_SALES >> >> > INNER JOIN DIM_STORE >> >> > ON FACT_SALES.STOREID = DIM_STORE.SROREID >> >> > INNER JOIN DIM_ITEM >> >> > ON FACT_SALES.ITEMID = DIM_ITEM.ITEMID >> >> > INNER JOIN DIM_CUSTOMER >> >> > ON FACT_SALES.CUSTID = DIM_CUSTOMER.CUSTID >> >> > ; >> >> > " >> >> > >> >> > >> >> > >> >> > On Thu, Feb 26, 2015 at 8:11 PM, Shi, Shaofeng <[email protected]> >> >>wrote: >> >> > >> >> > > The 0.7.1 is test version, its package contains the “snapshot” >> >>suffix; >> >> we >> >> > > will upload a new package there; Luke will also add a message >>there >> >>to >> >> > > avoid this confusion; >> >> > > >> >> > > Regarding the problem that you encountered, could you please >>open a >> >> JIRA >> >> > > ticket for tracking? Here is link of Apache JIRA: >> >> > > >> >> > > https://issues.apache.org/jira/secure/Dashboard.jspa >> >> > > >> >> > > >> >> > > Thanks for the feedback! >> >> > > >> >> > > On 2/26/15, 10:21 PM, "Santosh Akhilesh" >><[email protected] >> > >> >> > > wrote: >> >> > > >> >> > > >Actually I see this being published on kylin webpage. >> >> > > >http://kylin.incubator.apache.org/download/ >> >> > > >I am using 0.7.1 inverted index branch binary distribution. >> >> > > >If this is not stable please give me the link of stable branch I >> >>would >> >> > try >> >> > > >building and testing tonight. >> >> > > >On Thu, 26 Feb 2015 at 7:30 pm, Shi, Shaofeng <[email protected]> >> >> wrote: >> >> > > > >> >> > > >> Hi Santosh, it is not recommended to use the dev code branch >> >> > (actually I >> >> > > >> don’t know how you get the v0.7.x build and what’s the detail >> >> version >> >> > of >> >> > > >> that; each day we submit many changes to that); >> >> > > >> >> >> > > >> The options are 1) switch back to latest release v0.6.5; or 2) >> >>wait >> >> > for >> >> > > >> the formal release of 0.7, that should be in March; >>Otherwise, we >> >> > > >>couldn’t >> >> > > >> ensure there is no new problems come out in your next steps; >> >> > > >> >> >> > > >> On 2/26/15, 5:39 PM, "Santosh Akhilesh" >> >><[email protected]> >> >> > > >>wrote: >> >> > > >> >> >> > > >> >Hi Shaofeng >> >> > > >> >So what do you suggest , how should I proceed further with >>this >> >> > > >>release? >> >> > > >> >Will there be a patch? Any alternate way I can create cube? >> >> > > >> >Please suggest. >> >> > > >> >Regards >> >> > > >> >Santosh >> >> > > >> >On Thu, 26 Feb 2015 at 3:04 pm, Shi, Shaofeng >><[email protected] >> > >> >> > > wrote: >> >> > > >> > >> >> > > >> >> Hi Santosh, >> >> > > >> >> >> >> > > >> >> 0.7.1 hasn’t been formally released; from 0.6.x to 0.7.x we >> >>have >> >> > > >> >>metadata >> >> > > >> >> structure change; While, the web UI (cube wizard) for this >> >>change >> >> > > >>hasn’t >> >> > > >> >> been stabilized; So it is not strange that you got trouble >> >>when >> >> > > >>saving >> >> > > >> >>the >> >> > > >> >> cube; >> >> > > >> >> >> >> > > >> >> @Jason, any idea about the JS error? >> >> > > >> >> >> >> > > >> >> On 2/26/15, 5:08 PM, "Santosh Akhilesh" < >> >> [email protected] >> >> > > >> >> > > >> >>wrote: >> >> > > >> >> >> >> > > >> >> >Hi Shaofeng, >> >> > > >> >> > >> >> > > >> >> >I am using the binary distribution 0.7.1. I have not been >> >>able >> >> to >> >> > > >>save >> >> > > >> >> >cube >> >> > > >> >> >even once. I have tried creating new project and from >>local >> >> > machine >> >> > > >>and >> >> > > >> >> >server machine. But I am always stuck with this error. I >>am >> >> never >> >> > > >> >>allowed >> >> > > >> >> >to add measures and never been able to save the cube. I >>also >> >>see >> >> > the >> >> > > >> >> >kylin.log and it always tries to save cube with append >>mode. >> >>One >> >> > > >>thing >> >> > > >> >>I >> >> > > >> >> >need to tell that at partition stage since I don't have a >>big >> >> fact >> >> > > >> >>table >> >> > > >> >> >now I have not partititioned the fact table and I skip >>this >> >> step. >> >> > > >>Does >> >> > > >> >> >this >> >> > > >> >> >have affect in saving the cube. Is this because some >> >>metadata is >> >> > > >> >>available >> >> > > >> >> >and it tries to modify the cube? I am using latest Hadoop >> >>2.6.6. >> >> > Yes >> >> > > >> >>kylin >> >> > > >> >> >propert I have not added the jar. I will add them and >>check. >> >>But >> >> > > >>cube >> >> > > >> >> >creation failure is really puzzling me. I could see no >>error >> >> logs >> >> > in >> >> > > >> >> >kylin.log. >> >> > > >> >> >Regards >> >> > > >> >> >Santosh >> >> > > >> >> >On Thu, 26 Feb 2015 at 1:40 pm, Shi, Shaofeng >> >><[email protected] >> >> > >> >> > > >> wrote: >> >> > > >> >> > >> >> > > >> >> >> Which version or code branch are you using? I assume >>you’re >> >> > using >> >> > > >>the >> >> > > >> >> >> stable version from master; Seems you’re trying to edit >>an >> >> > > >>existing >> >> > > >> >>cube >> >> > > >> >> >> to add new measurement, try refresh your browser's >>cache; >> >>if >> >> it >> >> > > >>still >> >> > > >> >> >> couldn’t be saved, try to create a new cube; >> >> > > >> >> >> >> >> > > >> >> >> The two error traces in tomcat need be taken care: >> >> > > >> >> >> >> >> > > >> >> >> 1) java.lang.NoClassDefFoundError: >> >> > > >> >> >>org/apache/kylin/common/mr/KylinMapper >> >> > > >> >> >> Please check kylin.properties file, making sure >>the >> >> > > >> >> >>“kylin.job.jar” >> >> > > >> >> >> points to a right jar file; It will be loaded in >> >>Map-reduce; >> >> > > >> >> >> >> >> > > >> >> >> 2) java.lang.IllegalArgumentException: No enum constant >> >> > > >> >> >> org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS >> >> > > >> >> >> This indicates your hadoop version might be old; >> >> Please >> >> > > >>check >> >> > > >> >> >>and >> >> > > >> >> >> ensure >> >> > > >> >> >> hadoop version is 2.2 or above. >> >> > > >> >> >> >> >> > > >> >> >> On 2/26/15, 3:21 PM, "Santoshakhilesh" >> >> > > >><[email protected]> >> >> > > >> >> >> wrote: >> >> > > >> >> >> >> >> > > >> >> >> >Hi Shaofeng , >> >> > > >> >> >> > >> >> > > >> >> >> > I am using chrome , When I click on button to add >> >> measures , >> >> > > >> >> >>following >> >> > > >> >> >> >is error on chrome console. When I try to save the cube >> >>there >> >> > is >> >> > > >>no >> >> > > >> >> >>error >> >> > > >> >> >> >in console. I just get a error dialog saying failed to >> >>take >> >> > > >>action >> >> > > >> >>and >> >> > > >> >> >> >gives me the JSON cube schema. >> >> > > >> >> >> > >> >> > > >> >> >> >Error on chrome debug console is as below; >> >> > > >> >> >> > >> >> > > >> >> >> > ReferenceError: CubeDescModel is not defined >> >> > > >> >> >> > at h.$scope.addNewMeasure (scripts.min.0.js:15984) >> >> > > >> >> >> > at scripts.min.0.js:180 >> >> > > >> >> >> > at scripts.min.0.js:197 >> >> > > >> >> >> > at h.$eval (scripts.min.0.js:119) >> >> > > >> >> >> > at h.$apply (scripts.min.0.js:119) >> >> > > >> >> >> > at HTMLButtonElement.<anonymous> >> >>(scripts.min.0.js:197) >> >> > > >> >> >> > at HTMLButtonElement.m.event.dispatch >> >> (scripts.min.0.js:3) >> >> > > >> >> >> > at HTMLButtonElement.r.handle >> >> > > >> >> >> >(scripts.min.0.js:3)scripts.min.0.js:100 (anonymous >> >> > > >> >> >> >function)scripts.min.0.js:77 (anonymous >> >> > > >> >>function)scripts.min.0.js:119 >> >> > > >> >> >> >h.$applyscripts.min.0.js:197 (anonymous >> >> > > >>function)scripts.min.0.js:3 >> >> > > >> >> >> >m.event.dispatchscripts.min.0.js:3 r.handle >> >> > > >> >> >> > >> >> > > >> >> >> > About the hive table import , I got pass the run >>shell >> >> > command >> >> > > >> >> >> >exception but it still fails the hadoop log is; >> >> > > >> >> >> >2015-02-26 20:46:48,332 INFO [main] >> >>org.apache.hadoop.mapred. >> >> > > >> >> YarnChild: >> >> > > >> >> >> >mapreduce.cluster.local.dir for child: >> >> > > >> >> >> >> >>>/tmp/hadoop-root/nm-local-dir/usercache/root/appcache/appli >> >> > > >> >> >> cation_14249530 >> >> > > >> >> >> >91340_0002 >> >> > > >> >> >> >2015-02-26 20:46:48,776 INFO [main] >> >> > > >> >> >> >org.apache.hadoop.conf.Configuration.deprecation: >> >>session.id >> >> > is >> >> > > >> >> >> >deprecated. Instead, use dfs.metrics.session-id >> >> > > >> >> >> >2015-02-26 20:46:49,310 INFO [main] >> >> > > >>org.apache.hadoop.mapred.Task: >> >> > > >> >> >>Using >> >> > > >> >> >> >ResourceCalculatorProcessTree : [ ] >> >> > > >> >> >> >2015-02-26 20:46:49,386 FATAL [main] >> >> > > >> >> >>org.apache.hadoop.mapred.YarnChild: >> >> > > >> >> >> >Error running child : java.lang.NoClassDefFoundError: >> >> > > >> >> >> >org/apache/kylin/common/mr/KylinMapper >> >> > > >> >> >> > at java.lang.ClassLoader.defineClass1(Native Method) >> >> > > >> >> >> > at >> >>java.lang.ClassLoader.defineClass(ClassLoader.java:800) >> >> > > >> >> >> > at >> >> > > >> >> >> >java.security.SecureClassLoader.defineClass( >> >> > > >> >> SecureClassLoader.java:142) >> >> > > >> >> >> > at >> >> > java.net.URLClassLoader.defineClass(URLClassLoader.java:449) >> >> > > >> >> >> > at >> >> java.net.URLClassLoader.access$100(URLClassLoader.java:71) >> >> > > >> >> >> > at java.net.URLClassLoader$1.run( >> URLClassLoader.java:361) >> >> > > >> >> >> > at java.net.URLClassLoader$1.run( >> URLClassLoader.java:355) >> >> > > >> >> >> > at java.security.AccessController.doPrivileged(Native >> >> Method) >> >> > > >> >> >> > at >> >> java.net.URLClassLoader.findClass(URLClassLoader.java:354) >> >> > > >> >> >> > at >>java.lang.ClassLoader.loadClass(ClassLoader.java:425) >> >> > > >> >> >> > at >> >> > sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) >> >> > > >> >> >> > >> >> > > >> >> >> >tomcat logs: >> >> > > >> >> >> >usage: HiveColumnCardinalityJob >> >> > > >> >> >> > -output <path> Output path >> >> > > >> >> >> > -table <table name> The hive table name >> >> > > >> >> >> >[pool-4-thread-2]:[2015-02-26 >> >> > > >> >> >> >> >>>20:47:49,936][ERROR][org.apache.kylin.job.common.HadoopShel >> >> > > >> >> >> lExecutable.doW >> >> > > >> >> >> >ork(HadoopShellExecutable.java:64)] - error execute >> >> > > >> >> >> >> >> > >HadoopShellExecutable{id=d4730d26-7fe6-412e-9841-3288ab362c5b-00, >> >> > > >> >> >> >name=null, state=RUNNING} >> >> > > >> >> >> >java.lang.IllegalArgumentException: No enum constant >> >> > > >> >> >> >org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS >> >> > > >> >> >> > at java.lang.Enum.valueOf(Enum.java:236) >> >> > > >> >> >> > at >> >> > > >> >> >> >> >> > > >> >>>>>org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.valueOf( >> >> > > >> >> >> Framewo >> >> > > >> >> >> >rkCounterGroup.java:148) >> >> > > >> >> >> > at >> >> > > >> >> >> >> >>>org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup. >> >> > > >> >> >> findCounter(Fra >> >> > > >> >> >> >meworkCounterGroup.java:182) >> >> > > >> >> >> > at >> >> > > >> >> >> >> >>>org.apache.hadoop.mapreduce.counters.AbstractCounters.findC >> >> > > >> >> >> ounter(Abstract >> >> > > >> >> >> >Counters.java:154) >> >> > > >> >> >> > at >> >> > > >> >> >> >> >>>org.apache.hadoop.mapreduce.TypeConverter.fromYarn(TypeConv >> >> > > >> >> >> erter.java:240) >> >> > > >> >> >> > at >> >> > > >> >> >> >> >>>org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounte >> >> > > >> >> >> rs(ClientServic >> >> > > >> >> >> >eDelegate.java:370) >> >> > > >> >> >> > at >> >> > > >> >> >> >org.apache.hadoop.mapred.YARNRunner.getJobCounters( >> >> > > >> >> YARNRunner.java:511) >> >> > > >> >> >> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756) >> >> > > >> >> >> > at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753) >> >> > > >> >> >> > at java.security.AccessController.doPrivileged(Native >> >> Method) >> >> > > >> >> >> > at javax.security.auth.Subject.doAs(Subject.java:415) >> >> > > >> >> >> > at >> >> > > >> >> >> >> >>>org.apache.hadoop.security.UserGroupInformation.doAs(UserGr >> >> > > >> >> >> oupInformation. >> >> > > >> >> >> >java:1491) >> >> > > >> >> >> > at >> >>org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753) >> >> > > >> >> >> > at >> >> > > >> >> >>>>org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1361) >> >> > > >> >> >> > at org.apache.hadoop.mapreduce. >> Job.waitForCompletion(Job. >> >> > > >> java:1289) >> >> > > >> >> >> > at >> >> > > >> >> >> >> >>>org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForComple >> >> > > >> >> >> tion(AbstractHa >> >> > > >> >> >> >doopJob.java:134) >> >> > > >> >> >> > at >> >> > > >> >> >> >> >> > > >> >> >> > >> >>>>>>>org.apache.kylin.job.hadoop.cardinality.HiveColumnCardinalityJob.run >>>>>>>( >> >> > > >> >> >> HiveC >> >> > > >> >> >> >olumnCardinalityJob.java:114) >> >> > > >> >> >> > at >> >>org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) >> >> > > >> >> >> > at >> >>org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84) >> >> > > >> >> >> > at >> >> > > >> >> >> >> >>>org.apache.kylin.job.common.HadoopShellExecutable.doWork(Ha >> >> > > >> >> >> doopShellExecut >> >> > > >> >> >> >able.java:62) >> >> > > >> >> >> > at >> >> > > >> >> >> >> >>>org.apache.kylin.job.execution.AbstractExecutable.execute(A >> >> > > >> >> >> bstractExecutab >> >> > > >> >> >> >le.java:99) >> >> > > >> >> >> > at >> >> > > >> >> >> >> >>>org.apache.kylin.job.execution.DefaultChainedExecutable.doW >> >> > > >> >> >> ork(DefaultChai >> >> > > >> >> >> >nedExecutable.java:50) >> >> > > >> >> >> > at >> >> > > >> >> >> >> >>>org.apache.kylin.job.execution.AbstractExecutable.execute(A >> >> > > >> >> >> bstractExecutab >> >> > > >> >> >> >le.java:99) >> >> > > >> >> >> > at >> >> > > >> >> >> >> >>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRu >> >> > > >> >> >> nner.run(Defaul >> >> > > >> >> >> >tScheduler.java:132) >> >> > > >> >> >> > at >> >> > > >> >> >> >> >>>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoo >> >> > > >> >> >> lExecutor.java: >> >> > > >> >> >> >1145) >> >> > > >> >> >> > at >> >> > > >> >> >> >> >>>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPo >> >> > > >> >> >> olExecutor.java >> >> > > >> >> >> >:615) >> >> > > >> >> >> > at java.lang.Thread.run(Thread.java:745) >> >> > > >> >> >> > >> >> > > >> >> >> >Regards, >> >> > > >> >> >> >Santosh Akhilesh >> >> > > >> >> >> >Bangalore R&D >> >> > > >> >> >> >HUAWEI TECHNOLOGIES CO.,LTD. >> >> > > >> >> >> > >> >> > > >> >> >> >www.huawei.com >> >> > > >> >> >> >> >>>----------------------------------------------------------- >> >> > > >> >> >> --------------- >> >> > > >> >> >> >> >>>----------------------------------------------------------- >> >> > > >> >> >> >This e-mail and its attachments contain confidential >> >> > information >> >> > > >> >>from >> >> > > >> >> >> >HUAWEI, which >> >> > > >> >> >> >is intended only for the person or entity whose >>address is >> >> > listed >> >> > > >> >> >>above. >> >> > > >> >> >> >Any use of the >> >> > > >> >> >> >information contained herein in any way (including, but >> >>not >> >> > > >>limited >> >> > > >> >>to, >> >> > > >> >> >> >total or partial >> >> > > >> >> >> >disclosure, reproduction, or dissemination) by persons >> >>other >> >> > than >> >> > > >> >>the >> >> > > >> >> >> >intended >> >> > > >> >> >> >recipient(s) is prohibited. If you receive this e-mail >>in >> >> > error, >> >> > > >> >>please >> >> > > >> >> >> >notify the sender by >> >> > > >> >> >> >phone or email immediately and delete it! >> >> > > >> >> >> > >> >> > > >> >> >> >________________________________________ >> >> > > >> >> >> >From: Shi, Shaofeng [[email protected]] >> >> > > >> >> >> >Sent: Thursday, February 26, 2015 11:32 AM >> >> > > >> >> >> >To: [email protected] >> >> > > >> >> >> >Cc: Kulbhushan Rana >> >> > > >> >> >> >Subject: Re: Error while making cube & Measure option >>is >> >>not >> >> > > >> >>responding >> >> > > >> >> >> >on GUI >> >> > > >> >> >> > >> >> > > >> >> >> >Hi Santosh, hive table importing issue should not >>impact >> >>on >> >> > cube >> >> > > >> >> >>saving. >> >> > > >> >> >> > >> >> > > >> >> >> >If you couldn’t save the cube, firstly please check >> >>whether >> >> > > >>there is >> >> > > >> >> >>error >> >> > > >> >> >> >in the tomcat’s log; If not please check your web >> >>browser; We >> >> > > >> >>suggest >> >> > > >> >> >>use >> >> > > >> >> >> >Firefox (with firebug add-on) or Chrome, open the JS >> >>console >> >> > > >>(press >> >> > > >> >> >>F12) >> >> > > >> >> >> >and then operate web UI, check whether there is any >>error >> >> > > >>reported >> >> > > >> >>in >> >> > > >> >> >> >browser. >> >> > > >> >> >> > >> >> > > >> >> >> > >> >> > > >> >> >> >On 2/26/15, 1:08 PM, "Santoshakhilesh" >> >> > > >><[email protected] >> >> > > >> > >> >> > > >> >> >> >wrote: >> >> > > >> >> >> > >> >> > > >> >> >> >>Hi Shaofeng , >> >> > > >> >> >> >> Thanks for replying. >> >> > > >> >> >> >> Yes I am checking the yarn exception, But I find >>that >> >> this >> >> > > >>error >> >> > > >> >> >>comes >> >> > > >> >> >> >>while importing the hive table to kylin. >> >> > > >> >> >> >> Even if this error comes , hive tables is exported >> >> > > >>successfully >> >> > > >> >>in >> >> > > >> >> >> >>kylin. Is this the reason why cube saving has failed ? >> >> > > >> >> >> >> >> >> > > >> >> >> >> Next step when I go on creating the cube for >>following >> >> > > >>schema I >> >> > > >> >> >>get >> >> > > >> >> >> >>error at last step while saving and I am unable to >>add >> >>any >> >> > > >> >>measures , >> >> > > >> >> >> >>clicking on measure option just dont pop up any >>dialog, >> >> > > >> >> >> >> >> >> > > >> >> >> >>I am using star schema with fact_sales as fact table >>and >> >> dim_* >> >> > > >>as >> >> > > >> >> >> >>dimension tables. >> >> > > >> >> >> >> >> >> > > >> >> >> >> fact_sales: >> >> > > >> >> >> >> storeid int >> >> > > >> >> >> >> itemid int >> >> > > >> >> >> >> custid int >> >> > > >> >> >> >> qty int >> >> > > >> >> >> >> price double >> >> > > >> >> >> >> >> >> > > >> >> >> >>dim_customer >> >> > > >> >> >> >> custid int >> >> > > >> >> >> >> name string >> >> > > >> >> >> >> >> >> > > >> >> >> >> dim_item >> >> > > >> >> >> >> itemid int >> >> > > >> >> >> >> category string >> >> > > >> >> >> >> brand string >> >> > > >> >> >> >> color string >> >> > > >> >> >> >> >> >> > > >> >> >> >>dim_store >> >> > > >> >> >> >> storeid int >> >> > > >> >> >> >> city string >> >> > > >> >> >> >> state string >> >> > > >> >> >> >> >> >> > > >> >> >> >>JSON is as below. >> >> > > >> >> >> >> The JSON is as below. >> >> > > >> >> >> >> >> >> > > >> >> >> >> { >> >> > > >> >> >> >> "name": "Retail_Cube", >> >> > > >> >> >> >> "description": "", >> >> > > >> >> >> >> "dimensions": [ >> >> > > >> >> >> >> { >> >> > > >> >> >> >> "name": "RETAIL.FACT_SALES.STOREID", >> >> > > >> >> >> >> "table": "RETAIL.FACT_SALES", >> >> > > >> >> >> >> "hierarchy": false, >> >> > > >> >> >> >> "derived": null, >> >> > > >> >> >> >> "column": [ >> >> > > >> >> >> >> "STOREID" >> >> > > >> >> >> >> ], >> >> > > >> >> >> >> "id": 1 >> >> > > >> >> >> >> }, >> >> > > >> >> >> >> { >> >> > > >> >> >> >> "name": "RETAIL.FACT_SALES.ITEMID", >> >> > > >> >> >> >> "table": "RETAIL.FACT_SALES", >> >> > > >> >> >> >> "hierarchy": false, >> >> > > >> >> >> >> "derived": null, >> >> > > >> >> >> >> "column": [ >> >> > > >> >> >> >> "ITEMID" >> >> > > >> >> >> >> ], >> >> > > >> >> >> >> "id": 2 >> >> > > >> >> >> >> }, >> >> > > >> >> >> >> { >> >> > > >> >> >> >> "name": "RETAIL.FACT_SALES.CUSTID", >> >> > > >> >> >> >> "table": "RETAIL.FACT_SALES", >> >> > > >> >> >> >> "hierarchy": false, >> >> > > >> >> >> >> "derived": null, >> >> > > >> >> >> >> "column": [ >> >> > > >> >> >> >> "CUSTID" >> >> > > >> >> >> >> ], >> >> > > >> >> >> >> "id": 3 >> >> > > >> >> >> >> } >> >> > > >> >> >> >> ], >> >> > > >> >> >> >> "measures": [ >> >> > > >> >> >> >> { >> >> > > >> >> >> >> "id": 1, >> >> > > >> >> >> >> "name": "_COUNT_", >> >> > > >> >> >> >> "function": { >> >> > > >> >> >> >> "expression": "COUNT", >> >> > > >> >> >> >> "returntype": "bigint", >> >> > > >> >> >> >> "parameter": { >> >> > > >> >> >> >> "type": "constant", >> >> > > >> >> >> >> "value": 1 >> >> > > >> >> >> >> } >> >> > > >> >> >> >> } >> >> > > >> >> >> >> } >> >> > > >> >> >> >> ], >> >> > > >> >> >> >> "rowkey": { >> >> > > >> >> >> >> "rowkey_columns": [ >> >> > > >> >> >> >> { >> >> > > >> >> >> >> "column": "STOREID", >> >> > > >> >> >> >> "length": 0, >> >> > > >> >> >> >> "dictionary": "true", >> >> > > >> >> >> >> "mandatory": false >> >> > > >> >> >> >> }, >> >> > > >> >> >> >> { >> >> > > >> >> >> >> "column": "ITEMID", >> >> > > >> >> >> >> "length": 0, >> >> > > >> >> >> >> "dictionary": "true", >> >> > > >> >> >> >> "mandatory": false >> >> > > >> >> >> >> }, >> >> > > >> >> >> >> { >> >> > > >> >> >> >> "column": "CUSTID", >> >> > > >> >> >> >> "length": 0, >> >> > > >> >> >> >> "dictionary": "true", >> >> > > >> >> >> >> "mandatory": false >> >> > > >> >> >> >> } >> >> > > >> >> >> >> ], >> >> > > >> >> >> >> "aggregation_groups": [ >> >> > > >> >> >> >> [ >> >> > > >> >> >> >> "STOREID", >> >> > > >> >> >> >> "ITEMID", >> >> > > >> >> >> >> "CUSTID" >> >> > > >> >> >> >> ] >> >> > > >> >> >> >> ] >> >> > > >> >> >> >> }, >> >> > > >> >> >> >> "notify_list": [], >> >> > > >> >> >> >> "capacity": "", >> >> > > >> >> >> >> "hbase_mapping": { >> >> > > >> >> >> >> "column_family": [ >> >> > > >> >> >> >> { >> >> > > >> >> >> >> "name": "f1", >> >> > > >> >> >> >> "columns": [ >> >> > > >> >> >> >> { >> >> > > >> >> >> >> "qualifier": "m", >> >> > > >> >> >> >> "measure_refs": [ >> >> > > >> >> >> >> "_COUNT_" >> >> > > >> >> >> >> ] >> >> > > >> >> >> >> } >> >> > > >> >> >> >> ] >> >> > > >> >> >> >> } >> >> > > >> >> >> >> ] >> >> > > >> >> >> >> }, >> >> > > >> >> >> >> "project": "RetailProject", >> >> > > >> >> >> >> "model_name": "Retail_Cube" >> >> > > >> >> >> >> } >> >> > > >> >> >> >> >> >> > > >> >> >> >>Regards, >> >> > > >> >> >> >>Santosh Akhilesh >> >> > > >> >> >> >>Bangalore R&D >> >> > > >> >> >> >>HUAWEI TECHNOLOGIES CO.,LTD. >> >> > > >> >> >> >> >> >> > > >> >> >> >>www.huawei.com >> >> > > >> >> >> >> >>>>---------------------------------------------------------- >> >> > > >> >> >> --------------- >> >> > > >> >> >> >>- >> >> > > >> >> >> >> >>>>----------------------------------------------------------- >> >> > > >> >> >> >>This e-mail and its attachments contain confidential >> >> > information >> >> > > >> >>from >> >> > > >> >> >> >>HUAWEI, which >> >> > > >> >> >> >>is intended only for the person or entity whose >>address >> >>is >> >> > > >>listed >> >> > > >> >> >>above. >> >> > > >> >> >> >>Any use of the >> >> > > >> >> >> >>information contained herein in any way (including, >>but >> >>not >> >> > > >>limited >> >> > > >> >> >>to, >> >> > > >> >> >> >>total or partial >> >> > > >> >> >> >>disclosure, reproduction, or dissemination) by persons >> >>other >> >> > > >>than >> >> > > >> >>the >> >> > > >> >> >> >>intended >> >> > > >> >> >> >>recipient(s) is prohibited. If you receive this >>e-mail in >> >> > error, >> >> > > >> >> >>please >> >> > > >> >> >> >>notify the sender by >> >> > > >> >> >> >>phone or email immediately and delete it! >> >> > > >> >> >> >> >> >> > > >> >> >> >>________________________________________ >> >> > > >> >> >> >>From: Shi, Shaofeng [[email protected]] >> >> > > >> >> >> >>Sent: Thursday, February 26, 2015 7:01 AM >> >> > > >> >> >> >>To: [email protected] >> >> > > >> >> >> >>Subject: Re: Error while making cube & Measure option >>is >> >>not >> >> > > >> >> >>responding >> >> > > >> >> >> >>on GUI >> >> > > >> >> >> >> >> >> > > >> >> >> >>Hi Santosh, >> >> > > >> >> >> >> >> >> > > >> >> >> >>It looks like hadoop failed to execute some shell >> >>command in >> >> > the >> >> > > >> >> >> >>container; You need dive into hadoop to see what¹s the >> >> > concrete >> >> > > >> >>error. >> >> > > >> >> >> >>You >> >> > > >> >> >> >>can use yarn logs command to fetch all logs: >> >> > > >> >> >> >> >> >> > > >> >> >> >>yarn logs -applicationId <app_id> >> >> > > >> >> >> >> >> >> > > >> >> >> >> >> >> > > >> >> >> >>On 2/25/15, 7:39 PM, "Santosh Akhilesh" >> >> > > >><[email protected] >> >> > > >> > >> >> > > >> >> >> >>wrote: >> >> > > >> >> >> >> >> >> > > >> >> >> >>>Hi Luke / Shaofeng , >> >> > > >> >> >> >>> Can you please help me to check this >>issue. >> >> > > >> >> >> >>>Regards, >> >> > > >> >> >> >>>Santosh Akhilesh >> >> > > >> >> >> >>> >> >> > > >> >> >> >>>On Tue, Feb 24, 2015 at 10:41 PM, Santosh Akhilesh < >> >> > > >> >> >> >>>[email protected]> wrote: >> >> > > >> >> >> >>> >> >> > > >> >> >> >>>> Hi All , >> >> > > >> >> >> >>>> is it because of following error in map >>reduce >> >> job >> >> > ? >> >> > > >> >>what >> >> > > >> >> >> >>>>could >> >> > > >> >> >> >>>>be >> >> > > >> >> >> >>>> way to resolve this , a google search says that its >> >>issue >> >> > of >> >> > > >> >>Yarn >> >> > > >> >> >> >>>>class >> >> > > >> >> >> >>>> path , but I am not sure what it is ? >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> Kylin Hive Column Cardinality Job >> >>table=RETAIL.FACT_SALES >> >> > > >> >> >> >>>> output=/tmp/cardinality/RETAIL.FACT_SALES >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> Application application_1424791969399_0008 failed 2 >> >>times >> >> > due >> >> > > >> >>to AM >> >> > > >> >> >> >>>> Container for appattempt_1424791969399_0008_000002 >> >>exited >> >> > > >>with >> >> > > >> >> >> >>>>exitCode: 1 >> >> > > >> >> >> >>>> For more detailed output, check application >>tracking >> >> page: >> >> > > >> >> >> >>>> >> >> > > >>http://santosh:8088/proxy/application_1424791969399_0008/Then, >> >> > > >> >> >>click >> >> > > >> >> >> >>>>on >> >> > > >> >> >> >>>> links to logs of each attempt. >> >> > > >> >> >> >>>> Diagnostics: Exception from container-launch. >> >> > > >> >> >> >>>> Container id: >>container_1424791969399_0008_02_000001 >> >> > > >> >> >> >>>> Exit code: 1 >> >> > > >> >> >> >>>> Stack trace: ExitCodeException exitCode=1: >> >> > > >> >> >> >>>> at >> >> org.apache.hadoop.util.Shell.runCommand(Shell.java:538) >> >> > > >> >> >> >>>> at org.apache.hadoop.util.Shell.run(Shell.java:455) >> >> > > >> >> >> >>>> at >> >> > > >> >> >> >>>> >> >> > > >> >> >> >> >>>>>>org.apache.hadoop.util.Shell$ShellCommandExecutor.execut >> >> > > >> >> >> e(Shell.java:71 >> >> > > >> >> >> >>>>5 >> >> > > >> >> >> >>>>) >> >> > > >> >> >> >>>> at >> >> > > >> >> >> >>>> >> >> > > >> >> >> >> >>>>>>org.apache.hadoop.yarn.server.nodemanager.DefaultContain >> >> > > >> >> >> erExecutor.laun >> >> > > >> >> >> >>>>c >> >> > > >> >> >> >>>>h >> >> > > >> >> >> >>>>Container(DefaultContainerExecutor.java:211) >> >> > > >> >> >> >>>> at >> >> > > >> >> >> >>>> >> >> > > >> >> >> >> >>>>>>org.apache.hadoop.yarn.server.nodemanager.containermanag >> >> > > >> >> >> er.launcher.Con >> >> > > >> >> >> >>>>t >> >> > > >> >> >> >>>>a >> >> > > >> >> >> >>>>inerLaunch.call(ContainerLaunch.java:302) >> >> > > >> >> >> >>>> at >> >> > > >> >> >> >>>> >> >> > > >> >> >> >> >>>>>>org.apache.hadoop.yarn.server.nodemanager.containermanag >> >> > > >> >> >> er.launcher.Con >> >> > > >> >> >> >>>>t >> >> > > >> >> >> >>>>a >> >> > > >> >> >> >>>>inerLaunch.call(ContainerLaunch.java:82) >> >> > > >> >> >> >>>> at >> >> java.util.concurrent.FutureTask.run(FutureTask.java:262) >> >> > > >> >> >> >>>> at >> >> > > >> >> >> >>>> >> >> > > >> >> >> >> >>>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread >> >> > > >> >> >> PoolExecutor.ja >> >> > > >> >> >> >>>>v >> >> > > >> >> >> >>>>a >> >> > > >> >> >> >>>>:1145) >> >> > > >> >> >> >>>> at >> >> > > >> >> >> >>>> >> >> > > >> >> >> >> >>>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa >> >> > > >> >> >> dPoolExecutor.j >> >> > > >> >> >> >>>>a >> >> > > >> >> >> >>>>v >> >> > > >> >> >> >>>>a:615) >> >> > > >> >> >> >>>> at java.lang.Thread.run(Thread.java:745) >> >> > > >> >> >> >>>> Container exited with a non-zero exit code 1 >> >> > > >> >> >> >>>> Failing this attempt. Failing the application. >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> ---------- Forwarded message ---------- >> >> > > >> >> >> >>>> From: Santoshakhilesh <[email protected]> >> >> > > >> >> >> >>>> Date: Tue, Feb 24, 2015 at 7:41 PM >> >> > > >> >> >> >>>> Subject: FW: Error while making cube & Measure >>option >> >>is >> >> > not >> >> > > >> >> >> >>>>responding >> >> > > >> >> >> >>>>on >> >> > > >> >> >> >>>> GUI >> >> > > >> >> >> >>>> To: "[email protected]" >> >> > > >> >> >><[email protected]> >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> hi , >> >> > > >> >> >> >>>> please someone give me a hand to resolve this >> >>issue , >> >> > > >>thanks. >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> Regards, >> >> > > >> >> >> >>>> Santosh Akhilesh >> >> > > >> >> >> >>>> Bangalore R&D >> >> > > >> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD. >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> www.huawei.com >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> >> >> > > >> >> >> >> >>>>>>-------------------------------------------------------- >> >> > > >> >> >> --------------- >> >> > > >> >> >> >>>>- >> >> > > >> >> >> >>>>- >> >> > > >> >> >> >> >> >>>>------------------------------------------------------------ >> >> > > >> >> >> >>>> This e-mail and its attachments contain >>confidential >> >> > > >>information >> >> > > >> >> >>from >> >> > > >> >> >> >>>> HUAWEI, which >> >> > > >> >> >> >>>> is intended only for the person or entity whose >> >>address >> >> is >> >> > > >> >>listed >> >> > > >> >> >> >>>>above. >> >> > > >> >> >> >>>> Any use of the >> >> > > >> >> >> >>>> information contained herein in any way (including, >> >>but >> >> not >> >> > > >> >>limited >> >> > > >> >> >> >>>>to, >> >> > > >> >> >> >>>> total or partial >> >> > > >> >> >> >>>> disclosure, reproduction, or dissemination) by >>persons >> >> > other >> >> > > >> >>than >> >> > > >> >> >>the >> >> > > >> >> >> >>>> intended >> >> > > >> >> >> >>>> recipient(s) is prohibited. If you receive this >> >>e-mail in >> >> > > >>error, >> >> > > >> >> >> >>>>please >> >> > > >> >> >> >>>> notify the sender by >> >> > > >> >> >> >>>> phone or email immediately and delete it! >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> ________________________________________ >> >> > > >> >> >> >>>> From: Santoshakhilesh [[email protected]] >> >> > > >> >> >> >>>> Sent: Tuesday, February 24, 2015 12:55 PM >> >> > > >> >> >> >>>> To: [email protected] >> >> > > >> >> >> >>>> Cc: Kulbhushan Rana >> >> > > >> >> >> >>>> Subject: FW: Error while making cube & Measure >>option >> >>is >> >> > not >> >> > > >> >> >> >>>>responding >> >> > > >> >> >> >>>>on >> >> > > >> >> >> >>>> GUI >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> 2. If I ignore and continue and try to save the >>cube >> >>I >> >> get >> >> > > >>an >> >> > > >> >> >> >>>>exception >> >> > > >> >> >> >>>> in Kylin.log , I have checked the path is set >> >>correctly >> >> and >> >> > > >> >> >> >>>>HCatInputFormat >> >> > > >> >> >> >>>> this file is present in >>hive-hcatalog-core-0.14.0.jar >> >>. >> >> > > >>Please >> >> > > >> >>let >> >> > > >> >> >>me >> >> > > >> >> >> >>>>know >> >> > > >> >> >> >>>> what can I do to resolve this ? >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> -- This was path issue , now no more exception in >> >> > kylin.log >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> But saveing cube still fails with error. And still >> >>can't >> >> > add >> >> > > >> >> >>measures. >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> Error Message >> >> > > >> >> >> >>>> Failed to take action. >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> In log I can find no exception. Following is the >>last >> >>log >> >> > in >> >> > > >> >> >>kylin.log >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> [pool-3-thread-1]:[2015-02-24 >> >> > > >> >> >> >>>> >> >> > > >> >> >> >> >>>>>>20:47:15,613][INFO][org.apache.kylin.job.impl.threadpool >> >> > > >> >> >> .DefaultSchedul >> >> > > >> >> >> >>>>e >> >> > > >> >> >> >>>>r >> >> > > >> >> >> >>>>$FetcherRunner.run(DefaultScheduler.java:117)] >> >> > > >> >> >> >>>> - Job Fetcher: 0 running, 0 actual running, 0 >>ready, 6 >> >> > others >> >> > > >> >> >> >>>> [http-bio-7070-exec-2]:[2015-02-24 >> >> > > >> >> >> >>>> >> >> > > >> >> >> >> >>>>>>20:47:51,610][DEBUG][org.apache.kylin.rest.controller.Cu >> >> > > >> >> >> beController.de >> >> > > >> >> >> >>>>s >> >> > > >> >> >> >>>>e >> >> > > >> >> >> >>>>rializeDataModelDesc(CubeController.java:459)] >> >> > > >> >> >> >>>> - Saving cube { >> >> > > >> >> >> >>>> "name": "", >> >> > > >> >> >> >>>> "fact_table": "RETAIL.FACT_SALES", >> >> > > >> >> >> >>>> "lookups": [], >> >> > > >> >> >> >>>> "filter_condition": "", >> >> > > >> >> >> >>>> "capacity": "SMALL", >> >> > > >> >> >> >>>> "partition_desc": { >> >> > > >> >> >> >>>> "partition_date_column": "", >> >> > > >> >> >> >>>> "partition_date_start": 0, >> >> > > >> >> >> >>>> "partition_type": "APPEND" >> >> > > >> >> >> >>>> }, >> >> > > >> >> >> >>>> "last_modified": 0 >> >> > > >> >> >> >>>> } >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> local access logs all with 200 , so seems ok. >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:46:56 +0800] "GET >> >> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >> >> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:07 +0800] "GET >> >> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >> >> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:27 +0800] "GET >> >> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >> >> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:28 +0800] "GET >> >> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >> >> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:34 +0800] "GET >> >> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >> >> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:48 +0800] "GET >> >> > > >> >> >> >>>> /kylin/api/user/authentication HTTP/1.1" 200 246 >> >> > > >> >> >> >>>> 10.18.146.105 - - [24/Feb/2015:20:47:51 +0800] >>"POST >> >> > > >> >> >>/kylin/api/cubes >> >> > > >> >> >> >>>> HTTP/1.1" 200 701 >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> Regards, >> >> > > >> >> >> >>>> Santosh Akhilesh >> >> > > >> >> >> >>>> Bangalore R&D >> >> > > >> >> >> >>>> HUAWEI TECHNOLOGIES CO.,LTD. >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> www.huawei.com >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> >> >> > > >> >> >> >> >>>>>>-------------------------------------------------------- >> >> > > >> >> >> --------------- >> >> > > >> >> >> >>>>- >> >> > > >> >> >> >>>>- >> >> > > >> >> >> >> >> >>>>------------------------------------------------------------ >> >> > > >> >> >> >>>> This e-mail and its attachments contain >>confidential >> >> > > >>information >> >> > > >> >> >>from >> >> > > >> >> >> >>>> HUAWEI, which >> >> > > >> >> >> >>>> is intended only for the person or entity whose >> >>address >> >> is >> >> > > >> >>listed >> >> > > >> >> >> >>>>above. >> >> > > >> >> >> >>>> Any use of the >> >> > > >> >> >> >>>> information contained herein in any way (including, >> >>but >> >> not >> >> > > >> >>limited >> >> > > >> >> >> >>>>to, >> >> > > >> >> >> >>>> total or partial >> >> > > >> >> >> >>>> disclosure, reproduction, or dissemination) by >>persons >> >> > other >> >> > > >> >>than >> >> > > >> >> >>the >> >> > > >> >> >> >>>> intended >> >> > > >> >> >> >>>> recipient(s) is prohibited. If you receive this >> >>e-mail in >> >> > > >>error, >> >> > > >> >> >> >>>>please >> >> > > >> >> >> >>>> notify the sender by >> >> > > >> >> >> >>>> phone or email immediately and delete it! >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> ________________________________________ >> >> > > >> >> >> >>>> From: Santoshakhilesh [[email protected]] >> >> > > >> >> >> >>>> Sent: Tuesday, February 24, 2015 12:09 PM >> >> > > >> >> >> >>>> To: [email protected] >> >> > > >> >> >> >>>> Cc: Kulbhushan Rana >> >> > > >> >> >> >>>> Subject: Error while making cube & Measure option >>is >> >>not >> >> > > >> >> >>responding on >> >> > > >> >> >> >>>>GUI >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> Hi All , >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> I am building a simple cube for test and using >>the >> >> > binary >> >> > > >> >>build >> >> > > >> >> >> >>>>0.7.1 >> >> > > >> >> >> >>>> . I have following hive tables with columns. >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> fact_sales: >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> storeid int >> >> > > >> >> >> >>>> itemid int >> >> > > >> >> >> >>>> custid int >> >> > > >> >> >> >>>> qty int >> >> > > >> >> >> >>>> price double >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> dim_customer >> >> > > >> >> >> >>>> custid int >> >> > > >> >> >> >>>> name string >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> dim_item >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> itemid int >> >> > > >> >> >> >>>> category string >> >> > > >> >> >> >>>> brand string >> >> > > >> >> >> >>>> color string >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> dim_store >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> storeid int >> >> > > >> >> >> >>>> city string >> >> > > >> >> >> >>>> state string >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> Please help me to answer following issues; >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> 1. When I go to measure section and click on >>measure >> >> > option , >> >> > > >> >> >>there is >> >> > > >> >> >> >>>>no >> >> > > >> >> >> >>>> response , I want add measure on qty and price with >> >>sum >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> 2. If I ignore and continue and try to save the >>cube >> >>I >> >> get >> >> > > >>an >> >> > > >> >> >> >>>>exception >> >> > > >> >> >> >>>> in Kylin.log , I have checked the path is set >> >>correctly >> >> and >> >> > > >> >> >> >>>>HCatInputFormat >> >> > > >> >> >> >>>> this file is present in >>hive-hcatalog-core-0.14.0.jar >> >>. >> >> > > >>Please >> >> > > >> >>let >> >> > > >> >> >>me >> >> > > >> >> >> >>>>know >> >> > > >> >> >> >>>> what can I do to resolve this ? >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> 3. Also I have another question since this is a >>test >> >>and >> >> > > >>data is >> >> > > >> >> >>small >> >> > > >> >> >> >>>>I >> >> > > >> >> >> >>>> have not partitioned the fact table , is it ok to >>skip >> >> > > >>partition >> >> > > >> >> >>stage >> >> > > >> >> >> >>>> while cube build ? >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> Exception >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> pool-4-thread-4]:[2015-02-24 >> >> > > >> >> >> >>>> >> >> > > >> >> >> >> >>>>>>19:26:32,577][ERROR][org.apache.kylin.job.impl.threadpoo >> >> > > >> >> >> l.DefaultSchedu >> >> > > >> >> >> >>>>l >> >> > > >> >> >> >>>>e >> >> > > >> >> >> >>>>r$JobRunner.run(DefaultScheduler.java:134)] >> >> > > >> >> >> >>>> - ExecuteException >> >> job:c3532a6f-97ea-474a-b36a-218dd517cedb >> >> > > >> >> >> >>>> org.apache.kylin.job.exception.ExecuteException: >> >> > > >> >> >> >>>> org.apache.kylin.job.exception.ExecuteException: >> >> > > >> >> >> >>>> java.lang.NoClassDefFoundError: >> >> > > >> >> >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat >> >> > > >> >> >> >>>> at >> >> > > >> >> >> >>>> >> >> > > >> >> >> >> >>>>>>org.apache.kylin.job.execution.AbstractExecutable.execut >> >> > > >> >> >> e(AbstractExecu >> >> > > >> >> >> >>>>t >> >> > > >> >> >> >>>>a >> >> > > >> >> >> >>>>ble.java:102) >> >> > > >> >> >> >>>> at >> >> > > >> >> >> >>>> >> >> > > >> >> >> >> >>>>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$Jo >> >> > > >> >> >> bRunner.run(Def >> >> > > >> >> >> >>>>a >> >> > > >> >> >> >>>>u >> >> > > >> >> >> >>>>ltScheduler.java:132) >> >> > > >> >> >> >>>> at >> >> > > >> >> >> >>>> >> >> > > >> >> >> >> >>>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(Thread >> >> > > >> >> >> PoolExecutor.ja >> >> > > >> >> >> >>>>v >> >> > > >> >> >> >>>>a >> >> > > >> >> >> >>>>:1145) >> >> > > >> >> >> >>>> at >> >> > > >> >> >> >>>> >> >> > > >> >> >> >> >>>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(Threa >> >> > > >> >> >> dPoolExecutor.j >> >> > > >> >> >> >>>>a >> >> > > >> >> >> >>>>v >> >> > > >> >> >> >>>>a:615) >> >> > > >> >> >> >>>> at java.lang.Thread.run(Thread.java:745) >> >> > > >> >> >> >>>> Caused by: >> >> org.apache.kylin.job.exception.ExecuteException: >> >> > > >> >> >> >>>> java.lang.NoClassDefFoundError: >> >> > > >> >> >> >>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> The JSON is as below. >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> >> >> > > >> >> >> >>>> { >> >> > > >> >> >> >>>> "name": "Retail_Cube", >> >> > > >> >> >> >>>> "description": "", >> >> > > >> >> >> >>>> "dimensions": [ >> >> > > >> >> >> >>>> { >> >> > > >> >> >> >>>> "name": "RETAIL.FACT_SALES.STOREID", >> >> > > >> >> >> >>>> "table": "RETAIL.FACT_SALES", >> >> > > >> >> >> >>>> "hierarchy": false, >> >> > > >> >> >> >>>> "derived": null, >> >> > > >> >> >> >>>> "column": [ >> >> > > >> >> >> >>>> "STOREID" >> >> > > >> >> >> >>>> ], >> >> > > >> >> >> >>>> "id": 1 >> >> > > >> >> >> >>>> }, >> >> > > >> >> >> >>>> { >> >> > > >> >> >> >>>> "name": "RETAIL.FACT_SALES.ITEMID", >> >> > > >> >> >> >>>> "table": "RETAIL.FACT_SALES", >> >> > > >> >> >> >>>> "hierarchy": false, >> >> > > >> >> >> >>>> "derived": null, >> >> > > >> >> >> >>>> "column": [ >> >> > > >> >> >> >>>> "ITEMID" >> >> > > >> >> >> >>>> ], >> >> > > >> >> >> >>>> "id": 2 >> >> > > >> >> >> >>>> }, >> >> > > >> >> >> >>>> { >> >> > > >> >> >> >>>> "name": "RETAIL.FACT_SALES.CUSTID", >> >> > > >> >> >> >>>> "table": "RETAIL.FACT_SALES", >> >> > > >> >> >> >>>> "hierarchy": false, >> >> > > >> >> >> >>>> "derived": null, >> >> > > >> >> >> >>>> "column": [ >> >> > > >> >> >> >>>> "CUSTID" >> >> > > >> >> >> >>>> ], >> >> > > >> >> >> >>>> "id": 3 >> >> > > >> >> >> >>>> } >> >> > > >> >> >> >>>> ], >> >> > > >> >> >> >>>> "measures": [ >> >> > > >> >> >> >>>> { >> >> > > >> >> >> >>>> "id": 1, >> >> > > >> >> >> >>>> "name": "_COUNT_", >> >> > > >> >> >> >>>> "function": { >> >> > > >> >> >> >>>> "expression": "COUNT", >> >> > > >> >> >> >>>> "returntype": "bigint", >> >> > > >> >> >> >>>> "parameter": { >> >> > > >> >> >> >>>> "type": "constant", >> >> > > >> >> >> >>>> "value": 1 >> >> > > >> >> >> >>>> } >> >> > > >> >> >> >>>> } >> >> > > >> >> >> >>>> } >> >> > > >> >> >> >>>> ], >> >> > > >> >> >> >>>> "rowkey": { >> >> > > >> >> >> >>>> "rowkey_columns": [ >> >> > > >> >> >> >>>> { >> >> > > >> >> >> >>>> "column": "STOREID", >> >> > > >> >> >> >>>> "length": 0, >> >> > > >> >> >> >>>> "dictionary": "true", >> >> > > >> >> >> >>>> "mandatory": false >> >> > > >> >> >> >>>> }, >> >> > > >> >> >> >>>> { >> >> > > >> >> >> >>>> "column": "ITEMID", >> >> > > >> >> >> >>>> "length": 0, >> >> > > >> >> >> >>>> "dictionary": "true", >> >> > > >> >> >> >>>> "mandatory": false >> >> > > >> >> >> >>>> }, >> >> > > >> >> >> >>>> { >> >> > > >> >> >> >>>> "column": "CUSTID", >> >> > > >> >> >> >>>> "length": 0, >> >> > > >> >> >> >>>> "dictionary": "true", >> >> > > >> >> >> >>>> "mandatory": false >> >> > > >> >> >> >>>> } >> >> > > >> >> >> >>>> ], >> >> > > >> >> >> >>>> "aggregation_groups": [ >> >> > > >> >> >> >>>> [ >> >> > > >> >> >> >>>> "STOREID", >> >> > > >> >> >> >>>> "ITEMID", >> >> > > >> >> >> >>>> "CUSTID" >> >> > > >> >> >> >>>> ] >> >> > > >> >> >> >>>> ] >> >> > > >> >> >> >>>> }, >> >> > > >> >> >> >>>> "notify_list": [], >> >> > > >> >> >> >>>> "capacity": "", >> >> > > >> >> >> >>>> "hbase_mapping": { >> >> > > >> >> >> >>>> "column_family": [ >> >> > > >> >> >> >>>> { >> >> > > >> >> >> >>>> "name": "f1", >> >> > > >> >> >> >>>> "columns": [ >> >> > > >> >> >> >>>> { >> >> > > >> >> >> >>>> "qualifier": "m", >> >> > > >> >> >> >>>> "measure_refs": [ >> >> > > >> >> >> >>>> "_COUNT_" >> >> > > >> >> >> >>>> ] >> >> > > >> >> >> >>>> } >> >> > > >> >> >> >>>> ] >> >> > > >> >> >>
