OK,i will re-install CDH and hope not to have these problems. Than you very much!
2017-06-09 15:16 GMT+08:00 ShaoFeng Shi <shaofeng...@apache.org>: > It looks like another jar version mismatch error. As I mentioned > previously, if you install these Hadoop components separately, there are > kinds of such error. You need take the risk and spend the time to fix them > one by one. Such issue usually is out of Kylin scope and I may not help. > > 2017-06-09 14:27 GMT+08:00 周湘伦 <jayzho...@gmail.com>: > > > HI,ShaoFeng > > > > I had replaced hbase-server-1.1.1.jar which in hive/lib > > to hbase-server-1.2.5.jar. > > The above problem has been solved. > > Thank you! > > > > But another problem arises: > > > > 14:16:52.441 [Job 90064ee6-2b3a-4ef7-b035-33f40391aafb-141] ERROR > > org.apache.kylin.job.execution.AbstractExecutable - error running > > Executable: MapReduceExecutable{id=90064ee6-2b3a-4ef7-b035- > > 33f40391aafb-15, > > name=Convert Cuboid Data to HFile, state=RUNNING} > > 14:16:52.496 [Job 90064ee6-2b3a-4ef7-b035-33f40391aafb-141] ERROR > > org.apache.kylin.job.execution.AbstractExecutable - error running > > Executable: CubingJob{id=90064ee6-2b3a-4ef7-b035-33f40391aafb, > > name=Kylin_Sample_cube_1 - 20120101000000_20130101000000 - BUILD - > > GMT+08:00 2017-06-08 15:06:49, state=RUNNING} > > 14:16:52.537 [pool-9-thread-1] ERROR > > org.apache.kylin.job.impl.threadpool.DefaultScheduler - ExecuteException > > job:90064ee6-2b3a-4ef7-b035-33f40391aafb > > org.apache.kylin.job.exception.ExecuteException: > > org.apache.kylin.job.exception.ExecuteException: > > java.lang.NoSuchFieldError: DEFAULT_TEMPORARY_HDFS_DIRECTORY > > at > > org.apache.kylin.job.execution.AbstractExecutable. > > execute(AbstractExecutable.java:134) > > ~[kylin-core-job-2.0.0.jar:2.0.0] > > at > > org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run( > > DefaultScheduler.java:142) > > [kylin-core-job-2.0.0.jar:2.0.0] > > at > > java.util.concurrent.ThreadPoolExecutor.runWorker( > > ThreadPoolExecutor.java:1142) > > [?:1.8.0_131] > > at > > java.util.concurrent.ThreadPoolExecutor$Worker.run( > > ThreadPoolExecutor.java:617) > > [?:1.8.0_131] > > at java.lang.Thread.run(Thread.java:748) [?:1.8.0_131] > > Caused by: org.apache.kylin.job.exception.ExecuteException: > > java.lang.NoSuchFieldError: DEFAULT_TEMPORARY_HDFS_DIRECTORY > > at > > org.apache.kylin.job.execution.AbstractExecutable. > > execute(AbstractExecutable.java:134) > > ~[kylin-core-job-2.0.0.jar:2.0.0] > > at > > org.apache.kylin.job.execution.DefaultChainedExecutable.doWork( > > DefaultChainedExecutable.java:64) > > ~[kylin-core-job-2.0.0.jar:2.0.0] > > at > > org.apache.kylin.job.execution.AbstractExecutable. > > execute(AbstractExecutable.java:124) > > ~[kylin-core-job-2.0.0.jar:2.0.0] > > ... 4 more > > > > > > I had configured this option in hbase-site.xml: > > <property> > > <name>hbase.fs.tmp.dir</name> > > <value>hdfs://master:9000/hbase/tmp/hadoop-staging</value> > > <!-- <value>/tmp/hadoop-staging</value>--> > > <description>i have tried these configurations</description> > > </property> > > > > I had even configured this option in kylin_job_conf.xml. > > But there is no effect. > > > > Of course i have created director in local file system and hdfs. > > > > Can you tell me how to solve this problem? > > Than you! > > > > 2017-06-05 13:04 GMT+08:00 ShaoFeng Shi <shaofeng...@apache.org>: > > > > > The method "getChecksumObject" does not exist in " > > > org.apache.hadoop.hbase.util.ChecksumType" in hbase 1.2.5. While in > your > > > environment the "org.apache.hadoop.hbase.io.hfile.ChecksumUtil" still > > > invoking to it. It indicates there is a "hbase-server*.jar" which isn't > > > v1.2.5. Please search your environment globally to identify and remove > > it. > > > > > > Usually we suggest using a formal release of CDH/HDP/MapR, which won't > > have > > > such version conflict. If you install these components separately, > there > > > will be such environment issues. > > > > > > 2017-06-03 18:29 GMT+08:00 周湘伦 <jayzho...@gmail.com>: > > > > > > > Hi,all > > > > When i built cube,an error occurred in step 16(Convert Cuboid Data to > > > > HFile). > > > > > > > > The version as belows: > > > > hadoop-2.8.0,hbase-1.2.5,jdk1.8.0_131,kylin-2.0.0 > > > > > > > > The error logs in hadoop/userlogs,the log is shown below: > > > > > > > > 2017-06-03 17:57:04,106 FATAL [main] org.apache.hadoop.mapred. > > YarnChild: > > > > Error running child : java.lang.NoSuchMethodError: > > > > org.apache.hadoop.hbase.util.ChecksumType. > > getChecksumObject()Ljava/util/ > > > > zip/Checksum; > > > > at > > > > org.apache.hadoop.hbase.io.hfile.ChecksumUtil.generateChecksums( > > > > ChecksumUtil.java:73) > > > > at > > > > org.apache.hadoop.hbase.io.hfile.HFileBlock$Writer. > > > > finishBlock(HFileBlock.java:943) > > > > at > > > > org.apache.hadoop.hbase.io.hfile.HFileBlock$Writer. > > > > ensureBlockReady(HFileBlock.java:895) > > > > at > > > > org.apache.hadoop.hbase.io.hfile.HFileBlock$Writer. > > > > finishBlockAndWriteHeaderAndData(HFileBlock.java:1011) > > > > at > > > > org.apache.hadoop.hbase.io.hfile.HFileBlock$Writer. > > > > writeHeaderAndData(HFileBlock.java:997) > > > > at > > > > org.apache.hadoop.hbase.io.hfile.HFileBlockIndex$BlockIndexWriter. > > > > writeIndexBlocks(HFileBlockIndex.java:883) > > > > at > > > > org.apache.hadoop.hbase.io.hfile.HFileWriterV2.close( > > > > HFileWriterV2.java:331) > > > > at > > > > org.apache.hadoop.hbase.regionserver.StoreFile$Writer. > > > > close(StoreFile.java:996) > > > > at > > > > org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2$ > > > > 1.close(HFileOutputFormat2.java:269) > > > > at > > > > org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2$ > > > > 1.close(HFileOutputFormat2.java:277) > > > > at > > > > org.apache.hadoop.mapred.ReduceTask$NewTrackingRecordWriter.close( > > > > ReduceTask.java:550) > > > > at > > > > org.apache.hadoop.mapred.ReduceTask.runNewReducer( > ReduceTask.java:629) > > > > at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask. > > java:389) > > > > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild. > > java:175) > > > > at java.security.AccessController.doPrivileged(Native > Method) > > > > at javax.security.auth.Subject.doAs(Subject.java:422) > > > > at > > > > org.apache.hadoop.security.UserGroupInformation.doAs( > > > > UserGroupInformation.java:1807) > > > > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java: > 169) > > > > > > > > According to others' suggest,i had copy jars which in hadoop-2.8.0 to > > > > hbase-1.2.5,but the problem is still not resolved. > > > > > > > > How can we solve the problem? > > > > > > > > Thanks a lot. > > > > > > > > > > > > > > > > -- > > > Best regards, > > > > > > Shaofeng Shi 史少锋 > > > > > > > > > -- > Best regards, > > Shaofeng Shi 史少锋 >