Thank you! We checked the yarn and hard disk. But not found any error. Hard disk space and memory and so on is working well. Last time its error code was "unknownhost clusterB",now in new server env it cant find clusterB(hbase only). but cant find rowCount file. === the follow command runs ok: hdfs dfs -mkdir /young/kylin_test/kylin_metadata_nokia/kylin-678c15ba-5375-4f80-831e-1ae0af8ed576/row_count/tmp
And "ls" cant find file "000000_0" which it said "file does not exist". ------------------ ???????? ------------------ ??????: "Alberto Ram??n";<a.ramonporto...@gmail.com>; ????????: 2016??12??19??(??????) ????9:13 ??????: "user"<user@kylin.apache.org>; ????: Re: Error when #2 Step: Redistribute Flat Hive Table - File does not exist i think i had this error last nigth :) (go to yarn to find detailed error & find on internet) in my case was free space less than 10% of hard disk. Check this please El 19/12/2016 11:35, "????????" <491245...@qq.com> escribi??: When I build a cube in kylin1.6, I get error in step #2: Redistribute Flat Hive Table Please help! Thank you very much! env: kylin1.6 is in a independent server, and have 2 other server cluster: clusterA(hive only) and clusterB(hbase only). Error is: 2016-12-19 10:28:00,641 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : Compute row count of flat hive table, cmd: 2016-12-19 10:28:00,642 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : hive -e "USE boco; SET dfs.replication=2; SET hive.exec.compress.output=true; SET hive.auto.convert.join.noconditionaltask=true; SET hive.auto.convert.join.noconditionaltask.size=100000000; SET mapreduce.output.fileoutputformat.compress.type=BLOCK; SET mapreduce.job.split.metainfo.maxsize=-1; SET mapreduce.job.queuename=young; SET tez.queue.name=young; set hive.exec.compress.output=false; set hive.exec.compress.output=false; INSERT OVERWRITE DIRECTORY '/young/kylin_test/kylin_metadata_test/kylin-678266c0-ba0e-48b4-bdb5-6e578320375a/row_count' SELECT count(*) FROM kylin_intermediate_hbase_in_testCluster_CUBE_f9468805_eabf_4b54_bf2b_182e4c86214a; " 2016-12-19 10:28:03,277 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : WARNING: Use "yarn jar" to launch YARN applications. 2016-12-19 10:28:04,444 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : 2016-12-19 10:28:04,445 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : Logging initialized using configuration in file:/etc/hive/conf/hive-log4j.properties 2016-12-19 10:28:14,700 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : OK 2016-12-19 10:28:14,703 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : Time taken: 0.935 seconds 2016-12-19 10:28:15,559 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : Query ID = young_20161219102814_a7104fd4-ba83-47fc-ac0b-0c9bef4e1969 2016-12-19 10:28:15,560 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : Total jobs = 1 2016-12-19 10:28:15,575 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : Launching Job 1 out of 1 2016-12-19 10:28:22,842 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : 2016-12-19 10:28:22,842 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : 2016-12-19 10:28:23,104 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : Status: Running (Executing on YARN cluster with App id application_1473415773736_1063281) 2016-12-19 10:28:23,104 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : 2016-12-19 10:28:23,104 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : Map 1: -/- Reducer 2: 0/1 2016-12-19 10:28:23,307 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : Map 1: 0/2 Reducer 2: 0/1 2016-12-19 10:28:26,363 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : Map 1: 0/2 Reducer 2: 0/1 2016-12-19 10:28:26,567 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : Map 1: 0(+1)/2 Reducer 2: 0/1 2016-12-19 10:28:26,596 INFO [pool-7-thread-1] threadpool.DefaultScheduler:118 : Job Fetcher: 1 should running, 1 actual running, 0 ready, 0 already succeed, 3 error, 1 discarded, 0 others 2016-12-19 10:28:26,769 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : Map 1: 0(+2)/2 Reducer 2: 0/1 2016-12-19 10:28:29,810 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : Map 1: 0(+2)/2 Reducer 2: 0/1 2016-12-19 10:28:30,217 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : Map 1: 1(+1)/2 Reducer 2: 0(+1)/1 2016-12-19 10:28:30,826 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : Map 1: 2/2 Reducer 2: 0(+1)/1 2016-12-19 10:28:31,232 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : Map 1: 2/2 Reducer 2: 1/1 2016-12-19 10:28:31,319 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : Moving data to: /young/kylin_test/kylin_metadata_test/kylin-678266c0-ba0e-48b4-bdb5-6e578320375a/row_count 2016-12-19 10:28:31,406 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : OK 2016-12-19 10:28:31,454 INFO [pool-8-thread-7] execution.AbstractExecutable:36 : Time taken: 16.701 seconds 2016-12-19 10:28:35,074 ERROR [pool-8-thread-7] execution.AbstractExecutable:357 : job:678266c0-ba0e-48b4-bdb5-6e578320375a-01 execute finished with exception java.io.FileNotFoundException: File does not exist: /young/kylin_test/kylin_metadata_test/kylin-678266c0-ba0e-48b4-bdb5-6e578320375a/row_count/000000_0 at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:71) at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:61) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsInt(FSNamesystem.java:1820) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1791) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1704) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:587) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:365) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2081) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2077) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2075) at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:526) at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106) at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:73) at org.apache.hadoop.hdfs.DFSClient.callGetBlockLocations(DFSClient.java:1242) at org.apache.hadoop.hdfs.DFSClient.getLocatedBlocks(DFSClient.java:1227) at org.apache.hadoop.hdfs.DFSClient.getLocatedBlocks(DFSClient.java:1215) at org.apache.hadoop.hdfs.DFSInputStream.fetchLocatedBlocksAndGetLastBlockLength(DFSInputStream.java:303) at org.apache.hadoop.hdfs.DFSInputStream.openInfo(DFSInputStream.java:269) at org.apache.hadoop.hdfs.DFSInputStream.<init>(DFSInputStream.java:261) at org.apache.hadoop.hdfs.DFSClient.open(DFSClient.java:1540) at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:303) at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:299) at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) at org.apache.hadoop.hdfs.DistributedFileSystem.open(DistributedFileSystem.java:299) at org.apache.hadoop.fs.FileSystem.open(FileSystem.java:767) at org.apache.kylin.source.hive.HiveMRInput$RedistributeFlatHiveTableStep.readRowCountFromFile(HiveMRInput.java:294) at org.apache.kylin.source.hive.HiveMRInput$RedistributeFlatHiveTableStep.doWork(HiveMRInput.java:337) at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:113) at org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:57) at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:113) at org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:136) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) Caused by: org.apache.hadoop.ipc.RemoteException(java.io.FileNotFoundException): File does not exist: /young/kylin_test/kylin_metadata_test/kylin-678266c0-ba0e-48b4-bdb5-6e578320375a/row_count/000000_0 at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:71) at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:61) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsInt(FSNamesystem.java:1820) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1791) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1704) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:587) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:365) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2081) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2077) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2075) at org.apache.hadoop.ipc.Client.call(Client.java:1427) at org.apache.hadoop.ipc.Client.call(Client.java:1358) at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:229) at com.sun.proxy.$Proxy25.getBlockLocations(Unknown Source) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getBlockLocations(ClientNamenodeProtocolTranslatorPB.java:255) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:252) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:104) at com.sun.proxy.$Proxy26.getBlockLocations(Unknown Source) at org.apache.hadoop.hdfs.DFSClient.callGetBlockLocations(DFSClient.java:1240) ... 20 more 2016-12-19 10:28:35,078 DEBUG [pool-8-thread-7] dao.ExecutableDao:210 : updating job output, id: 678266c0-ba0e-48b4-bdb5-6e578320375a-01 2016-12-19 10:28:35,082 DEBUG [pool-8-thread-7] hbase.HBaseResourceStore:262 : Update row /execute_output/678266c0-ba0e-48b4-bdb5-6e578320375a-01 from oldTs: 1482114480638, to newTs: 1482114515078, operation result: true 2016-12-19 10:28:35,085 DEBUG [pool-8-thread-7] dao.ExecutableDao:210 : updating job output, id: 678266c0-ba0e-48b4-bdb5-6e578320375a-01