BTW: KYLIN-2165 will handle this in a more elegant way, it will be released
in next version;

2016-12-22 9:21 GMT+08:00 ShaoFeng Shi <shaofeng...@apache.org>:

> "java.io.FileNotFoundException: File does not exist:
> /young/kylin_test/kylin_metadata_test/kylin-678266c0-ba0e-
> 48b4-bdb5-6e578320375a/row_count/000000_0"
>
> Seems be the issue in https://issues.apache.org/jira/browse/KYLIN-2159
> It should be fixed in 1.6.0; could you please check whether there is any
> file in the folder?
>
> hadoop fs -ls /young/kylin_test/kylin_metadata_test/kylin-678266c0-
> ba0e-48b4-bdb5-6e578320375a/row_count/
>
> 2016-12-20 15:23 GMT+08:00 Alberto Ramón <a.ramonporto...@gmail.com>:
>
>> other idea:
>> Can be a problem with permissions?: the user that execute Kylin can't
>> read data generated by YARN
>> check if Kylin user can read your folder  /young/kylin_test/
>> Which Hadoop user are executing Kylin?
>>
>> (no more ideas, Good Luck)
>>
>> 2016-12-20 7:51 GMT+01:00 雨日听风 <491245...@qq.com>:
>>
>>> Thank you!
>>> We checked the yarn and hard disk. But not found any error. Hard disk
>>> space and memory and so on is working well.
>>> Last time its error code was "unknownhost clusterB",now in new server
>>> env it cant find clusterB(hbase only). but cant find rowCount file.
>>> ===
>>> the follow command runs ok:
>>> hdfs dfs -mkdir /young/kylin_test/kylin_metadata_nokia/kylin
>>> -678c15ba-5375-4f80-831e-1ae0af8ed576/row_count/tmp
>>> And "ls" cant find file "000000_0"  which it said "file does not exist".
>>>
>>> ------------------ 原始邮件 ------------------
>>> *发件人:* "Alberto Ramón";<a.ramonporto...@gmail.com>;
>>> *发送时间:* 2016年12月19日(星期一) 晚上9:13
>>> *收件人:* "user"<user@kylin.apache.org>;
>>> *主题:* Re: Error when #2 Step: Redistribute Flat Hive Table - File does
>>> not exist
>>>
>>> i think i had this error last nigth  :)
>>> (go to yarn to find detailed error & find on internet)
>>> in my case was free space less than 10% of hard disk. Check this please
>>>
>>> El 19/12/2016 11:35, "雨日听风" <491245...@qq.com> escribió:
>>>
>>>> When I build a cube in kylin1.6, I get error in step #2: Redistribute
>>>> Flat Hive Table
>>>>
>>>> Please help! Thank you very much!
>>>>
>>>> env: kylin1.6 is in a independent server, and have 2 other server
>>>> cluster: clusterA(hive only) and clusterB(hbase only).
>>>> Error is:
>>>>
>>>> 2016-12-19 10:28:00,641 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 : Compute row count of flat hive
>>>> table, cmd:
>>>> 2016-12-19 10:28:00,642 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 : hive -e "USE boco;
>>>> SET dfs.replication=2;
>>>> SET hive.exec.compress.output=true;
>>>> SET hive.auto.convert.join.noconditionaltask=true;
>>>> SET hive.auto.convert.join.noconditionaltask.size=100000000;
>>>> SET mapreduce.output.fileoutputformat.compress.type=BLOCK;
>>>> SET mapreduce.job.split.metainfo.maxsize=-1;
>>>> SET mapreduce.job.queuename=young;
>>>> SET tez.queue.name=young;
>>>>
>>>> set hive.exec.compress.output=false;
>>>>
>>>> set hive.exec.compress.output=false;
>>>> INSERT OVERWRITE DIRECTORY '/young/kylin_test/kylin_metad
>>>> ata_test/kylin-678266c0-ba0e-48b4-bdb5-6e578320375a/row_count' SELECT
>>>> count(*) FROM kylin_intermediate_hbase_in_te
>>>> stCluster_CUBE_f9468805_eabf_4b54_bf2b_182e4c86214a;
>>>>
>>>> "
>>>> 2016-12-19 10:28:03,277 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 : WARNING: Use "yarn jar" to launch
>>>> YARN applications.
>>>> 2016-12-19 10:28:04,444 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 :
>>>> 2016-12-19 10:28:04,445 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 : Logging initialized using
>>>> configuration in file:/etc/hive/conf/hive-log4j.properties
>>>> 2016-12-19 10:28:14,700 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 : OK
>>>> 2016-12-19 10:28:14,703 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 : Time taken: 0.935 seconds
>>>> 2016-12-19 10:28:15,559 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 : Query ID =
>>>> young_20161219102814_a7104fd4-ba83-47fc-ac0b-0c9bef4e1969
>>>> 2016-12-19 10:28:15,560 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 : Total jobs = 1
>>>> 2016-12-19 10:28:15,575 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 : Launching Job 1 out of 1
>>>> 2016-12-19 10:28:22,842 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 :
>>>> 2016-12-19 10:28:22,842 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 :
>>>> 2016-12-19 10:28:23,104 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 : Status: Running (Executing on YARN
>>>> cluster with App id application_1473415773736_1063281)
>>>> 2016-12-19 10:28:23,104 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 :
>>>> 2016-12-19 10:28:23,104 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 : Map 1: -/- Reducer 2: 0/1
>>>> 2016-12-19 10:28:23,307 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 : Map 1: 0/2 Reducer 2: 0/1
>>>> 2016-12-19 10:28:26,363 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 : Map 1: 0/2 Reducer 2: 0/1
>>>> 2016-12-19 10:28:26,567 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 : Map 1: 0(+1)/2 Reducer 2: 0/1
>>>> 2016-12-19 10:28:26,596 INFO  [pool-7-thread-1]
>>>> threadpool.DefaultScheduler:118 : Job Fetcher: 1 should running, 1
>>>> actual running, 0 ready, 0 already succeed, 3 error, 1 discarded, 0 others
>>>> 2016-12-19 10:28:26,769 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 : Map 1: 0(+2)/2 Reducer 2: 0/1
>>>> 2016-12-19 10:28:29,810 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 : Map 1: 0(+2)/2 Reducer 2: 0/1
>>>> 2016-12-19 10:28:30,217 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 : Map 1: 1(+1)/2 Reducer 2: 0(+1)/1
>>>> 2016-12-19 10:28:30,826 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 : Map 1: 2/2 Reducer 2: 0(+1)/1
>>>> 2016-12-19 10:28:31,232 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 : Map 1: 2/2 Reducer 2: 1/1
>>>> 2016-12-19 10:28:31,319 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 : Moving data to:
>>>> /young/kylin_test/kylin_metadata_test/kylin-678266c0-ba0e-48
>>>> b4-bdb5-6e578320375a/row_count
>>>> 2016-12-19 10:28:31,406 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 : OK
>>>> 2016-12-19 10:28:31,454 INFO  [pool-8-thread-7]
>>>> execution.AbstractExecutable:36 : Time taken: 16.701 seconds
>>>> 2016-12-19 10:28:35,074 ERROR [pool-8-thread-7]
>>>> execution.AbstractExecutable:357 : 
>>>> job:678266c0-ba0e-48b4-bdb5-6e578320375a-01
>>>> execute finished with exception
>>>> java.io.FileNotFoundException: File does not exist:
>>>> /young/kylin_test/kylin_metadata_test/kylin-678266c0-ba0e-48
>>>> b4-bdb5-6e578320375a/row_count/000000_0
>>>>  at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INo
>>>> deFile.java:71)
>>>>  at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INo
>>>> deFile.java:61)
>>>>  at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlock
>>>> LocationsInt(FSNamesystem.java:1820)
>>>>  at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlock
>>>> Locations(FSNamesystem.java:1791)
>>>>  at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlock
>>>> Locations(FSNamesystem.java:1704)
>>>>  at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.get
>>>> BlockLocations(NameNodeRpcServer.java:587)
>>>>  at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServ
>>>> erSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolS
>>>> erverSideTranslatorPB.java:365)
>>>>  at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocol
>>>> Protos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNam
>>>> enodeProtocolProtos.java)
>>>>  at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcIn
>>>> voker.call(ProtobufRpcEngine.java:616)
>>>>  at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969)
>>>>  at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2081)
>>>>  at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2077)
>>>>  at java.security.AccessController.doPrivileged(Native Method)
>>>>  at javax.security.auth.Subject.doAs(Subject.java:415)
>>>>  at org.apache.hadoop.security.UserGroupInformation.doAs(UserGro
>>>> upInformation.java:1657)
>>>>  at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2075)
>>>>
>>>>  at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
>>>> Method)
>>>>  at sun.reflect.NativeConstructorAccessorImpl.newInstance(Native
>>>> ConstructorAccessorImpl.java:57)
>>>>  at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(De
>>>> legatingConstructorAccessorImpl.java:45)
>>>>  at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
>>>>  at org.apache.hadoop.ipc.RemoteException.instantiateException(R
>>>> emoteException.java:106)
>>>>  at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(
>>>> RemoteException.java:73)
>>>>  at org.apache.hadoop.hdfs.DFSClient.callGetBlockLocations(DFSCl
>>>> ient.java:1242)
>>>>  at org.apache.hadoop.hdfs.DFSClient.getLocatedBlocks(DFSClient.
>>>> java:1227)
>>>>  at org.apache.hadoop.hdfs.DFSClient.getLocatedBlocks(DFSClient.
>>>> java:1215)
>>>>  at org.apache.hadoop.hdfs.DFSInputStream.fetchLocatedBlocksAndG
>>>> etLastBlockLength(DFSInputStream.java:303)
>>>>  at org.apache.hadoop.hdfs.DFSInputStream.openInfo(DFSInputStrea
>>>> m.java:269)
>>>>  at org.apache.hadoop.hdfs.DFSInputStream.<init>(DFSInputStream.
>>>> java:261)
>>>>  at org.apache.hadoop.hdfs.DFSClient.open(DFSClient.java:1540)
>>>>  at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(Distri
>>>> butedFileSystem.java:303)
>>>>  at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(Distri
>>>> butedFileSystem.java:299)
>>>>  at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSyst
>>>> emLinkResolver.java:81)
>>>>  at org.apache.hadoop.hdfs.DistributedFileSystem.open(Distribute
>>>> dFileSystem.java:299)
>>>>  at org.apache.hadoop.fs.FileSystem.open(FileSystem.java:767)
>>>>  at org.apache.kylin.source.hive.HiveMRInput$RedistributeFlatHiv
>>>> eTableStep.readRowCountFromFile(HiveMRInput.java:294)
>>>>  at org.apache.kylin.source.hive.HiveMRInput$RedistributeFlatHiv
>>>> eTableStep.doWork(HiveMRInput.java:337)
>>>>  at org.apache.kylin.job.execution.AbstractExecutable.execute(Ab
>>>> stractExecutable.java:113)
>>>>  at org.apache.kylin.job.execution.DefaultChainedExecutable.doWo
>>>> rk(DefaultChainedExecutable.java:57)
>>>>  at org.apache.kylin.job.execution.AbstractExecutable.execute(Ab
>>>> stractExecutable.java:113)
>>>>  at org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRun
>>>> ner.run(DefaultScheduler.java:136)
>>>>  at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool
>>>> Executor.java:1145)
>>>>  at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo
>>>> lExecutor.java:615)
>>>>  at java.lang.Thread.run(Thread.java:745)
>>>> Caused by: 
>>>> org.apache.hadoop.ipc.RemoteException(java.io.FileNotFoundException):
>>>> File does not exist: /young/kylin_test/kylin_metada
>>>> ta_test/kylin-678266c0-ba0e-48b4-bdb5-6e578320375a/row_count/000000_0
>>>>  at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INo
>>>> deFile.java:71)
>>>>  at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INo
>>>> deFile.java:61)
>>>>  at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlock
>>>> LocationsInt(FSNamesystem.java:1820)
>>>>  at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlock
>>>> Locations(FSNamesystem.java:1791)
>>>>  at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlock
>>>> Locations(FSNamesystem.java:1704)
>>>>  at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.get
>>>> BlockLocations(NameNodeRpcServer.java:587)
>>>>  at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServ
>>>> erSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolS
>>>> erverSideTranslatorPB.java:365)
>>>>  at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocol
>>>> Protos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNam
>>>> enodeProtocolProtos.java)
>>>>  at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcIn
>>>> voker.call(ProtobufRpcEngine.java:616)
>>>>  at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969)
>>>>  at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2081)
>>>>  at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2077)
>>>>  at java.security.AccessController.doPrivileged(Native Method)
>>>>  at javax.security.auth.Subject.doAs(Subject.java:415)
>>>>  at org.apache.hadoop.security.UserGroupInformation.doAs(UserGro
>>>> upInformation.java:1657)
>>>>  at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2075)
>>>>
>>>>  at org.apache.hadoop.ipc.Client.call(Client.java:1427)
>>>>  at org.apache.hadoop.ipc.Client.call(Client.java:1358)
>>>>  at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(Proto
>>>> bufRpcEngine.java:229)
>>>>  at com.sun.proxy.$Proxy25.getBlockLocations(Unknown Source)
>>>>  at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTran
>>>> slatorPB.getBlockLocations(ClientNamenodeProtocolTranslatorPB.java:255)
>>>>  at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>  at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAcce
>>>> ssorImpl.java:57)
>>>>  at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMe
>>>> thodAccessorImpl.java:43)
>>>>  at java.lang.reflect.Method.invoke(Method.java:606)
>>>>  at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMeth
>>>> od(RetryInvocationHandler.java:252)
>>>>  at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(Ret
>>>> ryInvocationHandler.java:104)
>>>>  at com.sun.proxy.$Proxy26.getBlockLocations(Unknown Source)
>>>>  at org.apache.hadoop.hdfs.DFSClient.callGetBlockLocations(DFSCl
>>>> ient.java:1240)
>>>>  ... 20 more
>>>> 2016-12-19 10:28:35,078 DEBUG [pool-8-thread-7] dao.ExecutableDao:210 :
>>>> updating job output, id: 678266c0-ba0e-48b4-bdb5-6e578320375a-01
>>>> 2016-12-19 10:28:35,082 DEBUG [pool-8-thread-7]
>>>> hbase.HBaseResourceStore:262 : Update row 
>>>> /execute_output/678266c0-ba0e-48b4-bdb5-6e578320375a-01
>>>> from oldTs: 1482114480638, to newTs: 1482114515078, operation result: true
>>>> 2016-12-19 10:28:35,085 DEBUG [pool-8-thread-7] dao.ExecutableDao:210 :
>>>> updating job output, id: 678266c0-ba0e-48b4-bdb5-6e578320375a-01
>>>>
>>>
>>
>
>
> --
> Best regards,
>
> Shaofeng Shi 史少锋
>
>


-- 
Best regards,

Shaofeng Shi 史少锋

Reply via email to