Can you trace namenode logs, whether this is file deleted/renamed(might be 
parent folder) before this reducer run..?




--Brahma Reddy Battula

From: Zhang Jianfeng [mailto:jzhang...@gmail.com]
Sent: 18 October 2016 18:55
To: Gaurav Kumar
Cc: user.hadoop; Rakesh Radhakrishnan
Subject: Re: LeaseExpiredException: No lease on 
/user/biadmin/analytic‐root/SX5XPWPPDPQH/.

Thanks Gaurav. For my case, I called the HDFS API to write the reducer result 
into HDFS directly, not using Spark.

2016-10-17 23:24 GMT+08:00 Gaurav Kumar 
<gauravkuma...@gmail.com<mailto:gauravkuma...@gmail.com>>:

Hi,

Please also check for coalesced RDD. I encountered the same error while writing 
a coalesced rdd/df to HDFS. If this is the case, please use repartition instead.

Sent from OnePlus 3

Thanks & Regards,
Gaurav Kumar

On Oct 17, 2016 11:22 AM, "Zhang Jianfeng" 
<jzhang...@gmail.com<mailto:jzhang...@gmail.com>> wrote:
Thanks Rakesh for your kind help. Actually during the job only one reducer 
result file (for example part-r-2) had this error, other reducers worked well.

Best Regards,
Jian Feng

2016-10-17 11:49 GMT+08:00 Rakesh Radhakrishnan 
<rake...@apache.org<mailto:rake...@apache.org>>:
Hi Jian Feng,

Could you please check your code and see any possibilities of simultaneous 
access to the same file. Mostly this situation happens when multiple clients 
tries to access the same file.

Code Reference:- 
https://github.com/apache/hadoop/blob/branch-2.2/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java#L2737

Best Regards,
Rakesh
Intel

On Mon, Oct 17, 2016 at 7:16 AM, Zhang Jianfeng 
<jzhang...@gmail.com<mailto:jzhang...@gmail.com>> wrote:
Hi ,

    I hit an wired error. On our hadoop cluster (2.2.0), occasionally a 
LeaseExpiredException is thrown.

The stacktrace is as below:


org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException):
 No lease on /user/biadmin/analytic‐root/SX5XPWPPDPQH/.executions/.at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:2737)

at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.completeFileInternal(FSNamesystem.java:2801)

at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.completeFile(FSNamesystem.java:2783)

at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.com<http://namenode.NameNodeRpcServer.com>plete(NameNodeRpcServer.java:611)

at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.complete(ClientNamenodeProtocolServerSideTranslatorPB.java:428)

at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:59586)

at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)

at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928)

at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2048)

at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2044)

at java.security.AccessController.doPrivileged(AccessController.java:310)

at 
javax.security.auth.Subject.do<http://javax.security.auth.Subject.do>As(Subject.java:573)

at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1502)

at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2042)

at org.apache.hadoop.ipc.Client.call(Client.java:1347)

at org.apache.hadoop.ipc.Client.call(Client.java:1300)

at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)

at $Proxy7.complete(Unknown Source)

at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:60)

at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:37)

at java.lang.reflect.Method.invoke(Method.java:611)

at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:186)

at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)

at $Proxy7.complete(Unknown Source)

at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.complete(ClientNamenodeProtocolTranslatorPB.java:371)

at 
org.apache.hadoop.hdfs.DFSOutputStream.completeFile(DFSOutputStream.java:1894)

at org.apache.hadoop.hdfs.DFSOutputStream.close(DFSOutputStream.java:1881)

at 
org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStream.java:71)

at org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:104)

at java.io.FilterOutputStream.close(FilterOutputStream.java:154)
Any help will be appreciated!

--
Best Regards,
Jian Feng




--
Best Regards,
Jian Feng




--
Best Regards,
Jian Feng

Reply via email to