Home | About | Sematext search-lucene.com search-hadoop.com
 Search Hadoop and all its subprojects:

Switch to Threaded View
MapReduce, mail # user - Re: LeaseExpiredException : Lease mismatch in Hadoop mapReduce| How to solve?


Copy link to this message
-
Re: LeaseExpiredException : Lease mismatch in Hadoop mapReduce| How to solve?
unmesha sreeveni 2013-11-15, 06:18
@chandu banavaram:
This exception usually happens if hdfs is trying to write into a file which
is no more in hdfs..

I think in my case certain files are not created in my hdfs.it failed to
create due to some permissions.

I am trying out.
On Wed, Nov 13, 2013 at 9:25 AM, unmesha sreeveni <[EMAIL PROTECTED]>wrote:

> :) Ok
> Why u also experienced the same?
>
>
> On Tue, Nov 12, 2013 at 5:14 PM, chandu banavaram <
> [EMAIL PROTECTED]> wrote:
>
>> plz send the answer to me  for this query
>>
>>
>> On Tue, Nov 12, 2013 at 2:52 AM, unmesha sreeveni <[EMAIL PROTECTED]>wrote:
>>
>>> While running job with 90 Mb file i am getting LeaseExpiredException
>>>
>>> 13/11/12 15:46:41 WARN mapred.JobClient: Use GenericOptionsParser for
>>> parsing the arguments. Applications should implement Tool for the same.
>>> 13/11/12 15:46:42 INFO input.FileInputFormat: Total input paths to
>>> process : 1
>>> 13/11/12 15:46:43 INFO mapred.JobClient: Running job:
>>> job_201310301645_25033
>>>  13/11/12 15:46:44 INFO mapred.JobClient:  map 0% reduce 0%
>>> 13/11/12 15:46:56 INFO mapred.JobClient: Task Id :
>>> attempt_201310301645_25033_m_000000_0, Status : FAILED
>>> org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException):
>>> Lease mismatch on /user/hdfs/in/map owned by
>>> DFSClient_NONMAPREDUCE_-1622335545_1 but is accessed by
>>> DFSClient_NONMAPREDUCE_-1561990512_1
>>>  at
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:2459)
>>> at
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:2437)
>>>  at
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.completeFileInternal(FSNamesystem.java:2503)
>>> at
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.completeFile(FSNamesystem.java:2480)
>>>  at
>>> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.complete(NameNodeRpcServer.java:556)
>>> at
>>> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.complete(ClientNamenodeProtocolServerSideTranslatorPB.java:337)
>>>  at
>>> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:44958)
>>>  at org.
>>> attempt_201310301645_25033_m_000000_0: SLF4J: Class path contains
>>> multiple SLF4J bindings.
>>> attempt_201310301645_25033_m_000000_0: SLF4J: Found binding in
>>> [jar:file:/usr/lib/zookeeper/lib/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>> attempt_201310301645_25033_m_000000_0: SLF4J: Found binding in
>>> [jar:file:/tmp/hadoop-mapred/mapred/local/taskTracker/hdfs/jobcache/job_201310301645_25033/jars/job.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>> attempt_201310301645_25033_m_000000_0: SLF4J: See
>>> http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
>>>  13/11/12 15:47:02 INFO mapred.JobClient: Task Id :
>>> attempt_201310301645_25033_m_000000_1, Status : FAILED
>>> org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException):
>>> Lease mismatch on /user/hdfs/in/map owned by
>>> DFSClient_NONMAPREDUCE_-1622335545_1 but is accessed by
>>> DFSClient_NONMAPREDUCE_-1662926329_1
>>>  at
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:2459)
>>> at
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.analyzeFileState(FSNamesystem.java:2262)
>>>  at
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2175)
>>> at
>>> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:501)
>>>  at
>>> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:299)
>>>  at
>>> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:44954)
>>>  at
*Thanks & Regards*

Unmesha Sreeveni U.B

*Junior Developer*