Hi Shashwat,

As already mentioned already in my mail setting 
dfs.client.use.legacy.blockreader to true fixes the problem.
This looks to be workaround or moreover disabling a feature.
Would like to know, what is the exact problem?

Cheers,
Subroto Sanyal
On Mar 5, 2013, at 6:33 PM, shashwat shriparv wrote:

> Try setting dfs.client.use.legacy.blockreader to true
> 
> ∞
> Shashwat Shriparv
> 
> 
> 
> On Tue, Mar 5, 2013 at 8:39 PM, 卖报的小行家 <85469...@qq.com> wrote:
> Yes.It's from hadoop 2.0. I just now read the code 1.1.1.There are no such 
> classes the log mentioned.Maybe you can read the code first.
> 
> 
> ------------------ 原始邮件 ------------------
> 发件人: "Subroto"<ssan...@datameer.com>;
> 发送时间: 2013年3月5日(星期二) 晚上10:56
> 收件人: "user"<user@hadoop.apache.org>;
> 主题: Re: Socket does not have a channel
> 
> Hi Julian,
> 
> This is from CDH4.1.2 and I think its based on Apache Hadoop 2.0.
> 
> Cheers,
> Subroto Sanyal
> On Mar 5, 2013, at 3:50 PM, 卖报的小行家 wrote:
> 
>> Hi,
>> Which revision of hadoop?
>> and  what's the  situation  to report the Exception?
>> BRs//Julian
>> 
>> ------------------ Original ------------------
>> From:  "Subroto"<ssan...@datameer.com>;
>> Date:  Tue, Mar 5, 2013 04:46 PM
>> To:  "user"<user@hadoop.apache.org>;
>> Subject:  Socket does not have a channel
>> 
>> Hi
>> 
>> java.lang.IllegalStateException: Socket 
>> Socket[addr=/10.86.203.112,port=1004,localport=35170] does not have a channel
>>      at 
>> com.google.common.base.Preconditions.checkState(Preconditions.java:172)
>>      at 
>> org.apache.hadoop.net.SocketInputWrapper.getReadableByteChannel(SocketInputWrapper.java:83)
>>      at 
>> org.apache.hadoop.hdfs.RemoteBlockReader2.newBlockReader(RemoteBlockReader2.java:432)
>>      at 
>> org.apache.hadoop.hdfs.BlockReaderFactory.newBlockReader(BlockReaderFactory.java:82)
>>      at 
>> org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:832)
>>      at 
>> org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:444)
>> 
>> While accessing the HDFS  I keep getting the above mentioned error.
>> Setting the dfs.client.use.legacy.blockreader to true fixes the problem.
>> I would like to know what exactly is the problem? Is it a problem/bug in 
>> hadoop ?
>> Is there is JIRA ticket for this??
>> 
>> 
>> Cheers,
>> Subroto Sanyal
> 
> 

Reply via email to