RE: problem about put a lot of files

2009-06-21 Thread zhuweimin
Hi

The max open files have limit in LINUX box. Please using ulimit to view and
modify the limit
1.view limit
   # ulimit -a
2.modify limit
   For example
   # ulimit -n 10240 

Best wish

> -Original Message-
> From: stchu [mailto:stchu.cl...@gmail.com]
> Sent: Monday, June 22, 2009 12:57 PM
> To: core-user@hadoop.apache.org
> Subject: problem about put a lot of files
> 
> Hi,
> Is there any restriction on the amount of putting files? I tried to
> put/copyFromLocal about 50,573 files to HDFS, but I faced a problem:
> ==
> ==
> 09/06/22 11:34:34 INFO dfs.DFSClient: Exception in
> createBlockOutputStream
> java.io.IOException: Bad connect ack with firstBadLink 140.96.89.57:51010
> 09/06/22 11:34:34 INFO dfs.DFSClient: Abandoning block
> blk_8245450203753506945_65955
> 09/06/22 11:34:40 INFO dfs.DFSClient: Exception in
> createBlockOutputStream
> java.io.IOException: Bad connect ack with firstBadLink 140.96.89.57:51010
> 09/06/22 11:34:40 INFO dfs.DFSClient: Abandoning block
> blk_-8257846965500649510_65956
> 09/06/22 11:34:46 INFO dfs.DFSClient: Exception in
> createBlockOutputStream
> java.io.IOException: Bad connect ack with firstBadLink 140.96.89.57:51010
> 09/06/22 11:34:46 INFO dfs.DFSClient: Abandoning block
> blk_4751737303082929912_65956
> 09/06/22 11:34:56 INFO dfs.DFSClient: Exception in
> createBlockOutputStream
> java.io.IOException: Bad connect ack with firstBadLink 140.96.89.57:51010
> 09/06/22 11:34:56 INFO dfs.DFSClient: Abandoning block
> blk_5912850890372596972_66040
> 09/06/22 11:35:02 INFO dfs.DFSClient: Exception in
> createBlockOutputStream
> java.io.IOException: Bad connect ack with firstBadLink
> 140.96.89.193:51010
> 09/06/22 11:35:02 INFO dfs.DFSClient: Abandoning block
> blk_6609198685444611538_66040
> 09/06/22 11:35:08 INFO dfs.DFSClient: Exception in
> createBlockOutputStream
> java.io.IOException: Bad connect ack with firstBadLink
> 140.96.89.193:51010
> 09/06/22 11:35:08 INFO dfs.DFSClient: Abandoning block
> blk_6696101244177965180_66040
> 09/06/22 11:35:17 INFO dfs.DFSClient: Exception in
> createBlockOutputStream
> java.io.IOException: Bad connect ack with firstBadLink 140.96.89.57:51010
> 09/06/22 11:35:17 INFO dfs.DFSClient: Abandoning block
> blk_-5430033105510098342_66105
> 09/06/22 11:35:26 INFO dfs.DFSClient: Exception in
> createBlockOutputStream
> java.io.IOException: Bad connect ack with firstBadLink 140.96.89.57:51010
> 09/06/22 11:35:26 INFO dfs.DFSClient: Abandoning block
> blk_5325140471333041601_66165
> 09/06/22 11:35:32 INFO dfs.DFSClient: Exception in
> createBlockOutputStream
> java.io.IOException: Bad connect ack with firstBadLink
> 140.96.89.205:51010
> 09/06/22 11:35:32 INFO dfs.DFSClient: Abandoning block
> blk_1121864992752821949_66165
> 09/06/22 11:35:39 INFO dfs.DFSClient: Exception in
> createBlockOutputStream
> java.io.IOException: Bad connect ack with firstBadLink
> 140.96.89.205:51010
> 09/06/22 11:35:39 INFO dfs.DFSClient: Abandoning block
> blk_-2096783021040778965_66184
> 09/06/22 11:35:45 INFO dfs.DFSClient: Exception in
> createBlockOutputStream
> java.io.IOException: Bad connect ack with firstBadLink
> 140.96.89.205:51010
> 09/06/22 11:35:45 INFO dfs.DFSClient: Abandoning block
> blk_6949821898790162970_66184
> 09/06/22 11:35:51 INFO dfs.DFSClient: Exception in
> createBlockOutputStream
> java.io.IOException: Bad connect ack with firstBadLink
> 140.96.89.205:51010
> 09/06/22 11:35:51 INFO dfs.DFSClient: Abandoning block
> blk_4708848202696905125_66184
> 09/06/22 11:35:57 INFO dfs.DFSClient: Exception in
> createBlockOutputStream
> java.io.IOException: Bad connect ack with firstBadLink
> 140.96.89.205:51010
> 09/06/22 11:35:57 INFO dfs.DFSClient: Abandoning block
> blk_8031882012801762201_66184
> 09/06/22 11:36:03 WARN dfs.DFSClient: DataStreamer Exception:
> java.io.IOException: Unable to create new block.
> at
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.nextBlockOutputStream(
> DFSClient.java:2359)
> at
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.access$1800(DFSClient.
> java:1745)
> at
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSCl
> ient.java:1922)
> 
> 09/06/22 11:36:03 WARN dfs.DFSClient: Error Recovery for block
> blk_8031882012801762201_66184 bad datanode[2]
> put: Could not get block locations. Aborting...
> Exception closing file /osmFiles/a/109103.gpx.txt
> java.io.IOException: Could not get block locations. Aborting...
> at
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.processDatanodeError(D
> FSClient.java:2153)
> at
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.access$1400(DFSClient.
> java:1745)
> at
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSCl
> ient.java:1899)
> 
> ==
> =
>

can't edit the file that mounted by fuse_dfs by editor

2009-02-11 Thread zhuweimin
Hey all

I was trying to edit the file that mounted by fuse_dfs by vi editor, but the
contents could not save.
The command is like the following:
[had...@vm-centos-5-shu-4 src]$ vi /mnt/dfs/test.txt
The error message from system log (/var/log/messages) is the following:
Feb 12 09:53:48 VM-CentOS-5-SHU-4 fuse_dfs: ERROR: could not connect open
file fuse_dfs.c:1340

I using the hadoop0.19.0 and fuse-dfs version 26 with centos5.2.
Does anyone have an idea as to what could be wrong!

Thanks!
zhuweimin




RE: Reduce won't start until Map stage reaches 100%?

2009-02-09 Thread zhuweimin
Hi

I think the number of your job's reduce task is 1
because if the number of reduce task is 1 then reduce stage does not start
until Map stage 100% completion.

zhuweimin

-Original Message-
From: Taeho Kang [mailto:tka...@gmail.com] 
Sent: Monday, February 09, 2009 4:26 PM
To: hadoop-u...@lucene.apache.org
Subject: Reduce won't start until Map stage reaches 100%?

Dear All,

With Hadoop 0.19.0, Reduce stage does not start until Map stage gets to the
100% completion.
Has anyone faced the similar situation?

 ... ...
 -  map 90% reduce 0%
-  map 91% reduce 0%
-  map 92% reduce 0%
-  map 93% reduce 0%
-  map 94% reduce 0%
-  map 95% reduce 0%
-  map 96% reduce 0%
-  map 97% reduce 0%
-  map 98% reduce 0%
-  map 99% reduce 0%
-  map 100% reduce 0%
-  map 100% reduce 1%
-  map 100% reduce 2%
-  map 100% reduce 3%
-  map 100% reduce 4%
-  map 100% reduce 5%
-  map 100% reduce 6%
-  map 100% reduce 7%
-  map 100% reduce 8%
-  map 100% reduce 9%

Thank you all in advance,

/Taeho




RE: The error occurred when a lot of files created use fuse-dfs

2008-12-16 Thread zhuweimin
Brian

Thank you very much.

The version of Hadoop is 0.19.0,I think 4616 and 4635 patches is necessary.

I will try it.



-Original Message-
From: Brian Bockelman [mailto:bbock...@cse.unl.edu] 
Sent: Monday, December 15, 2008 10:00 PM
To: core-user@hadoop.apache.org
Subject: Re: The error occurred when a lot of files created use fuse-dfs

Hey,

What version of Hadoop are you running?  Have you taken a look at  
HADOOP-4775?

https://issues.apache.org/jira/browse/HADOOP-4775

Basically, fuse-dfs is not usable on Hadoop 0.19.0 without a patch.

Brian

On Dec 15, 2008, at 12:24 AM, zhuweimin wrote:

> Dear fuse-dfs users
>
> I copy 1000 files into hadoop from local disk use fuse-dfs,
> Display the following error when the 600th files are copied:
>
> cp: cannot create regular file `/mnt/dfs/user/hadoop/fuse3/10m/ 
> 10m_33.dat':
> Input/output error
> cp: cannot create regular file `/mnt/dfs/user/hadoop/fuse3/10m/ 
> 10m_34.dat':
> Input/output error
> cp: cannot create regular file `/mnt/dfs/user/hadoop/fuse3/10m/ 
> 10m_35.dat':
> Input/output error
> cp: cannot create regular file `/mnt/dfs/user/hadoop/fuse3/10m/ 
> 10m_36.dat':
> Input/output error
> cp: cannot create regular file `/mnt/dfs/user/hadoop/fuse3/10m/ 
> 10m_37.dat':
> Input/output error
> cp: cannot create regular file `/mnt/dfs/user/hadoop/fuse3/10m/ 
> 10m_38.dat':
> Input/output error
> ...
>
> It is necessary to remount the fuse-dfs.
>
> Do you think about of the error.
>
> thanks
>
>




The error occurred when a lot of files created use fuse-dfs

2008-12-14 Thread zhuweimin
Dear fuse-dfs users

I copy 1000 files into hadoop from local disk use fuse-dfs, 
Display the following error when the 600th files are copied:

cp: cannot create regular file `/mnt/dfs/user/hadoop/fuse3/10m/10m_33.dat':
Input/output error
cp: cannot create regular file `/mnt/dfs/user/hadoop/fuse3/10m/10m_34.dat':
Input/output error
cp: cannot create regular file `/mnt/dfs/user/hadoop/fuse3/10m/10m_35.dat':
Input/output error
cp: cannot create regular file `/mnt/dfs/user/hadoop/fuse3/10m/10m_36.dat':
Input/output error
cp: cannot create regular file `/mnt/dfs/user/hadoop/fuse3/10m/10m_37.dat':
Input/output error
cp: cannot create regular file `/mnt/dfs/user/hadoop/fuse3/10m/10m_38.dat':
Input/output error
...

It is necessary to remount the fuse-dfs.

Do you think about of the error.

thanks





RE: Did Hadoop support gz/zip format file?

2008-11-30 Thread zhuweimin
Hello

I have a requirement of use file with the extension .Z(the file is created
by UNIX compress command) in Hadoop.
It will too be automatically recognized / handled ?

Any suggestion on how to handle the .Z file in the map task?

Thanks
Best Regards

Alamo

-Original Message-
From: Abdul Qadeer [mailto:[EMAIL PROTECTED] 
Sent: Monday, November 03, 2008 4:19 PM
To: core-user@hadoop.apache.org
Subject: Re: Did Hadoop support gz/zip format file?

Hadoop supports gzip format by means of its gzip codecs.
If you meant to use default input format / record reader of
Hadoop, it will automatically recognize / handle your gzip
input files.  And if you are using some other input formats / record
readers, you can use the gzip codec provided by Hadoop to
de-compress the data and then interpret it in your record reader.

On Sun, Nov 2, 2008 at 7:49 PM, howardh <[EMAIL PROTECTED]> wrote:

> Hello,
>
> I have a requirement of use gz/zip format in Hadoop,   After these days
> research and learning,   It seems Hadoop didn't support gz/zip file yet,
>  is that true?
> Like I'm going to create file with gz format,  and read it later by
> FileSystem interface. Does it feasible? Experts!  Could you show
me
> some advices?
>
> Best Regards
> 2008-11-03
>
>
>
> howardh
>