Re: hadoop file system error

2008-06-18 Thread 晋光峰
i'm sure i close all the files in the reduce step. Any other reasons cause this problem? 2008/6/18 Konstantin Shvachko <[EMAIL PROTECTED]>: > Did you close those files? > If not they may be empty. > > > > ??? wrote: > >> Dears, >> >> I use hadoop-0.16.4 to do some work and found a error which i c

Re: hadoop file system error

2008-06-18 Thread Konstantin Shvachko
Did you close those files? If not they may be empty. ??? wrote: Dears, I use hadoop-0.16.4 to do some work and found a error which i can't get the reasons. The scenario is like this: In the reduce step, instead of using OutputCollector to write result, i use FSDataOutputStream to write result

Re: dfs put fails

2008-06-18 Thread Alexander Arimond
Got a similar error when doing a mapreduce job on the master machine. Mapping job is ok and in the end there are the right results in my output folder, but the reduce hangs at 17% a very long time. Found this in one of the task logs a view times: ... 2008-06-18 17:31:02,297 INFO org.apache.hadoop

Re: Internet-Based Secure Clustered FS?

2008-06-18 Thread Chris Collins
Have you considered Amazon S3? I dont know how secure your requirements are. There are lots of companies using this for just offsite data storage and also with EC2. C On Jun 17, 2008, at 6:48 PM, Kenneth Miller wrote: All, I'm looking for a solution that would allow me to securely us

how can i save the JobClient info?

2008-06-18 Thread Daniel
Hi all, I'm new to Hadoop framework, i want to know when one MapReduce task is finished, is there any easy way to save the total number of input/output records to some file or variables? Thanks.

Re: All datanodes getting marked as dead

2008-06-18 Thread Dhruba Borthakur
You are running out of file handles on the namenode. When this happens, the namenode cannot receive heartbeats from datanodes because these heartbeats arrive on a tcp/ip socket connection and the namenode does not have any free file descriptors to accept these socket connections. Your data is stil

Re: data locality in HDFS

2008-06-18 Thread Dhruba Borthakur
HDFS uses the network topology to distribute and replicate data. An admin has to configure a script that describes the network topology to HDFS. This is specified by using the parameter "topology.script.file.name" in the Configuration file. This has been tested when nodes are on different subnets i

hadoop file system error

2008-06-18 Thread 晋光峰
Dears, I use hadoop-0.16.4 to do some work and found a error which i can't get the reasons. The scenario is like this: In the reduce step, instead of using OutputCollector to write result, i use FSDataOutputStream to write result to files on HDFS(becouse i want to split the result by some rules).

Re: is there a way to to debug hadoop from Eclipse

2008-06-18 Thread Brian Vargas
JMock also works rather well, using its cglib extensions, for mocking out fake FileSystem implementations, if you're expecting your code to make calls directly to the filesystem for some reason. Brian Matt Kent wrote: JMock is a unit testing tool for creating mock objects. I use it to mock th

Re: Task failing, cause FileSystem close?

2008-06-18 Thread Christophe Taton
Right. But actually as far as I have seen, we don't call FileSystem.close(). Here is a corresponding stacktrace: it seems the exception happens when the task child vm closes the sequence file (which is out of control of the Mapper.map() function). java.io.IOException: Filesystem closed at

Re: dfs put fails

2008-06-18 Thread Alexander Arimond
Thank you, first tried the put from the master machine, which leads to the error. The put from the slave machine works. Guess youre right with the configuration parameters. Appears a bit strange to me, because the firewall settings and the hadoop-site.xml on both machines are equal. On Tue, 2008-

dfs copyFromLocal/put fails

2008-06-18 Thread Alexander Arimond
hi, i'm new in hadoop and im just testing it at the moment. i set up a cluster with 2 nodes and it seems like they are running normally, the log files of the namenode and the datanodes dont show errors. Firewall should be set right. but when i try to upload a file to the dfs i get following m