Re: Very strange Java Collection behavior in Hadoop

2012-03-19 Thread Owen O'Malley
On Mon, Mar 19, 2012 at 11:05 PM, madhu phatak  wrote:

> Hi Owen O'Malley,
>  Thank you for that Instant reply. It's working now. Can you explain me
> what you mean by "input to reducer is reused" in little detail?


Each time the statement "Text value = values.next();" is executed it always
returns the same Text object with the contents of that object changed. When
you add the Text to the list, you are adding a pointer to the same Text
object. At the end you have 6 copies of the same pointer instead of 6
different Text objects.

The reason that I said it is my fault, is because I added the optimization
that causes it. If you are interested in Hadoop archeology, it was
HADOOP-2399 that made the change. I also did HADOOP-3522 to improve the
documentation in the area.

-- Owen


Re: Very strange Java Collection behavior in Hadoop

2012-03-19 Thread madhu phatak
Hi Owen O'Malley,
 Thank you for that Instant reply. It's working now. Can you explain me
what you mean by "input to reducer is reused" in little detail?

On Tue, Mar 20, 2012 at 11:28 AM, Owen O'Malley  wrote:

> On Mon, Mar 19, 2012 at 10:52 PM, madhu phatak 
> wrote:
>
> > Hi All,
> >  I am using Hadoop 0.20.2 . I am observing a Strange behavior of Java
> > Collection's . I have following code in reducer
>
>
> That is my fault. *sigh* The input to the reducer is reused. Replace:
>
> list.add(value);
>
> with:
>
> list.add(new Text(value));
>
> and the problem will go away.
>
> -- Owen
>



-- 
https://github.com/zinnia-phatak-dev/Nectar


Re: Very strange Java Collection behavior in Hadoop

2012-03-19 Thread Owen O'Malley
On Mon, Mar 19, 2012 at 10:52 PM, madhu phatak  wrote:

> Hi All,
>  I am using Hadoop 0.20.2 . I am observing a Strange behavior of Java
> Collection's . I have following code in reducer


That is my fault. *sigh* The input to the reducer is reused. Replace:

list.add(value);

with:

list.add(new Text(value));

and the problem will go away.

-- Owen


Very strange Java Collection behavior in Hadoop

2012-03-19 Thread madhu phatak
Hi All,
 I am using Hadoop 0.20.2 . I am observing a Strange behavior of Java
Collection's . I have following code in reducer

   public void reduce(Text text, Iterator values,
OutputCollector collector, Reporter reporter)
throws IOException {
// TODO Auto-generated method stub
List list = new ArrayList();
while(values.hasNext())
{
Text value = values.next();
list.add(value);
System.out.println(value.toString());   }

for(Text value : list)
{
System.out.println(value.toString());
}

}

The  first sysout prints following

4   5   6

1   2   3

But when I print from the List, it prints following

1   2   3
1   2   3

*
*The all List values are getting replaced by last added value.

I am not able to understand this behavior. Did anyone seen this behavior ?

Regards,
Madhukara Phatak

-- 
https://github.com/zinnia-phatak-dev/Nectar


Re: Using mapred.reducer (old api) but don't write anything to HDFS using output collector

2012-03-19 Thread Harsh J
Hi,

If you do not want to output anything, use NullOutputFormat and simply
do not call output.collect(…) anywhere. The framework does not demand
that you call output.collect(…).

On Tue, Mar 20, 2012 at 3:03 AM, aji1705  wrote:
> I am using the old api and my Reducer class has the following signature (only
> relevant parts shown):
>
>
> import org.apache.hadoop.mapred.MapReduceBase;
> import org.apache.hadoop.mapred.OutputCollector;
> import org.apache.hadoop.mapred.Reducer;
> import org.apache.hadoop.mapred.Reporter;
>
> public class FlagDropComparisionReducer extends MapReduceBase
>               implements Reducer  {
>
>        public void reduce(Text key, Iterator values,
>                        OutputCollector output, Reporter reporter) 
> {
>
>        ..
>        String result = some code here to compute result using key and value
>        Use HTTP Client to POST "result" to server
>
>
>         }
> }
>
> I don't want to write the output of reducer to HDFS ie no
> "output.collect(key, somevalue). Is there a way for me to do this using the
> old api? Note: I must have a reducer because I am doing some calculation
> here (not shown).
>
> Suggestions? Thanks.
>
>
>
> --
> View this message in context: 
> http://hadoop-common.472056.n3.nabble.com/Using-mapred-reducer-old-api-but-don-t-write-anything-to-HDFS-using-output-collector-tp3840442p3840442.html
> Sent from the Users mailing list archive at Nabble.com.



-- 
Harsh J


question about processing large zip

2012-03-19 Thread Andrew McNair
Hi,

I have a large (~300 gig) zip of images that I need to process. My
current workflow is to copy the zip to HDFS, use a custom input format
to read the zip entries, do the processing in a map, and then generate
a processing report in the reduce. I'm struggling to tune params right
now with my cluster to make everything run smoothly, but I'm also
worried that I'm missing a better way of processing.

Does anybody have suggestions for how to make the processing of a zip
more parallel? The only other idea I had was uploading the zip as a
sequence file, but that proved incredibly slow (~30 hours on my 3 node
cluster to upload).

Thanks in advance.

-Andrew


Using mapred.reducer (old api) but don't write anything to HDFS using output collector

2012-03-19 Thread aji1705
I am using the old api and my Reducer class has the following signature (only
relevant parts shown):


import org.apache.hadoop.mapred.MapReduceBase;
import org.apache.hadoop.mapred.OutputCollector;
import org.apache.hadoop.mapred.Reducer;
import org.apache.hadoop.mapred.Reporter;

public class FlagDropComparisionReducer extends MapReduceBase 
   implements Reducer  {

public void reduce(Text key, Iterator values,
OutputCollector output, Reporter reporter) {

..
String result = some code here to compute result using key and value
Use HTTP Client to POST "result" to server


 }
}

I don't want to write the output of reducer to HDFS ie no
"output.collect(key, somevalue). Is there a way for me to do this using the
old api? Note: I must have a reducer because I am doing some calculation
here (not shown).

Suggestions? Thanks.



--
View this message in context: 
http://hadoop-common.472056.n3.nabble.com/Using-mapred-reducer-old-api-but-don-t-write-anything-to-HDFS-using-output-collector-tp3840442p3840442.html
Sent from the Users mailing list archive at Nabble.com.


Re: Problem setting super user

2012-03-19 Thread Harsh J
The right property for your version of Hadoop is
"dfs.permissions.supergroup". Change the property name, restart NN,
and your 'root' user should behave as a superuser afterwards.

Or, you can just do "sudo -u hdfs " as root.

On Mon, Mar 19, 2012 at 11:02 PM, Olivier Sallou
 wrote:
> Hi,
> I have installed Hadoop 1.0 using .deb package.
> I tried to configure superuser groups but it somehow fail. I do not know
> what's wrong:
>
> I expect root to be able to run hadoop dfsadmin -report command.
>
> # id hdfs
> uid=201(hdfs) gid=201(hdfs) groupes=123(hadoop),201(hdfs)
> # id root
> uid=0(root) gid=0(root) groupes=123(hadoop),0(root)
>
> Both my hdfs user (superuser by default) and my root user are in group
> hadoop.
>
> In hdfs-site.xml I have:
>  
>    dfs.permissions.superusergroup
>    hadoop
>    The name of the group of super-users.
>  
>
> I tried running refreshServiceAcl and
> refreshSuperUserGroupsConfiguration as hdfs, but I still have the error:
> root# hadoop dfsadmin -report
> report: org.apache.hadoop.security.AccessControlException: Access denied
> for user root. Superuser privilege is required
>
> Has anyone an idea?
>
> Thanks!
>
> Olivier
>
> --
>
> gpg key id: 4096R/326D8438  (keyring.debian.org)
> Key fingerprint = 5FB4 6F83 D3B9 5204 6335  D26D 78DC 68DB 326D 8438
>
>



-- 
Harsh J


Re: _temporary doesn't exist

2012-03-19 Thread Vipul Bharakhada
I found that "_temporary" directory is getting created and getting deleted
it self by hadoop. So is there any timeout for "_temporary" folder? I
noticed that if the job finishes within 5 minutes then I am not getting
this problem, but when it takes more than 50 minutes, the
"/output/_temporary" directory is missing.
Just to remind, I am working on 0.17.2 version.
-Vipul

On Fri, Mar 16, 2012 at 11:29 AM, Vipul Bharakhada wrote:

> Hi,
> I am using the old hadoop version 0.17.2 and I am getting the following
> exception when I am trying to run a Job. But it only happens at particular
> time, as cron jobs do run those task at particular intervals but it only
> fails at one particular time in day.
>
> Mar 14 06:49:23 7 08884: java.io.IOException: The directory
> hdfs://<{IPADDRESS}>:<{PORT}>/myserver/matcher/output/_temporary doesnt
> exist
> Mar 14 06:49:23 7 08884: \09at
> org.apache.hadoop.mapred.TaskTracker$TaskInProgress.localizeTask(TaskTracker.java:1439)
> Mar 14 06:49:23 7 08884: \09at
> org.apache.hadoop.mapred.TaskTracker$TaskInProgress.launchTask(TaskTracker.java:1511)
> Mar 14 06:49:23 7 08884: \09at
> org.apache.hadoop.mapred.TaskTracker.launchTaskForJob(TaskTracker.java:723)
> Mar 14 06:49:23 7 08884: \09at
> org.apache.hadoop.mapred.TaskTracker.localizeJob(TaskTracker.java:716)
> Mar 14 06:49:23 7 08884: \09at
> org.apache.hadoop.mapred.TaskTracker.startNewTask(TaskTracker.java:1274)
> Mar 14 06:49:23 7 08884: \09at
> org.apache.hadoop.mapred.TaskTracker.offerService(TaskTracker.java:915)
> Mar 14 06:49:23 7 08884: \09at
> org.apache.hadoop.mapred.TaskTracker.run(TaskTracker.java:1310)
> Mar 14 06:49:23 7 08884: \09at
> org.apache.hadoop.mapred.TaskTracker.main(TaskTracker.java:2251)
>
> What can be a problem? as this folder is created by hadoop internally and
> used internally, clean up is also done internally by hadoop. So why this
> directory is missing at some particular time?
> Any clue?
> -Vipul
>


Re: Problem setting super user

2012-03-19 Thread Mathias Herberts
does it work under user hdfs?
On Mar 19, 2012 6:32 PM, "Olivier Sallou"  wrote:

> Hi,
> I have installed Hadoop 1.0 using .deb package.
> I tried to configure superuser groups but it somehow fail. I do not know
> what's wrong:
>
> I expect root to be able to run hadoop dfsadmin -report command.
>
> # id hdfs
> uid=201(hdfs) gid=201(hdfs) groupes=123(hadoop),201(hdfs)
> # id root
> uid=0(root) gid=0(root) groupes=123(hadoop),0(root)
>
> Both my hdfs user (superuser by default) and my root user are in group
> hadoop.
>
> In hdfs-site.xml I have:
>  
>dfs.permissions.superusergroup
>hadoop
>The name of the group of super-users.
>  
>
> I tried running refreshServiceAcl and
> refreshSuperUserGroupsConfiguration as hdfs, but I still have the error:
> root# hadoop dfsadmin -report
> report: org.apache.hadoop.security.AccessControlException: Access denied
> for user root. Superuser privilege is required
>
> Has anyone an idea?
>
> Thanks!
>
> Olivier
>
> --
>
> gpg key id: 4096R/326D8438  (keyring.debian.org)
> Key fingerprint = 5FB4 6F83 D3B9 5204 6335  D26D 78DC 68DB 326D 8438
>
>
>


Problem setting super user

2012-03-19 Thread Olivier Sallou
Hi,
I have installed Hadoop 1.0 using .deb package.
I tried to configure superuser groups but it somehow fail. I do not know
what's wrong:

I expect root to be able to run hadoop dfsadmin -report command.

# id hdfs
uid=201(hdfs) gid=201(hdfs) groupes=123(hadoop),201(hdfs)
# id root
uid=0(root) gid=0(root) groupes=123(hadoop),0(root)

Both my hdfs user (superuser by default) and my root user are in group
hadoop.

In hdfs-site.xml I have:
  
dfs.permissions.superusergroup
hadoop
The name of the group of super-users.
  

I tried running refreshServiceAcl and
refreshSuperUserGroupsConfiguration as hdfs, but I still have the error:
root# hadoop dfsadmin -report
report: org.apache.hadoop.security.AccessControlException: Access denied
for user root. Superuser privilege is required

Has anyone an idea?

Thanks!

Olivier

-- 

gpg key id: 4096R/326D8438  (keyring.debian.org)
Key fingerprint = 5FB4 6F83 D3B9 5204 6335  D26D 78DC 68DB 326D 8438




Re: need help regarding eclipse plugin for hadoop

2012-03-19 Thread Shailesh Samudrala
Hi,

On top of the Hadoop eclipse plugin, you already need to have a Hadoop
Virtual Machine, where eclipse can run the code, and a local library of
hadoop, which will be used by Eclipse to compile your code.
Below are the steps that I used to run a sample wordcount example on the
Hadoop VM.
1. Download this virtual appliance:
http://shared.zenfs.com/hadoop-vm-appliance-0-20-S.zip

Also, download a relevant version of hadoop onto your computer, you will
need to provide the path to the hadoop library to Eclipse.

2. Login using: hadoop-user, pwd: hadoop

3. Try this command: hadoop dfs -ls
if you get an error, then use this command
kinit -k -t /home/hadoop-user/hadoop-user.keytab
hadoop-user/localhost@LOCALDOMAIN

This will setup your Hadoop VM

Next, to install the plugin on eclipse,

1. Download this plugin, and place it in the plugin directory in eclipse
http://code.google.com/edu/parallel/tools/hadoopvm/hadoop-eclipse-plugin.jar

2. Use the steps mentioned here to configure the plugin in eclipse:
http://code.google.com/edu/parallel/tools/hadoopvm/index.html

Host id is the IP address that is printed on VM startup
Hadoop directory: /home/hadoop-user/hadoop

3. Create input directory in hadoop DFS, which will be input directory for
map reduce, and create a text file which will be the input file.

a) vi temp.txt
b) mkdir input
c) mv temp.txt input/temp.txt
d) hadoop dfs copyFromLocal input .
e) mkdir output
f) hadoop dfs copyFromLoacl output .

4. Then run the code attached on the hadoop server, if it runs
successfully, the output will be stored in the output directory.
a) vi output/part-r-0


On Mon, Mar 19, 2012 at 12:29 PM, tabraiz anwer wrote:

> Hi,
>  I am newbie to the hadoop, and trying to configure eclipse plugin of
> hadoop but its resposnse is very awkward and giving me the error.
> Error : Unable to login. when trying to connect to hadoop DFS  using
> Hadoop plugin for eclipse
>
>
> using hadoop : hadoop-0.20.203.0
> eclipse plugin : hadoop-eclipse-plugin-0.20.203.0.jar
> eclipse : eclipse indigo
> using Ubuntu 11.04 64bit.


need help regarding eclipse plugin for hadoop

2012-03-19 Thread tabraiz anwer
Hi,
 I am newbie to the hadoop, and trying to configure eclipse plugin of 
hadoop but its resposnse is very awkward and giving me the error.
Error : Unable to login. when trying to connect to hadoop DFS  using Hadoop 
plugin for eclipse


using hadoop : hadoop-0.20.203.0
eclipse plugin : hadoop-eclipse-plugin-0.20.203.0.jar
eclipse : eclipse indigo
using Ubuntu 11.04 64bit.

Re: EOFException

2012-03-19 Thread Mohit Anchlia
I guess I am trying to see how to debug such problems? I don't see enough
info in the logs.


On Mon, Mar 19, 2012 at 12:48 AM, madhu phatak  wrote:

> Hi,
>  Seems like HDFS is in safemode.
>
> On Fri, Mar 16, 2012 at 1:37 AM, Mohit Anchlia  >wrote:
>
> > This is actually just hadoop job over HDFS. I am assuming you also know
> why
> > this is erroring out?
> >
> > On Thu, Mar 15, 2012 at 1:02 PM, Gopal  wrote:
> >
> > >  On 03/15/2012 03:06 PM, Mohit Anchlia wrote:
> > >
> > >> When I start a job to read data from HDFS I start getting these
> errors.
> > >> Does anyone know what this means and how to resolve it?
> > >>
> > >> 2012-03-15 10:41:31,402 [Thread-5] INFO
> >  org.apache.hadoop.hdfs.**DFSClient
> > >> -
> > >> Exception in createBlockOutputStream 164.28.62.204:50010java.io.**
> > >> EOFException
> > >> 2012-03-15 10:41:31,402 [Thread-5] INFO
> >  org.apache.hadoop.hdfs.**DFSClient
> > >> -
> > >> Abandoning block blk_-6402969611996946639_11837
> > >> 2012-03-15 10:41:31,403 [Thread-5] INFO
> >  org.apache.hadoop.hdfs.**DFSClient
> > >> -
> > >> Excluding datanode 164.28.62.204:50010
> > >> 2012-03-15 10:41:31,406 [Thread-5] INFO
> >  org.apache.hadoop.hdfs.**DFSClient
> > >> -
> > >> Exception in createBlockOutputStream 164.28.62.198:50010java.io.**
> > >> EOFException
> > >> 2012-03-15 10:41:31,406 [Thread-5] INFO
> >  org.apache.hadoop.hdfs.**DFSClient
> > >> -
> > >> Abandoning block blk_-5442664108986165368_11838
> > >> 2012-03-15 10:41:31,407 [Thread-5] INFO
> >  org.apache.hadoop.hdfs.**DFSClient
> > >> -
> > >> Exception in createBlockOutputStream 164.28.62.197:50010java.io.**
> > >> EOFException
> > >> 2012-03-15 10:41:31,407 [Thread-5] INFO
> >  org.apache.hadoop.hdfs.**DFSClient
> > >> -
> > >> Abandoning block blk_-3373089616877234160_11838
> > >> 2012-03-15 10:41:31,407 [Thread-5] INFO
> >  org.apache.hadoop.hdfs.**DFSClient
> > >> -
> > >> Excluding datanode 164.28.62.198:50010
> > >> 2012-03-15 10:41:31,409 [Thread-5] INFO
> >  org.apache.hadoop.hdfs.**DFSClient
> > >> -
> > >> Excluding datanode 164.28.62.197:50010
> > >> 2012-03-15 10:41:31,410 [Thread-5] INFO
> >  org.apache.hadoop.hdfs.**DFSClient
> > >> -
> > >> Exception in createBlockOutputStream 164.28.62.204:50010java.io.**
> > >> EOFException
> > >> 2012-03-15 10:41:31,410 [Thread-5] INFO
> >  org.apache.hadoop.hdfs.**DFSClient
> > >> -
> > >> Abandoning block blk_4481292025401332278_11838
> > >> 2012-03-15 10:41:31,411 [Thread-5] INFO
> >  org.apache.hadoop.hdfs.**DFSClient
> > >> -
> > >> Excluding datanode 164.28.62.204:50010
> > >> 2012-03-15 10:41:31,412 [Thread-5] INFO
> >  org.apache.hadoop.hdfs.**DFSClient
> > >> -
> > >> Exception in createBlockOutputStream 164.28.62.200:50010java.io.**
> > >> EOFException
> > >> 2012-03-15 10:41:31,412 [Thread-5] INFO
> >  org.apache.hadoop.hdfs.**DFSClient
> > >> -
> > >> Abandoning block blk_-5326771177080888701_11838
> > >> 2012-03-15 10:41:31,413 [Thread-5] INFO
> >  org.apache.hadoop.hdfs.**DFSClient
> > >> -
> > >> Excluding datanode 164.28.62.200:50010
> > >> 2012-03-15 10:41:31,414 [Thread-5] INFO
> >  org.apache.hadoop.hdfs.**DFSClient
> > >> -
> > >> Exception in createBlockOutputStream 164.28.62.197:50010java.io.**
> > >> EOFException
> > >> 2012-03-15 10:41:31,414 [Thread-5] INFO
> >  org.apache.hadoop.hdfs.**DFSClient
> > >> -
> > >> Abandoning block blk_-8073750683705518772_11839
> > >> 2012-03-15 10:41:31,415 [Thread-5] INFO
> >  org.apache.hadoop.hdfs.**DFSClient
> > >> -
> > >> Excluding datanode 164.28.62.197:50010
> > >> 2012-03-15 10:41:31,416 [Thread-5] INFO
> >  org.apache.hadoop.hdfs.**DFSClient
> > >> -
> > >> Exception in createBlockOutputStream 164.28.62.199:50010java.io.**
> > >> EOFException
> > >> 2012-03-15 10:41:31,416 [Thread-5] INFO
> >  org.apache.hadoop.hdfs.**DFSClient
> > >> -
> > >> Exception in createBlockOutputStream 164.28.62.198:50010java.io.**
> > >> EOFException
> > >> 2012-03-15 10:41:31,416 [Thread-5] INFO
> >  org.apache.hadoop.hdfs.**DFSClient
> > >> -
> > >> Abandoning block blk_441003866688859169_11838
> > >> 2012-03-15 10:41:31,416 [Thread-5] INFO
> >  org.apache.hadoop.hdfs.**DFSClient
> > >> -
> > >> Abandoning block blk_-466858474055876377_11839
> > >> 2012-03-15 10:41:31,417 [Thread-5] INFO
> >  org.apache.hadoop.hdfs.**DFSClient
> > >> -
> > >> Excluding datanode 164.28.62.198:50010
> > >> 2012-03-15 10:41:31,417 [Thread-5] WARN
> >  org.apache.hadoop.hdfs.**DFSClient
>  > >> -
> > >>
> > >>
> > > Try shutting down and  restarting hbase.
> > >
> >
>
>
>
> --
> https://github.com/zinnia-phatak-dev/Nectar
>


Re: EOFException

2012-03-19 Thread madhu phatak
Hi,
 Seems like HDFS is in safemode.

On Fri, Mar 16, 2012 at 1:37 AM, Mohit Anchlia wrote:

> This is actually just hadoop job over HDFS. I am assuming you also know why
> this is erroring out?
>
> On Thu, Mar 15, 2012 at 1:02 PM, Gopal  wrote:
>
> >  On 03/15/2012 03:06 PM, Mohit Anchlia wrote:
> >
> >> When I start a job to read data from HDFS I start getting these errors.
> >> Does anyone know what this means and how to resolve it?
> >>
> >> 2012-03-15 10:41:31,402 [Thread-5] INFO
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >> Exception in createBlockOutputStream 164.28.62.204:50010java.io.**
> >> EOFException
> >> 2012-03-15 10:41:31,402 [Thread-5] INFO
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >> Abandoning block blk_-6402969611996946639_11837
> >> 2012-03-15 10:41:31,403 [Thread-5] INFO
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >> Excluding datanode 164.28.62.204:50010
> >> 2012-03-15 10:41:31,406 [Thread-5] INFO
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >> Exception in createBlockOutputStream 164.28.62.198:50010java.io.**
> >> EOFException
> >> 2012-03-15 10:41:31,406 [Thread-5] INFO
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >> Abandoning block blk_-5442664108986165368_11838
> >> 2012-03-15 10:41:31,407 [Thread-5] INFO
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >> Exception in createBlockOutputStream 164.28.62.197:50010java.io.**
> >> EOFException
> >> 2012-03-15 10:41:31,407 [Thread-5] INFO
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >> Abandoning block blk_-3373089616877234160_11838
> >> 2012-03-15 10:41:31,407 [Thread-5] INFO
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >> Excluding datanode 164.28.62.198:50010
> >> 2012-03-15 10:41:31,409 [Thread-5] INFO
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >> Excluding datanode 164.28.62.197:50010
> >> 2012-03-15 10:41:31,410 [Thread-5] INFO
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >> Exception in createBlockOutputStream 164.28.62.204:50010java.io.**
> >> EOFException
> >> 2012-03-15 10:41:31,410 [Thread-5] INFO
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >> Abandoning block blk_4481292025401332278_11838
> >> 2012-03-15 10:41:31,411 [Thread-5] INFO
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >> Excluding datanode 164.28.62.204:50010
> >> 2012-03-15 10:41:31,412 [Thread-5] INFO
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >> Exception in createBlockOutputStream 164.28.62.200:50010java.io.**
> >> EOFException
> >> 2012-03-15 10:41:31,412 [Thread-5] INFO
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >> Abandoning block blk_-5326771177080888701_11838
> >> 2012-03-15 10:41:31,413 [Thread-5] INFO
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >> Excluding datanode 164.28.62.200:50010
> >> 2012-03-15 10:41:31,414 [Thread-5] INFO
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >> Exception in createBlockOutputStream 164.28.62.197:50010java.io.**
> >> EOFException
> >> 2012-03-15 10:41:31,414 [Thread-5] INFO
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >> Abandoning block blk_-8073750683705518772_11839
> >> 2012-03-15 10:41:31,415 [Thread-5] INFO
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >> Excluding datanode 164.28.62.197:50010
> >> 2012-03-15 10:41:31,416 [Thread-5] INFO
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >> Exception in createBlockOutputStream 164.28.62.199:50010java.io.**
> >> EOFException
> >> 2012-03-15 10:41:31,416 [Thread-5] INFO
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >> Exception in createBlockOutputStream 164.28.62.198:50010java.io.**
> >> EOFException
> >> 2012-03-15 10:41:31,416 [Thread-5] INFO
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >> Abandoning block blk_441003866688859169_11838
> >> 2012-03-15 10:41:31,416 [Thread-5] INFO
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >> Abandoning block blk_-466858474055876377_11839
> >> 2012-03-15 10:41:31,417 [Thread-5] INFO
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >> Excluding datanode 164.28.62.198:50010
> >> 2012-03-15 10:41:31,417 [Thread-5] WARN
>  org.apache.hadoop.hdfs.**DFSClient
> >> -
> >>
> >>
> > Try shutting down and  restarting hbase.
> >
>



-- 
https://github.com/zinnia-phatak-dev/Nectar


Re: Setting up MapReduce 2 on a test cluster

2012-03-19 Thread fleandr
Try to check if all necessary files exist in /share/hdfs and /share/mapreduce
folder.