Can Hadoop handle many appliations

2009-06-20 Thread ashish pareek
Hello Everybody,

   I have a few questions regarding running applications
on Hadoop.

1) Can a single Hadoop cluster simultaneously handle applications of same
type ( or of different type).

2) What are the challenges in running applications of different types in
Hadoop. How Does it differ from running applications of same type.


Thanking you,

Regards,
Ashish.


Fwd: Need help

2009-06-18 Thread ashish pareek
Hello,
I am doing my master my final year project is on Hadoop ...so I
would like to know some thing about Hadoop cluster i.e, Do new version of
Hadoop are able to handle heterogeneous hardware.If you have any
informantion regarding these please mail me as my project is in heterogenous
environment.


Thanks!

Reagrds,
Ashish Pareek


Re: Need help

2009-06-18 Thread ashish pareek
Does that mean hadoop is not scalable wrt heterogeneous environment? and one
more question is can we run different application on the same hadoop cluster
.

Thanks.
Regards,
Ashish

On Thu, Jun 18, 2009 at 8:30 PM, jason hadoop jason.had...@gmail.comwrote:

 Hadoop has always been reasonably agnostic wrt hardware and homogeneity.
 There are optimizations in configuration for near homogeneous machines.



 On Thu, Jun 18, 2009 at 7:46 AM, ashish pareek pareek...@gmail.com
 wrote:

  Hello,
 I am doing my master my final year project is on Hadoop ...so
 I
  would like to know some thing about Hadoop cluster i.e, Do new version of
  Hadoop are able to handle heterogeneous hardware.If you have any
  informantion regarding these please mail me as my project is in
  heterogenous
  environment.
 
 
  Thanks!
 
  Reagrds,
  Ashish Pareek
 



 --
 Pro Hadoop, a book to guide you from beginner to hadoop mastery,
 http://www.amazon.com/dp/1430219424?tag=jewlerymall
 www.prohadoopbook.com a community for Hadoop Professionals



Re: Need help

2009-06-18 Thread ashish pareek
Can you tell few of the challenges in configuring heterogeneous cluster...or
can pass on some link where I would get some information regarding
challenges in running Hadoop on heterogeneous hardware

One more things is How about running different applications on the same
Hadoop cluster?and what challenges are involved in it ?

Thanks,
Regards,
Ashish


On Thu, Jun 18, 2009 at 8:53 PM, jason hadoop jason.had...@gmail.comwrote:

 I don't know anyone who has a completely homogeneous cluster.

 So hadoop is scalable across heterogeneous environments.

 I stated that configuration is simpler if the machines are similar (There
 are optimizations in configuration for near homogeneous machines.)

 On Thu, Jun 18, 2009 at 8:10 AM, ashish pareek pareek...@gmail.com
 wrote:

  Does that mean hadoop is not scalable wrt heterogeneous environment? and
  one
  more question is can we run different application on the same hadoop
  cluster
  .
 
  Thanks.
  Regards,
  Ashish
 
  On Thu, Jun 18, 2009 at 8:30 PM, jason hadoop jason.had...@gmail.com
  wrote:
 
   Hadoop has always been reasonably agnostic wrt hardware and
 homogeneity.
   There are optimizations in configuration for near homogeneous machines.
  
  
  
   On Thu, Jun 18, 2009 at 7:46 AM, ashish pareek pareek...@gmail.com
   wrote:
  
Hello,
   I am doing my master my final year project is on Hadoop
  ...so
   I
would like to know some thing about Hadoop cluster i.e, Do new
 version
  of
Hadoop are able to handle heterogeneous hardware.If you have any
informantion regarding these please mail me as my project is in
heterogenous
environment.
   
   
Thanks!
   
Reagrds,
Ashish Pareek
   
  
  
  
   --
   Pro Hadoop, a book to guide you from beginner to hadoop mastery,
   http://www.amazon.com/dp/1430219424?tag=jewlerymall
   www.prohadoopbook.com a community for Hadoop Professionals
  
 



 --
 Pro Hadoop, a book to guide you from beginner to hadoop mastery,
 http://www.amazon.com/dp/1430219424?tag=jewlerymall
 www.prohadoopbook.com a community for Hadoop Professionals



Re: Need help

2009-06-18 Thread ashish pareek
Hello Everybody,

  How can we handle different applications having
different requirement being run on the same hadoop cluster ? What are the
various approaches to solve such problem.. if possible please mention some
of those ideas.

Does such implementation exists ?

Thanks ,

Regards,
Ashish

On Thu, Jun 18, 2009 at 9:36 PM, jason hadoop jason.had...@gmail.comwrote:

 For me, I like to have one configuration file that I distribute to all of
 the machines in my cluster via rsync.

 In there are things like the number of tasks per node to run, and where to
 store dfs data and local temporary data, and the limits to storage for the
 machines.

 If the machines are very different, it becomes important to tailor the
 configuration file per machine or type of machine.

 At this point, you are pretty much going to have to spend the time, reading
 through the details of configuring a hadoop cluster.


 On Thu, Jun 18, 2009 at 8:33 AM, ashish pareek pareek...@gmail.com
 wrote:

  Can you tell few of the challenges in configuring heterogeneous
  cluster...or
  can pass on some link where I would get some information regarding
  challenges in running Hadoop on heterogeneous hardware
 
  One more things is How about running different applications on the same
  Hadoop cluster?and what challenges are involved in it ?
 
  Thanks,
  Regards,
  Ashish
 
 
  On Thu, Jun 18, 2009 at 8:53 PM, jason hadoop jason.had...@gmail.com
  wrote:
 
   I don't know anyone who has a completely homogeneous cluster.
  
   So hadoop is scalable across heterogeneous environments.
  
   I stated that configuration is simpler if the machines are similar
 (There
   are optimizations in configuration for near homogeneous machines.)
  
   On Thu, Jun 18, 2009 at 8:10 AM, ashish pareek pareek...@gmail.com
   wrote:
  
Does that mean hadoop is not scalable wrt heterogeneous environment?
  and
one
more question is can we run different application on the same hadoop
cluster
.
   
Thanks.
Regards,
Ashish
   
On Thu, Jun 18, 2009 at 8:30 PM, jason hadoop 
 jason.had...@gmail.com
wrote:
   
 Hadoop has always been reasonably agnostic wrt hardware and
   homogeneity.
 There are optimizations in configuration for near homogeneous
  machines.



 On Thu, Jun 18, 2009 at 7:46 AM, ashish pareek 
 pareek...@gmail.com
 wrote:

  Hello,
 I am doing my master my final year project is on
 Hadoop
...so
 I
  would like to know some thing about Hadoop cluster i.e, Do new
   version
of
  Hadoop are able to handle heterogeneous hardware.If you have any
  informantion regarding these please mail me as my project is in
  heterogenous
  environment.
 
 
  Thanks!
 
  Reagrds,
  Ashish Pareek
 



 --
 Pro Hadoop, a book to guide you from beginner to hadoop mastery,
 http://www.amazon.com/dp/1430219424?tag=jewlerymall
 www.prohadoopbook.com a community for Hadoop Professionals

   
  
  
  
   --
   Pro Hadoop, a book to guide you from beginner to hadoop mastery,
   http://www.amazon.com/dp/1430219424?tag=jewlerymall
   www.prohadoopbook.com a community for Hadoop Professionals
  
 



 --
 Pro Hadoop, a book to guide you from beginner to hadoop mastery,
 http://www.amazon.com/dp/1430219424?tag=jewlerymall
 www.prohadoopbook.com a community for Hadoop Professionals



Re: Hadoop error help- file system closed, could only be replicated to 0 nodes, instead of 1

2009-06-18 Thread ashish pareek
HI ,
 What seems from your details is that datanode is not running.can
you run *bin/hadoop dfsadmin -report*  and find out whether your datanodes
are up ? then post your observation and it would be better if you post even
your hadoop-site.xml file deatils also.

Regards,
Ashish.

On Fri, Jun 19, 2009 at 3:16 AM, terrianne.erick...@accenture.com wrote:

 Hi,

 I am extremely new to Hadoop and have come across a few errors that I'm not
 sure how to fix. I am running Hadoop version 0.19.0 from an image through
 Elasticfox and S3. I am on windows and use puTTY as my ssh. I am trying to
 run a wordcount with 5 slaves. This is what I do so far:

 1. boot up the instance through ElasticFox
 2. cd /usr/local/hadoop-0.19.0
 3. bin/hadoop namenode -format
 4. bin/start-all.sh
 5. jps --( shows jps, jobtracker, secondarynamenode)
 6.bin/stop-all.sh
 7. ant examples
 8. bin/start-all.sh
 9. bin/hadoop jar build/hadoop-0.19.0-examples.jar pi 0 100

 Then I get this error trace:

 Number of Maps = 0 Samples per Map = 100
 Starting Job
 09/06/18 17:31:25 INFO hdfs.DFSClient:
 org.apache.hadoop.ipc.RemoteException: java.io.IOException: File
 /mnt/hadoop/mapred/system/job_200906181730_0001/job.jar could only be
 replicated to 0 nodes, instead of 1
at
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1270)
at
 org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:351)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:452)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:892)

at org.apache.hadoop.ipc.Client.call(Client.java:696)
at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:216)
at $Proxy0.addBlock(Unknown Source)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at
 org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
at
 org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
at $Proxy0.addBlock(Unknown Source)
at
 org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.locateFollowingBlock(DFSClient.java:2815)
at
 org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2697)
at
 org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:1997)
at
 org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2183)

 09/06/18 17:31:25 WARN hdfs.DFSClient: NotReplicatedYetException sleeping
 /mnt/hadoop/mapred/system/job_200906181730_0001/job.jar retries left 4
 09/06/18 17:31:25 INFO hdfs.DFSClient:
 org.apache.hadoop.ipc.RemoteException: java.io.IOException: File
 /mnt/hadoop/mapred/system/job_200906181730_0001/job.jar could only be
 replicated to 0 nodes, instead of 1
at
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1270)
at
 org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:351)
at sun.reflec,t.NativeMethodAccessorImpl.invoke0(Native Method)
at
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:452)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:892)

at org.apache.hadoop.ipc.Client.call(Client.java:696)
at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:216)
at $Proxy0.addBlock(Unknown Source)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at
 org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
at
 org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
at $Proxy0.addBlock(Unknown Source)
at
 org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.locateFollowingBlock(DFSClient.java:2815)
at
 

Problem in viewing WEB UI

2009-06-16 Thread ashish pareek
Hi,

  When I run command *bin/hadoop dfsadmin -report *it shows that 2
datanodes are alive but when i try to http://hadoopmster:50070/ but the
problem is that it opens doesnot opne
http://hadoopmaster:50070/dfshealth.jsp page and throws *error HTTP: 404 .
So why is't happening like this?
*
Regards,
Ashish Pareek


 On Wed, Jun 17, 2009 at 10:06 AM, Sugandha Neaolekar 
sugandha@gmail.com wrote:

 Well, You just have to specify the address in the URL address bar as::
 http://hadoopmaster:50070 U'll be able to see the web UI..!


 On Tue, Jun 16, 2009 at 7:17 PM, ashish pareek pareek...@gmail.comwrote:

 HI Sugandha,
Hmmm your suggestion helped and Now I am able
 to run two datanode one on the same machine as name node and other on
 the different machine Thanks a lot :)

  But the problem is now I am not able to see web UI .
 for  both datanode and as well as name node
 should I have to consider some more things in the site.xml ? if so please
 help...

 Thanking you again,
 regards,
 Ashish Pareek.

 On Tue, Jun 16, 2009 at 3:10 PM, Sugandha Naolekar 
 sugandha@gmail.com wrote:

 hi,,!


 First of all, get your concepts clear of hadoop.
 You can refer to the following

 site::
 http://www.google.co.in/url?sa=tsource=webct=rescd=1url=http%3A%2F%2Fwww.michael-noll.com%2Fwiki%2FRunning_Hadoop_On_Ubuntu_Linux_(Single-Node_Cluster)ei=lGU3Spv2FZbLjAe19KmiDQusg=AFQjCNFbmVGsoChOSMzCB3tRhoV0ylHOzAsig2=t2AJ_nf24SFtveN4PHS_TAhttp://www.google.co.in/url?sa=tsource=webct=rescd=1url=http%3A%2F%2Fwww.michael-noll.com%2Fwiki%2FRunning_Hadoop_On_Ubuntu_Linux_%28Single-Node_Cluster%29ei=lGU3Spv2FZbLjAe19KmiDQusg=AFQjCNFbmVGsoChOSMzCB3tRhoV0ylHOzAsig2=t2AJ_nf24SFtveN4PHS_TA


 I have small doubt whether in the mater.xml and slave.xml we can have
 same port numbers to both of them like


 for slave :

 property
 namefs.default.name/name
 valuehdfs://hadoopslave:

 9000/value
   /property


  for master:::

 property
 namefs.default.name/name
 valuehdfs://hadoopmaster:9000/value
   /property



 Well, any  two daemons or services can run on the same port unless, they
 are not run on the same machine.If you wish to run DN and NN on the same
 machine, their port numbers have to be different.




 On Tue, Jun 16, 2009 at 2:55 PM, ashish pareek pareek...@gmail.comwrote:

 HI sugandha,



 and one more thing can we have in slave:::

 property
   namedfs.datanode.address/

 name
   valuehadoopmaster:9000/value
 valuehadoopslave:9001/value
   /property



 Also, fs,default.name is the tag which specifies the default filesystem.
 And generaLLY, it is run on namenode. So, it;s value has to be a namenode's
 address only and not slave's.



 Else if you have complete procedure for installing and running Hadoop in
 cluster can you please send it to me .. I need to step up hadoop with 
 in
 two days and show it to my guide.Currently I am doing my masters.

 Thanks for your spending time


 Try for the above, and this should work!



 regards,
 Ashish Pareek


 On Tue, Jun 16, 2009 at 2:33 PM, Sugandha Naolekar 
 sugandha@gmail.com wrote:

 Following changes are to be done::

 Under master folder::

 - put slaves address as well under the values of
 tag(dfs.datanode.address)

 - You want to make namenode as datanode as well. As per your config
 file, you have specified hadoopmaster  in your slave file. If you don't 
 want
 that, remove ti from slaves file.

 UNder slave folder::

 - put only slave's (the m/c where you intend to run your datanode)'s
 address.under datanode.address tag. Else
 it should go as such::

 property
   namedfs.datanode.address/name
   valuehadoopmaster:9000/value
 valuehadoopslave:9001/value
   /property

 Also, your port numbers hould be different. the daemons NN,DN,JT,TT
 should run independently on different ports.


 On Tue, Jun 16, 2009 at 2:05 PM, Sugandha Naolekar 
 sugandha@gmail.com wrote:



 -- Forwarded message --
 From: ashish pareek pareek...@gmail.com
 Date: Tue, Jun 16, 2009 at 2:00 PM
 Subject: Re: org.apache.hadoop.ipc.client : trying connect to server
 failed
 To: Sugandha Naolekar sugandha@gmail.com




 On Tue, Jun 16, 2009 at 1:58 PM, ashish pareek 
 pareek...@gmail.comwrote:

 HI ,
  I am sending .tar.gz format containing both master and datanode
 config files ...

 Regards,
 Ashish Pareek


 On Tue, Jun 16, 2009 at 1:47 PM, Sugandha Naolekar 
 sugandha@gmail.com wrote:

 can u pls send me a zip or a tar file? I don't have windows systems
 but, linux


 On Tue, Jun 16, 2009 at 1:19 PM, ashish pareek pareek...@gmail.com
  wrote:

 HI Sungandha ,
   Thanks for your reply  I am sending you
 master and slave configuration files if you can go through it and 
 tell me
 where I am going wrong it would be helpful.

 Hope to get a reply soon ... Thanks
 again!

 Regards,
 Ashish Pareek

 On Tue, Jun 16

org.apache.hadoop.ipc.client : trying connect to server failed

2009-05-28 Thread ashish pareek
HI ,
 I am trying to step up a hadoop cluster on 512 MB machine and using
hadoop 0.18 and  have followed procedure given in  apache hadoop site for
hadoop cluster.
 I included  in conf/slaves two datanode i.e including the namenode
vitrual machine and other machine virtual machine  . and have set up
passwordless ssh between both virtual machines. But now problem is when
is run command 

bin/hadoop start-all.sh

It start only one datanode on the same namenode vitrual machine but it
doesn't start the datanode on other machine.

in logs/hadoop-datanode  i get message


 INFO org.apache.hadoop.ipc.Client: Retrying
 connect to server: hadoop1/192.168.1.28:9000. Already
  tried 1 time(s).
 2009-05-09 18:35:14,266 INFO org.apache.hadoop.ipc.Client: Retrying
 connect to server: hadoop1/192.168.1.28:9000. Already tried 2 time(s).
 2009-05-09 18:35:14,266 INFO org.apache.hadoop.ipc.Client: Retrying
 connect to server: hadoop1/192.168.1.28:9000. Already tried 3 time(s).

.
.
.
.
.
.
.
.
.

.
.

.


So can any one help in solving this problem. :)

Thanks

Regards
Ashish Pareek


Re: org.apache.hadoop.ipc.client : trying connect to server failed

2009-05-28 Thread ashish pareek
Hi some help me out 



On Thu, May 28, 2009 at 10:32 PM, ashish pareek pareek...@gmail.com wrote:

 HI ,
  I am trying to step up a hadoop cluster on 512 MB machine and using
 hadoop 0.18 and  have followed procedure given in  apache hadoop site for
 hadoop cluster.
  I included  in conf/slaves two datanode i.e including the namenode
 vitrual machine and other machine virtual machine  . and have set up
 passwordless ssh between both virtual machines. But now problem is when
 is run command 

 bin/hadoop start-all.sh

 It start only one datanode on the same namenode vitrual machine but it
 doesn't start the datanode on other machine.

 in logs/hadoop-datanode  i get message


  INFO org.apache.hadoop.ipc.Client: Retrying
  connect to server: hadoop1/192.168.1.28:9000. Already
   tried 1 time(s).

  2009-05-09 18:35:14,266 INFO org.apache.hadoop.ipc.Client: Retrying
  connect to server: hadoop1/192.168.1.28:9000. Already tried 2 time(s).
  2009-05-09 18:35:14,266 INFO org.apache.hadoop.ipc.Client: Retrying
  connect to server: hadoop1/192.168.1.28:9000. Already tried 3 time(s).

 .
 .
 .
 .
 .
 .
 .
 .
 .

 .
 .

 .


 So can any one help in solving this problem. :)

 Thanks

 Regards

 Ashish Pareek





Re: org.apache.hadoop.ipc.client : trying connect to server failed

2009-05-28 Thread ashish pareek
Yes I am able to ping and ssh between two virtual machine and even i
have set ip address of both the virtual machines in their respective
/etc/hosts file ...

thanx for reply .. if you suggest some other thing which i could
have missed or any remedy 

Regards,
Ashish Pareek.


On Fri, May 29, 2009 at 10:04 AM, Pankil Doshi forpan...@gmail.com wrote:

 make sure u can ping that data node and ssh it.


 On Thu, May 28, 2009 at 12:02 PM, ashish pareek pareek...@gmail.com
 wrote:

  HI ,
  I am trying to step up a hadoop cluster on 512 MB machine and using
  hadoop 0.18 and  have followed procedure given in  apache hadoop site for
  hadoop cluster.
  I included  in conf/slaves two datanode i.e including the namenode
  vitrual machine and other machine virtual machine  . and have set up
  passwordless ssh between both virtual machines. But now problem is
 when
  is run command 
 
  bin/hadoop start-all.sh
 
  It start only one datanode on the same namenode vitrual machine but it
  doesn't start the datanode on other machine.
 
  in logs/hadoop-datanode  i get message
 
 
   INFO org.apache.hadoop.ipc.Client: Retrying
   connect to server: hadoop1/192.168.1.28:9000. Already
   tried 1 time(s).
   2009-05-09 18:35:14,266 INFO org.apache.hadoop.ipc.Client: Retrying
   connect to server: hadoop1/192.168.1.28:9000. Already tried 2 time(s).
   2009-05-09 18:35:14,266 INFO org.apache.hadoop.ipc.Client: Retrying
   connect to server: hadoop1/192.168.1.28:9000. Already tried 3 time(s).
 
  .
  .
  .
  .
  .
  .
  .
  .
  .
 
  .
  .
 
  .
 
 
  So can any one help in solving this problem. :)
 
  Thanks
 
  Regards
  Ashish Pareek
 



Re: Could only be replicated to 0 nodes, instead of 1

2009-05-21 Thread ashish pareek
Hi ,

I have two suggestion

i)Choose a right version ( Hadoop- 0.18 is good)
ii)replication should be 3 as ur having 3 modes.( Indirectly see to it that
ur configuration is correct !!)

Hey even i am just suggesting this as i am also a new to hadoop

Ashish Pareek


On Thu, May 21, 2009 at 2:41 PM, Stas Oskin stas.os...@gmail.com wrote:

 Hi.

 I'm testing Hadoop in our lab, and started getting the following message
 when trying to copy a file:
 Could only be replicated to 0 nodes, instead of 1

 I have the following setup:

 * 3 machines, 2 of them with only 80GB of space, and 1 with 1.5GB
 * Two clients are copying files all the time (one of them is the 1.5GB
 machine)
 * The replication is set on 2
 * I let the space on 2 smaller machines to end, to test the behavior

 Now, one of the clients (the one located on 1.5GB) works fine, and the
 other
 one - the external, unable to copy and displays the error + the exception
 below

 Any idea if this expected on my scenario? Or how it can be solved?

 Thanks in advance.



 09/05/21 10:51:03 WARN dfs.DFSClient: NotReplicatedYetException sleeping
 /test/test.bin retries left 1

 09/05/21 10:51:06 WARN dfs.DFSClient: DataStreamer Exception:
 org.apache.hadoop.ipc.RemoteException: java.io.IOException: File
 /test/test.bin could only be replicated to 0 nodes, instead of 1

at

 org.apache.hadoop.dfs.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1123
 )

at org.apache.hadoop.dfs.NameNode.addBlock(NameNode.java:330)

at sun.reflect.GeneratedMethodAccessor8.invoke(Unknown Source)

at

 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25
 )

at java.lang.reflect.Method.invoke(Method.java:597)

at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:481)

at org.apache.hadoop.ipc.Server$Handler.run(Server.java:890)



at org.apache.hadoop.ipc.Client.call(Client.java:716)

at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:216)

at org.apache.hadoop.dfs.$Proxy0.addBlock(Unknown Source)

at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

at

 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39
 )

at

 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25
 )

at java.lang.reflect.Method.invoke(Method.java:597)

at

 org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82
 )

at

 org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59
 )

at org.apache.hadoop.dfs.$Proxy0.addBlock(Unknown Source)

at

 org.apache.hadoop.dfs.DFSClient$DFSOutputStream.locateFollowingBlock(DFSClient.java:2450
 )

at

 org.apache.hadoop.dfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2333
 )

at

 org.apache.hadoop.dfs.DFSClient$DFSOutputStream.access$1800(DFSClient.java:1745
 )

at

 org.apache.hadoop.dfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:1922
 )



 09/05/21 10:51:06 WARN dfs.DFSClient: Error Recovery for block null bad
 datanode[0]

 java.io.IOException: Could not get block locations. Aborting...

at

 org.apache.hadoop.dfs.DFSClient$DFSOutputStream.processDatanodeError(DFSClient.java:2153
 )

at

 org.apache.hadoop.dfs.DFSClient$DFSOutputStream.access$1400(DFSClient.java:1745
 )

at

 org.apache.hadoop.dfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:1899
 )



TASKS KILLED WHEN RUNNING : bin/hadoop jar hadoop-*-examples.jar grep input output 'dfs[a-z.]+'

2009-05-16 Thread ashish pareek
hELLO TO EVERY BODY
   I AM A NEW HAOODP USER I STARTED RUNNING HADOOP USING SITE
http://hadoop.apache.org/core/docs/current/quickstart.html BUT WHEN I RUN
COMMAND   bin/hadoop jar hadoop-*-examples.jar grep input output
'dfs[a-z.]+'  , IN PSEDUO DISTRIBUTED MODE I GET ERROR LIKE :::

Task  task_200801251420_0007_m_06_0 failed to report status for 601
seconds.  Killing!

Task  task_200801251420_0007_m_07_0 failed to report status for 602
seconds.  Killing!

AND SO ON...

THEN ALL THE TAKED GETS KILLED 

BUT STILL DATANODE IS ALIVE... I HAVE BEEN RUNNING HADOOP IN VMWARE AND ON
512MB RAM MACHINE..


SO PLZ HELP ME IN SOLVING THIS PROBLEM...

THANKS IN ADVANCE.

REGARDS,
ASHISH