Can Hadoop handle many appliations

2009-06-20 Thread ashish pareek
Hello Everybody,

   I have a few questions regarding running applications
on Hadoop.

1) Can a single Hadoop cluster simultaneously handle applications of same
type ( or of different type).

2) What are the challenges in running applications of different types in
Hadoop. How Does it differ from running applications of same type.


Thanking you,

Regards,
Ashish.


Re: Hadoop error help- file system closed, could only be replicated to 0 nodes, instead of 1

2009-06-18 Thread ashish pareek
HI ,
 What seems from your details is that datanode is not running.can
you run *bin/hadoop dfsadmin -report*  and find out whether your datanodes
are up ? then post your observation and it would be better if you post even
your hadoop-site.xml file deatils also.

Regards,
Ashish.

On Fri, Jun 19, 2009 at 3:16 AM,  wrote:

> Hi,
>
> I am extremely new to Hadoop and have come across a few errors that I'm not
> sure how to fix. I am running Hadoop version 0.19.0 from an image through
> Elasticfox and S3. I am on windows and use puTTY as my ssh. I am trying to
> run a wordcount with 5 slaves. This is what I do so far:
>
> 1. boot up the instance through ElasticFox
> 2. cd /usr/local/hadoop-0.19.0
> 3. bin/hadoop namenode -format
> 4. bin/start-all.sh
> 5. jps --( shows jps, jobtracker, secondarynamenode)
> 6.bin/stop-all.sh
> 7. ant examples
> 8. bin/start-all.sh
> 9. bin/hadoop jar build/hadoop-0.19.0-examples.jar pi 0 100
>
> Then I get this error trace:
>
> Number of Maps = 0 Samples per Map = 100
> Starting Job
> 09/06/18 17:31:25 INFO hdfs.DFSClient:
> org.apache.hadoop.ipc.RemoteException: java.io.IOException: File
> /mnt/hadoop/mapred/system/job_200906181730_0001/job.jar could only be
> replicated to 0 nodes, instead of 1
>at
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1270)
>at
> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:351)
>at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>at java.lang.reflect.Method.invoke(Method.java:597)
>at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:452)
>at org.apache.hadoop.ipc.Server$Handler.run(Server.java:892)
>
>at org.apache.hadoop.ipc.Client.call(Client.java:696)
>at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:216)
>at $Proxy0.addBlock(Unknown Source)
>at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>at java.lang.reflect.Method.invoke(Method.java:597)
>at
> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
>at
> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
>at $Proxy0.addBlock(Unknown Source)
>at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.locateFollowingBlock(DFSClient.java:2815)
>at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2697)
>at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:1997)
>at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2183)
>
> 09/06/18 17:31:25 WARN hdfs.DFSClient: NotReplicatedYetException sleeping
> /mnt/hadoop/mapred/system/job_200906181730_0001/job.jar retries left 4
> 09/06/18 17:31:25 INFO hdfs.DFSClient:
> org.apache.hadoop.ipc.RemoteException: java.io.IOException: File
> /mnt/hadoop/mapred/system/job_200906181730_0001/job.jar could only be
> replicated to 0 nodes, instead of 1
>at
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1270)
>at
> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:351)
>at sun.reflec,t.NativeMethodAccessorImpl.invoke0(Native Method)
>at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>at java.lang.reflect.Method.invoke(Method.java:597)
>at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:452)
>at org.apache.hadoop.ipc.Server$Handler.run(Server.java:892)
>
>at org.apache.hadoop.ipc.Client.call(Client.java:696)
>at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:216)
>at $Proxy0.addBlock(Unknown Source)
>at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>at java.lang.reflect.Method.invoke(Method.java:597)
>at
> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
>at
> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
>at $Proxy0.addBlock(Unknown Source)
>at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.locateFollowingBlock(DFSClient.java:2815)
>at
> org.apac

Re: Need help

2009-06-18 Thread ashish pareek
Hello Everybody,

  How can we handle different applications having
different requirement being run on the same hadoop cluster ? What are the
various approaches to solve such problem.. if possible please mention some
of those ideas.

Does such implementation exists ?

Thanks ,

Regards,
Ashish

On Thu, Jun 18, 2009 at 9:36 PM, jason hadoop wrote:

> For me, I like to have one configuration file that I distribute to all of
> the machines in my cluster via rsync.
>
> In there are things like the number of tasks per node to run, and where to
> store dfs data and local temporary data, and the limits to storage for the
> machines.
>
> If the machines are very different, it becomes important to tailor the
> configuration file per machine or type of machine.
>
> At this point, you are pretty much going to have to spend the time, reading
> through the details of configuring a hadoop cluster.
>
>
> On Thu, Jun 18, 2009 at 8:33 AM, ashish pareek 
> wrote:
>
> > Can you tell few of the challenges in configuring heterogeneous
> > cluster...or
> > can pass on some link where I would get some information regarding
> > challenges in running Hadoop on heterogeneous hardware
> >
> > One more things is How about running different applications on the same
> > Hadoop cluster?and what challenges are involved in it ?
> >
> > Thanks,
> > Regards,
> > Ashish
> >
> >
> > On Thu, Jun 18, 2009 at 8:53 PM, jason hadoop  > >wrote:
> >
> > > I don't know anyone who has a completely homogeneous cluster.
> > >
> > > So hadoop is scalable across heterogeneous environments.
> > >
> > > I stated that configuration is simpler if the machines are similar
> (There
> > > are optimizations in configuration for near homogeneous machines.)
> > >
> > > On Thu, Jun 18, 2009 at 8:10 AM, ashish pareek 
> > > wrote:
> > >
> > > > Does that mean hadoop is not scalable wrt heterogeneous environment?
> > and
> > > > one
> > > > more question is can we run different application on the same hadoop
> > > > cluster
> > > > .
> > > >
> > > > Thanks.
> > > > Regards,
> > > > Ashish
> > > >
> > > > On Thu, Jun 18, 2009 at 8:30 PM, jason hadoop <
> jason.had...@gmail.com
> > > > >wrote:
> > > >
> > > > > Hadoop has always been reasonably agnostic wrt hardware and
> > > homogeneity.
> > > > > There are optimizations in configuration for near homogeneous
> > machines.
> > > > >
> > > > >
> > > > >
> > > > > On Thu, Jun 18, 2009 at 7:46 AM, ashish pareek <
> pareek...@gmail.com>
> > > > > wrote:
> > > > >
> > > > > > Hello,
> > > > > >I am doing my master my final year project is on
> Hadoop
> > > > ...so
> > > > > I
> > > > > > would like to know some thing about Hadoop cluster i.e, Do new
> > > version
> > > > of
> > > > > > Hadoop are able to handle heterogeneous hardware.If you have any
> > > > > > informantion regarding these please mail me as my project is in
> > > > > > heterogenous
> > > > > > environment.
> > > > > >
> > > > > >
> > > > > > Thanks!
> > > > > >
> > > > > > Reagrds,
> > > > > > Ashish Pareek
> > > > > >
> > > > >
> > > > >
> > > > >
> > > > > --
> > > > > Pro Hadoop, a book to guide you from beginner to hadoop mastery,
> > > > > http://www.amazon.com/dp/1430219424?tag=jewlerymall
> > > > > www.prohadoopbook.com a community for Hadoop Professionals
> > > > >
> > > >
> > >
> > >
> > >
> > > --
> > > Pro Hadoop, a book to guide you from beginner to hadoop mastery,
> > > http://www.amazon.com/dp/1430219424?tag=jewlerymall
> > > www.prohadoopbook.com a community for Hadoop Professionals
> > >
> >
>
>
>
> --
> Pro Hadoop, a book to guide you from beginner to hadoop mastery,
> http://www.amazon.com/dp/1430219424?tag=jewlerymall
> www.prohadoopbook.com a community for Hadoop Professionals
>


Re: Need help

2009-06-18 Thread ashish pareek
Can you tell few of the challenges in configuring heterogeneous cluster...or
can pass on some link where I would get some information regarding
challenges in running Hadoop on heterogeneous hardware

One more things is How about running different applications on the same
Hadoop cluster?and what challenges are involved in it ?

Thanks,
Regards,
Ashish


On Thu, Jun 18, 2009 at 8:53 PM, jason hadoop wrote:

> I don't know anyone who has a completely homogeneous cluster.
>
> So hadoop is scalable across heterogeneous environments.
>
> I stated that configuration is simpler if the machines are similar (There
> are optimizations in configuration for near homogeneous machines.)
>
> On Thu, Jun 18, 2009 at 8:10 AM, ashish pareek 
> wrote:
>
> > Does that mean hadoop is not scalable wrt heterogeneous environment? and
> > one
> > more question is can we run different application on the same hadoop
> > cluster
> > .
> >
> > Thanks.
> > Regards,
> > Ashish
> >
> > On Thu, Jun 18, 2009 at 8:30 PM, jason hadoop  > >wrote:
> >
> > > Hadoop has always been reasonably agnostic wrt hardware and
> homogeneity.
> > > There are optimizations in configuration for near homogeneous machines.
> > >
> > >
> > >
> > > On Thu, Jun 18, 2009 at 7:46 AM, ashish pareek 
> > > wrote:
> > >
> > > > Hello,
> > > >I am doing my master my final year project is on Hadoop
> > ...so
> > > I
> > > > would like to know some thing about Hadoop cluster i.e, Do new
> version
> > of
> > > > Hadoop are able to handle heterogeneous hardware.If you have any
> > > > informantion regarding these please mail me as my project is in
> > > > heterogenous
> > > > environment.
> > > >
> > > >
> > > > Thanks!
> > > >
> > > > Reagrds,
> > > > Ashish Pareek
> > > >
> > >
> > >
> > >
> > > --
> > > Pro Hadoop, a book to guide you from beginner to hadoop mastery,
> > > http://www.amazon.com/dp/1430219424?tag=jewlerymall
> > > www.prohadoopbook.com a community for Hadoop Professionals
> > >
> >
>
>
>
> --
> Pro Hadoop, a book to guide you from beginner to hadoop mastery,
> http://www.amazon.com/dp/1430219424?tag=jewlerymall
> www.prohadoopbook.com a community for Hadoop Professionals
>


Re: Need help

2009-06-18 Thread ashish pareek
Does that mean hadoop is not scalable wrt heterogeneous environment? and one
more question is can we run different application on the same hadoop cluster
.

Thanks.
Regards,
Ashish

On Thu, Jun 18, 2009 at 8:30 PM, jason hadoop wrote:

> Hadoop has always been reasonably agnostic wrt hardware and homogeneity.
> There are optimizations in configuration for near homogeneous machines.
>
>
>
> On Thu, Jun 18, 2009 at 7:46 AM, ashish pareek 
> wrote:
>
> > Hello,
> >I am doing my master my final year project is on Hadoop ...so
> I
> > would like to know some thing about Hadoop cluster i.e, Do new version of
> > Hadoop are able to handle heterogeneous hardware.If you have any
> > informantion regarding these please mail me as my project is in
> > heterogenous
> > environment.
> >
> >
> > Thanks!
> >
> > Reagrds,
> > Ashish Pareek
> >
>
>
>
> --
> Pro Hadoop, a book to guide you from beginner to hadoop mastery,
> http://www.amazon.com/dp/1430219424?tag=jewlerymall
> www.prohadoopbook.com a community for Hadoop Professionals
>


Fwd: Need help

2009-06-18 Thread ashish pareek
Hello,
I am doing my master my final year project is on Hadoop ...so I
would like to know some thing about Hadoop cluster i.e, Do new version of
Hadoop are able to handle heterogeneous hardware.If you have any
informantion regarding these please mail me as my project is in heterogenous
environment.


Thanks!

Reagrds,
Ashish Pareek


Problem in viewing WEB UI

2009-06-16 Thread ashish pareek
Hi,

  When I run command *bin/hadoop dfsadmin -report *it shows that 2
datanodes are alive but when i try to http://hadoopmster:50070/ but the
problem is that it opens doesnot opne
http://hadoopmaster:50070/dfshealth.jsp page and throws *error HTTP: 404 .
So why is't happening like this?
*
Regards,
Ashish Pareek


 On Wed, Jun 17, 2009 at 10:06 AM, Sugandha Neaolekar <
sugandha@gmail.com> wrote:

> Well, You just have to specify the address in the URL address bar as::
> http://hadoopmaster:50070 U'll be able to see the web UI..!
>
>
> On Tue, Jun 16, 2009 at 7:17 PM, ashish pareek wrote:
>
>> HI Sugandha,
>>Hmmm your suggestion helped and Now I am able
>> to run two datanode one on the same machine as name node and other on
>> the different machine Thanks a lot :)
>>
>>  But the problem is now I am not able to see web UI .
>> for  both datanode and as well as name node
>> should I have to consider some more things in the site.xml ? if so please
>> help...
>>
>> Thanking you again,
>> regards,
>> Ashish Pareek.
>>
>> On Tue, Jun 16, 2009 at 3:10 PM, Sugandha Naolekar <
>> sugandha@gmail.com> wrote:
>>
>>> hi,,!
>>>
>>>
>>> First of all, get your concepts clear of hadoop.
>>> You can refer to the following
>>>
>>> site::
>>> http://www.google.co.in/url?sa=t&source=web&ct=res&cd=1&url=http%3A%2F%2Fwww.michael-noll.com%2Fwiki%2FRunning_Hadoop_On_Ubuntu_Linux_(Single-Node_Cluster)&ei=lGU3Spv2FZbLjAe19KmiDQ&usg=AFQjCNFbmVGsoChOSMzCB3tRhoV0ylHOzA&sig2=t2AJ_nf24SFtveN4PHS_TA<http://www.google.co.in/url?sa=t&source=web&ct=res&cd=1&url=http%3A%2F%2Fwww.michael-noll.com%2Fwiki%2FRunning_Hadoop_On_Ubuntu_Linux_%28Single-Node_Cluster%29&ei=lGU3Spv2FZbLjAe19KmiDQ&usg=AFQjCNFbmVGsoChOSMzCB3tRhoV0ylHOzA&sig2=t2AJ_nf24SFtveN4PHS_TA>
>>>
>>>
>>> I have small doubt whether in the mater.xml and slave.xml we can have
>>> same port numbers to both of them like
>>>
>>>
>>> for slave :
>>>
>>> 
>>> fs.default.name
>>> hdfs://hadoopslave:
>>>>
>>>> 9000
>>>>   
>>>>
>>>>
>>>>  for master:::
>>>>
>>>> 
>>>> fs.default.name
>>>> hdfs://hadoopmaster:9000
>>>>   
>>>>
>>>>
>>>
>>> Well, any  two daemons or services can run on the same port unless, they
>>> are not run on the same machine.If you wish to run DN and NN on the same
>>> machine, their port numbers have to be different.
>>>
>>>
>>>
>>>
>>> On Tue, Jun 16, 2009 at 2:55 PM, ashish pareek wrote:
>>>
>>>> HI sugandha,
>>>>
>>>>
>>>>
>>>> and one more thing can we have in slave:::
>>>>
>>>> 
>>>>   dfs.datanode.address>>>>
>>>>> name>
>>>>>   hadoopmaster:9000
>>>>> hadoopslave:9001
>>>>>   
>>>>>
>>>>
>>>
>>> Also, fs,default.name is the tag which specifies the default filesystem.
>>> And generaLLY, it is run on namenode. So, it;s value has to be a namenode's
>>> address only and not slave's.
>>>
>>>
>>>>
>>>> Else if you have complete procedure for installing and running Hadoop in
>>>> cluster can you please send it to me .. I need to step up hadoop with 
>>>> in
>>>> two days and show it to my guide.Currently I am doing my masters.
>>>>
>>>> Thanks for your spending time
>>>
>>>
>>> Try for the above, and this should work!
>>>
>>>>
>>>>
>>>> regards,
>>>> Ashish Pareek
>>>>
>>>>
>>>> On Tue, Jun 16, 2009 at 2:33 PM, Sugandha Naolekar <
>>>> sugandha@gmail.com> wrote:
>>>>
>>>>> Following changes are to be done::
>>>>>
>>>>> Under master folder::
>>>>>
>>>>> -> put slaves address as well under the values of
>>>>> tag(dfs.datanode.address)
>>>>>
>>>>> -> You want to make namenode as datanode as well. As per your config
>>>>> file, you have specified hadoopmaster  in your slave file. If you don't 
>>>>> want
>>

org.apache.hadoop.ipc.client : trying connect to server failed

2009-06-15 Thread ashish pareek
HI ,
 I am trying to step up a hadoop cluster on 3GB machine and using hadoop
0.18.3 and  have followed procedure given in  apache hadoop site for hadoop
cluster.
 In conf/slaves I have added two datanode i.e including the namenode
vitrual machine and other machine virtual machine (datanode)  . and have
set up passwordless ssh between both virtual machines . But now problem
is when I run command :

bin/hadoop start-all.sh

It start only one datanode on the same namenode vitrual machine but it
doesn't start the datanode on other machine.

in logs/hadoop-datanode.log  i get message


 INFO org.apache.hadoop.ipc.Client: Retrying
 connect to server: hadoop1/192.168.1.28:9000. Already

  tried 1 time(s).

 2009-05-09 18:35:14,266 INFO org.apache.hadoop.ipc.Client: Retrying
 connect to server: hadoop1/192.168.1.28:9000. Already tried 2 time(s).

 2009-05-09 18:35:14,266 INFO org.apache.hadoop.ipc.Client: Retrying
 connect to server: hadoop1/192.168.1.28:9000. Already tried 3 time(s).


.
.
.
.
.
.
.
.
.

.
.

.

I have tried formatting and start the cluster again .but still I
get the same error.

So can any one help in solving this problem. :)

Thanks

Regards

Ashish Pareek


Re: org.apache.hadoop.ipc.client : trying connect to server failed

2009-05-28 Thread ashish pareek
Yes I am able to ping and ssh between two virtual machine and even i
have set ip address of both the virtual machines in their respective
/etc/hosts file ...

thanx for reply .. if you suggest some other thing which i could
have missed or any remedy 

Regards,
Ashish Pareek.


On Fri, May 29, 2009 at 10:04 AM, Pankil Doshi  wrote:

> make sure u can ping that data node and ssh it.
>
>
> On Thu, May 28, 2009 at 12:02 PM, ashish pareek 
> wrote:
>
> > HI ,
> > I am trying to step up a hadoop cluster on 512 MB machine and using
> > hadoop 0.18 and  have followed procedure given in  apache hadoop site for
> > hadoop cluster.
> > I included  in conf/slaves two datanode i.e including the namenode
> > vitrual machine and other machine virtual machine  . and have set up
> > passwordless ssh between both virtual machines. But now problem is
> when
> > is run command >>
> >
> > bin/hadoop start-all.sh
> >
> > It start only one datanode on the same namenode vitrual machine but it
> > doesn't start the datanode on other machine.
> >
> > in logs/hadoop-datanode  i get message
> >
> >
> >  INFO org.apache.hadoop.ipc.Client: Retrying
> >  connect to server: hadoop1/192.168.1.28:9000. Already
> >  tried 1 time(s).
> >  2009-05-09 18:35:14,266 INFO org.apache.hadoop.ipc.Client: Retrying
> >  connect to server: hadoop1/192.168.1.28:9000. Already tried 2 time(s).
> >  2009-05-09 18:35:14,266 INFO org.apache.hadoop.ipc.Client: Retrying
> >  connect to server: hadoop1/192.168.1.28:9000. Already tried 3 time(s).
> >
> > .
> > .
> > .
> > .
> > .
> > .
> > .
> > .
> > .
> >
> > .
> > .
> >
> > .
> >
> >
> > So can any one help in solving this problem. :)
> >
> > Thanks
> >
> > Regards
> > Ashish Pareek
> >
>


Re: org.apache.hadoop.ipc.client : trying connect to server failed

2009-05-28 Thread ashish pareek
Hi some help me out 



On Thu, May 28, 2009 at 10:32 PM, ashish pareek  wrote:

> HI ,
>  I am trying to step up a hadoop cluster on 512 MB machine and using
> hadoop 0.18 and  have followed procedure given in  apache hadoop site for
> hadoop cluster.
>  I included  in conf/slaves two datanode i.e including the namenode
> vitrual machine and other machine virtual machine  . and have set up
> passwordless ssh between both virtual machines. But now problem is when
> is run command >>
>
> bin/hadoop start-all.sh
>
> It start only one datanode on the same namenode vitrual machine but it
> doesn't start the datanode on other machine.
>
> in logs/hadoop-datanode  i get message
>
>
>  INFO org.apache.hadoop.ipc.Client: Retrying
>  connect to server: hadoop1/192.168.1.28:9000. Already
>   tried 1 time(s).
>
>  2009-05-09 18:35:14,266 INFO org.apache.hadoop.ipc.Client: Retrying
>  connect to server: hadoop1/192.168.1.28:9000. Already tried 2 time(s).
>  2009-05-09 18:35:14,266 INFO org.apache.hadoop.ipc.Client: Retrying
>  connect to server: hadoop1/192.168.1.28:9000. Already tried 3 time(s).
>
> .
> .
> .
> .
> .
> .
> .
> .
> .
>
> .
> .
>
> .
>
>
> So can any one help in solving this problem. :)
>
> Thanks
>
> Regards
>
> Ashish Pareek
>
>
>


org.apache.hadoop.ipc.client : trying connect to server failed

2009-05-28 Thread ashish pareek
HI ,
 I am trying to step up a hadoop cluster on 512 MB machine and using
hadoop 0.18 and  have followed procedure given in  apache hadoop site for
hadoop cluster.
 I included  in conf/slaves two datanode i.e including the namenode
vitrual machine and other machine virtual machine  . and have set up
passwordless ssh between both virtual machines. But now problem is when
is run command >>

bin/hadoop start-all.sh

It start only one datanode on the same namenode vitrual machine but it
doesn't start the datanode on other machine.

in logs/hadoop-datanode  i get message


 INFO org.apache.hadoop.ipc.Client: Retrying
 connect to server: hadoop1/192.168.1.28:9000. Already
  tried 1 time(s).
 2009-05-09 18:35:14,266 INFO org.apache.hadoop.ipc.Client: Retrying
 connect to server: hadoop1/192.168.1.28:9000. Already tried 2 time(s).
 2009-05-09 18:35:14,266 INFO org.apache.hadoop.ipc.Client: Retrying
 connect to server: hadoop1/192.168.1.28:9000. Already tried 3 time(s).

.
.
.
.
.
.
.
.
.

.
.

.


So can any one help in solving this problem. :)

Thanks

Regards
Ashish Pareek


Re: Could only be replicated to 0 nodes, instead of 1

2009-05-21 Thread ashish pareek
Hi ,

I have two suggestion

i)Choose a right version ( Hadoop- 0.18 is good)
ii)replication should be 3 as ur having 3 modes.( Indirectly see to it that
ur configuration is correct !!)

Hey even i am just suggesting this as i am also a new to hadoop

Ashish Pareek


On Thu, May 21, 2009 at 2:41 PM, Stas Oskin  wrote:

> Hi.
>
> I'm testing Hadoop in our lab, and started getting the following message
> when trying to copy a file:
> Could only be replicated to 0 nodes, instead of 1
>
> I have the following setup:
>
> * 3 machines, 2 of them with only 80GB of space, and 1 with 1.5GB
> * Two clients are copying files all the time (one of them is the 1.5GB
> machine)
> * The replication is set on 2
> * I let the space on 2 smaller machines to end, to test the behavior
>
> Now, one of the clients (the one located on 1.5GB) works fine, and the
> other
> one - the external, unable to copy and displays the error + the exception
> below
>
> Any idea if this expected on my scenario? Or how it can be solved?
>
> Thanks in advance.
>
>
>
> 09/05/21 10:51:03 WARN dfs.DFSClient: NotReplicatedYetException sleeping
> /test/test.bin retries left 1
>
> 09/05/21 10:51:06 WARN dfs.DFSClient: DataStreamer Exception:
> org.apache.hadoop.ipc.RemoteException: java.io.IOException: File
> /test/test.bin could only be replicated to 0 nodes, instead of 1
>
>at
>
> org.apache.hadoop.dfs.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1123
> )
>
>at org.apache.hadoop.dfs.NameNode.addBlock(NameNode.java:330)
>
>at sun.reflect.GeneratedMethodAccessor8.invoke(Unknown Source)
>
>at
>
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25
> )
>
>at java.lang.reflect.Method.invoke(Method.java:597)
>
>at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:481)
>
>at org.apache.hadoop.ipc.Server$Handler.run(Server.java:890)
>
>
>
>at org.apache.hadoop.ipc.Client.call(Client.java:716)
>
>at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:216)
>
>at org.apache.hadoop.dfs.$Proxy0.addBlock(Unknown Source)
>
>at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>
>at
>
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39
> )
>
>at
>
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25
> )
>
>at java.lang.reflect.Method.invoke(Method.java:597)
>
>at
>
> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82
> )
>
>at
>
> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59
> )
>
>at org.apache.hadoop.dfs.$Proxy0.addBlock(Unknown Source)
>
>at
>
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.locateFollowingBlock(DFSClient.java:2450
> )
>
>at
>
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2333
> )
>
>at
>
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.access$1800(DFSClient.java:1745
> )
>
>at
>
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:1922
> )
>
>
>
> 09/05/21 10:51:06 WARN dfs.DFSClient: Error Recovery for block null bad
> datanode[0]
>
> java.io.IOException: Could not get block locations. Aborting...
>
>at
>
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.processDatanodeError(DFSClient.java:2153
> )
>
>at
>
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.access$1400(DFSClient.java:1745
> )
>
>at
>
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:1899
> )
>


TASKS KILLED WHEN RUNNING : bin/hadoop jar hadoop-*-examples.jar grep input output 'dfs[a-z.]+'

2009-05-16 Thread ashish pareek
hELLO TO EVERY BODY
   I AM A NEW HAOODP USER I STARTED RUNNING HADOOP USING SITE
http://hadoop.apache.org/core/docs/current/quickstart.html BUT WHEN I RUN
COMMAND  " bin/hadoop jar hadoop-*-examples.jar grep input output
'dfs[a-z.]+'  ", IN PSEDUO DISTRIBUTED MODE I GET ERROR LIKE :::

Task  task_200801251420_0007_m_06_0 failed to report status for 601
seconds.  Killing!

Task  task_200801251420_0007_m_07_0 failed to report status for 602
seconds.  Killing!

AND SO ON...

THEN ALL THE TAKED GETS KILLED 

BUT STILL DATANODE IS ALIVE... I HAVE BEEN RUNNING HADOOP IN VMWARE AND ON
512MB RAM MACHINE..


SO PLZ HELP ME IN SOLVING THIS PROBLEM...

THANKS IN ADVANCE.

REGARDS,
ASHISH