Hi dennis,
         Ur's intitution was spot on!!!!!!!!!!!!!!!!!!!!!!!!!! After we
changed to fully qualified domain name it's working perfect........Thanks
alot for your help!!!!!!!!!!!!!! We require more help from u in coming
days!!!!!!!!!!!!!

THANK YOU!!!!!



Dennis Kubes wrote:
> 
> I may be wrong but my intuition keeps telling me it is a dns issue. 
> Below you have websearch1.  Have you tried fully qualified domain names 
> in both the hadoop-site.xml and slaves files.  For example 
> websearch1.internal.com:9000 where internal is your internal domain.  I 
> have seen some computers have a hard time connecting when not using 
> fully qualified names.  Try that and let me know the result.
> 
> The other thing I see, although I don't think it would be causing any 
> problems, is on some of your properties below there are spaces in the 
> names.  I would remove the spaces.
> 
> Dennis
> 
> srinath wrote:
>> Hi Dennis,
>>       
>>                  Yes i can ping from slavenode (there is no problems with
>> network connections and i have successfully cofingured Nutch0.8.1 and
>> crawled some pages without any issues) ...and below i'm attaching the
>> configurations which i'm using in hadoop-site.xml file and name of
>> systems i
>> mentioned in slaves file
>> 
>> The content of hadoop-site.xml is as follows : 
>> 
>> <configuration> 
>> 
>> <property> 
>>   <name>fs.default.name</name> 
>>   <value>websearch1:9000</value> 
>> </property> 
>> 
>> <property> 
>>   <name>mapred.job.tracker</name> 
>>   <value>websearch1:9001</value> 
>> </property> 
>> 
>> <property> 
>>   <name>mapred.tasktracker.tasks.maximum </name> 
>>   <value>20</value> 
>> </property> 
>> 
>> <property> 
>>   <name>mapred.map.tasks</name> 
>>   <value>60</value> 
>> </property> 
>> 
>> <property> 
>>   <name>mapred.reduce.tasks</name> 
>>   <value>6</value> 
>> </property> 
>> 
>> <property> 
>>   <name>dfs.name.dir</name> 
>>   <value>/data/nutch/nutch0.9/filesystem0.9/name</value> 
>> </property> 
>> 
>> <property> 
>>   <name>dfs.data.dir</name> 
>>   <value>/data/nutch/nutch0.9/filesystem0.9/data</value> 
>> </property> 
>> 
>> <property> 
>>   <name> dfs.replication</name> 
>>   <value>2</value> 
>> </property> 
>> </configuration> 
>> 
>> and slaves.txt
>> 
>> websearch1
>> websearch2
>>  
>>                 on Websearch1 datanode is starting without any
>> problems..but
>> on websearch3 it says it's started task tracker and datanode but they
>> both
>> can't send heartbeat to masternode(i mean they can't able to register
>> themselves with masternode ) ..... i'm continuosuly getting the same
>> error
>> as metioned before...
>>               One more thing i tried yesterday ...was to change the
>> hadoop
>> version... to 0.5 from 0.9.1 then namenode and datanode started
>> successfully
>> but ...jobtracker failed to start because nutch version(0.9.1) is using
>> some
>> classes which are not available on 0.5.....and i tried 0.7.1 and 0.8.2
>> ....same problems....and also i tried with 0.9.2 version i can't succeed
>> ..then i feel there is something to do with configurations?   
>> 
>> 
>> Dennis Kubes wrote:
>>> Can you ping the master computer (name node) from the slave (data node) 
>>> computers.  Also is your namenode configuration fs.default.name variable 
>>> pointing to 127.0.0.1 or is it pointing to the fully qualified domain 
>>> name of the master computer?
>>>
>>> Dennis Kubes
>>>
>>> srinath wrote:
>>>> Hi,
>>>>    Thx For Your Reply .. But namenode was started successfully on
>>>> masternode
>>>> .... and datanode where we started on the machine where masternode  is
>>>> running is able to connect but datanode on the other machine is not
>>>> able
>>>> to
>>>> connect back!!!!!!!!!! if u like to see i will post configuration
>>>> params
>>>> what we set????
>>>>
>>>>
>>>> Dennis Kubes wrote:
>>>>> I would take a look at the processes on the namenode server and see if 
>>>>> the namenode has started up.  It doesn't look like it did.  If this is
>>>>> a 
>>>>>   new install, did you format the namenode?
>>>>>
>>>>> Dennis
>>>>>
>>>>> srinath wrote:
>>>>>> Hi,
>>>>>>     While starting hadoop process we are getting the following error
>>>>>> in
>>>>>> logs
>>>>>> tasktracker in datanode is not able to connect back to jobtracker
>>>>>> (but
>>>>>> jobtracker on the other machine started successfully and listening on
>>>>>> port
>>>>>> 9001) ... i'm using Nutch0.9.1 version and Hadoop0.9.1 ..
>>>>>>
>>>>>> 2007-01-04 23:57:35,559 INFO  ipc.Server - IPC Server handler 17 on
>>>>>> 50050:
>>>>>> starting
>>>>>> 2007-01-04 23:57:35,559 INFO  ipc.Server - IPC Server handler 18 on
>>>>>> 50050:
>>>>>> starting
>>>>>> 2007-01-04 23:57:35,559 INFO  mapred.TaskTracker - Starting tracker
>>>>>> tracker_websearch3:50050
>>>>>> 2007-01-04 23:57:35,559 INFO  ipc.Server - IPC Server handler 19 on
>>>>>> 50050:
>>>>>> starting
>>>>>> 2007-01-04 23:57:35,566 INFO  ipc.Client -
>>>>>> org.apache.hadoop.io.ObjectWritableConnection culler maxidletime=
>>>>>> 1000ms
>>>>>> 2007-01-04 23:57:35,567 INFO  ipc.Client -
>>>>>> org.apache.hadoop.io.ObjectWritable Connection Culler: starting
>>>>>> 2007-01-04 23:57:35,589 INFO  ipc.Client - Retrying connect to
>>>>>> server:
>>>>>> websearch1/10.50.12.220:9001. Already tried 1 time(s).
>>>>>> 2007-01-04 23:57:36,590 INFO  ipc.Client - Retrying connect to
>>>>>> server:
>>>>>> websearch1/10.50.12.220:9001. Already tried 2 time(s).
>>>>>> 2007-01-04 23:57:37,600 INFO  ipc.Client - Retrying connect to
>>>>>> server:
>>>>>> websearch1/10.50.12.220:9001. Already tried 3 time(s).
>>>>>> 2007-01-04 23:57:38,610 INFO  ipc.Client - Retrying connect to
>>>>>> server:
>>>>>> websearch1/10.50.12.220:9001. Already tried 4 time(s).
>>>>>> 2007-01-04 23:57:39,620 INFO  ipc.Client - Retrying connect to
>>>>>> server:
>>>>>> websearch1/10.50.12.220:9001. Already tried 5 time(s).
>>>>>> 2007-01-04 23:57:40,630 INFO  ipc.Client - Retrying connect to
>>>>>> server:
>>>>>> websearch1/10.50.12.220:9001. Already tried 6 time(s).
>>>>>> 2007-01-04 23:57:41,640 INFO  ipc.Client - Retrying connect to
>>>>>> server:
>>>>>> websearch1/10.50.12.220:9001. Already tried 7 time(s).
>>>>>> 2007-01-04 23:57:42,650 INFO  ipc.Client - Retrying connect to
>>>>>> server:
>>>>>> websearch1/10.50.12.220:9001. Already tried 8 time(s).
>>>>>> 2007-01-04 23:57:43,660 INFO  ipc.Client - Retrying connect to
>>>>>> server:
>>>>>> websearch1/10.50.12.220:9001. Already tried 9 time(s).
>>>>>> 2007-01-04 23:57:44,670 INFO  ipc.Client - Retrying connect to
>>>>>> server:
>>>>>> websearch1/10.50.12.220:9001. Already tried 10 time(s).
>>>>>> 2007-01-04 23:57:45,680 INFO  ipc.RPC - Server at
>>>>>> websearch1/10.50.12.220:9001 not available yet, Zzzzz...
>>>>>> 2007-01-04 23:57:46,690 INFO  ipc.Client - Retrying connect to
>>>>>> server:
>>>>>> websearch1/10.50.12.220:9001. Already tried 1 time(s).
>>>>>> 2007-01-04 23:57:47,700 INFO  ipc.Client - Retrying connect to
>>>>>> server:
>>>>>> websearch1/10.50.12.220:9001. Already tried 2 time(s).
>>>>>> 2007-01-04 23:57:48,710 INFO  ipc.Client - Retrying connect to
>>>>>> server:
>>>>>> websearch1/10.50.12.220:9001. Already tried 3 time(s).
>>>>>> 2007-01-04 23:57:49,720 INFO  ipc.Client - Retrying connect to
>>>>>> server:
>>>>>> websearch1/10.50.12.220:9001. Already tried 4 time(s).
>>>>>>
>>>>>>
>>>>>> can any one help? regarding this????? does something to do with
>>>>>> hadoop
>>>>>> configuration?
>>>
>> 
> 
> 

-- 
View this message in context: 
http://www.nabble.com/Issues-Starting-Hadoop-Process-in-Nutch0.9l.1-tf2920348.html#a8213165
Sent from the Nutch - User mailing list archive at Nabble.com.


-------------------------------------------------------------------------
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT & business topics through brief surveys - and earn cash
http://www.techsay.com/default.php?page=join.php&p=sourceforge&CID=DEVDEV
_______________________________________________
Nutch-general mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/nutch-general

Reply via email to