Ok thanks will do

Sent from my iPhone

> On 22-Apr-2015, at 2:39 pm, sandeep vura <sandeepv...@gmail.com> wrote:
> 
> hosts file will be available in /etc directory please check once.
> 
>> On Wed, Apr 22, 2015 at 2:36 PM, Anand Murali <anand_vi...@yahoo.com> wrote:
>> I don't seem to have etc/host
>> 
>> 
>> Sent from my iPhone
>> 
>>> On 22-Apr-2015, at 2:30 pm, sandeep vura <sandeepv...@gmail.com> wrote:
>>> 
>>> Hi Anand,
>>> 
>>> comment the ip address - 127.0.1.1 in /etc/hosts
>>> add the following ip address - 127.0.0.1  localhost  in /etc/hosts.
>>> 
>>> Restart your hadoop cluster after made changes in /etc/hosts
>>> 
>>> Regards,
>>> Sandeep.v
>>> 
>>>> On Wed, Apr 22, 2015 at 2:16 PM, Anand Murali <anand_vi...@yahoo.com> 
>>>> wrote:
>>>> Dear All:
>>>> 
>>>> Has anyone encountered this error and if so how have you fixed it other 
>>>> then re-installing Hadoop or re-starting start-dfs.sh when you have 
>>>> already started after boot. Find below
>>>> 
>>>> anand_vihar@Latitude-E5540:~$ ssh localhost
>>>> Welcome to Ubuntu 14.10 (GNU/Linux 3.16.0-34-generic x86_64)
>>>> 
>>>>  * Documentation:  https://help.ubuntu.com/
>>>> 
>>>> 1 package can be updated.
>>>> 1 update is a security update.
>>>> 
>>>> Last login: Wed Apr 22 13:33:26 2015 from localhost
>>>> anand_vihar@Latitude-E5540:~$ cd hadoop-2.6.0
>>>> anand_vihar@Latitude-E5540:~/hadoop-2.6.0$ . .hadoop
>>>> /home/anand_vihar/hadoop-2.6.0
>>>> /home/anand_vihar/jdk1.7.0_75
>>>> /home/anand_vihar/hadoop-2.6.0
>>>> anand_vihar@Latitude-E5540:~/hadoop-2.6.0$ hadoop version
>>>> Hadoop 2.6.0
>>>> Subversion https://git-wip-us.apache.org/repos/asf/hadoop.git -r 
>>>> e3496499ecb8d220fba99dc5ed4c99c8f9e33bb1
>>>> Compiled by jenkins on 2014-11-13T21:10Z
>>>> Compiled with protoc 2.5.0
>>>> From source with checksum 18e43357c8f927c0695f1e9522859d6a
>>>> This command was run using 
>>>> /home/anand_vihar/hadoop-2.6.0/share/hadoop/common/hadoop-common-2.6.0.jar
>>>> anand_vihar@Latitude-E5540:~/hadoop-2.6.0$ start-dfs.sh
>>>> Starting namenodes on [localhost]
>>>> localhost: starting namenode, logging to 
>>>> /home/anand_vihar/hadoop-2.6.0/logs/hadoop-anand_vihar-namenode-Latitude-E5540.out
>>>> localhost: starting datanode, logging to 
>>>> /home/anand_vihar/hadoop-2.6.0/logs/hadoop-anand_vihar-datanode-Latitude-E5540.out
>>>> Starting secondary namenodes [0.0.0.0]
>>>> 0.0.0.0: starting secondarynamenode, logging to 
>>>> /home/anand_vihar/hadoop-2.6.0/logs/hadoop-anand_vihar-secondarynamenode-Latitude-E5540.out
>>>> anand_vihar@Latitude-E5540:~/hadoop-2.6.0$ hdfs dfs -ls
>>>> ls: Call From Latitude-E5540/127.0.1.1 to localhost:9000 failed on 
>>>> connection exception: java.net.ConnectException: Connection refused; For 
>>>> more details see:  http://wiki.apache.org/hadoop/ConnectionRefused
>>>> anand_vihar@Latitude-E5540:~/hadoop-2.6.0$ 
>>>> 
>>>> 
>>>> 
>>>> I have checked http://wiki.apache.org/hadoop/ConnectionRefused but there 
>>>> is no fix to the problem rather it seems to be a Ubuntu network problem. I 
>>>> have many times killed nanenode/datanode/secondary data note, shutdown and 
>>>> restarted, but this error still appears. The only way seems to be 
>>>> re-installing hadoop. Please advise or refer.
>>>> 
>>>> Many thanks,
>>>> 
>>>> Regards,
>>>> 
>>>> 
>>>>  
>>>> Anand Murali  
>>>> 11/7, 'Anand Vihar', Kandasamy St, Mylapore
>>>> Chennai - 600 004, India
>>>> Ph: (044)- 28474593/ 43526162 (voicemail)
> 

Reply via email to