I am getting the following error. Does this mean the tar file is corrupted? Do 
i need to download it again? Please advise.
$ tar xzf hadoop-1.1.2.tar.gz
gzip: stdin: unexpected end of file
tar: Unexpected EOF in archive
tar: Unexpected EOF in archive
tar: Error is not recoverable: exiting now



________________________________
From: Mohammad Tariq <donta...@gmail.com>
To: "user@hadoop.apache.org" <user@hadoop.apache.org>; Raj Hadoop 
<hadoop...@yahoo.com> 
Sent: Thursday, May 16, 2013 12:02 PM
Subject: Re: Configuring SSH - is it required? for a psedo distriburted mode?



Hello Raj, 

     ssh is actually 2 things :
1- ssh : The command we use to connect to remote machines - the client. 
2- sshd : The daemon that is running on the server and allows clients to 
connect to the server.
ssh is pre-enabled on Linux, but in order to start sshd daemon, we need to 
install ssh first.

To start the Hadoop daemons you have to make ssh passwordless and issue 
bin/start-dfs.sh and bin/start-mapred.sh.

You might find this link useful.


Warm Regards, 
Tariq
cloudfront.blogspot.com



On Thu, May 16, 2013 at 9:26 PM, Raj Hadoop <hadoop...@yahoo.com> wrote:

Hi,
>
>I am a bit confused here. I am planning to run on a single machine.
>
>So what should i do to start hadoop processes. How should I do an SSH? Can you 
>please breifly explain me what SSH is?
>
>Thanks,
>Raj
>From: Jay Vyas <jayunit...@gmail.com>
>To: "common-u...@hadoop.apache.org" <user@hadoop.apache.org> 
>Cc: Raj Hadoop <hadoop...@yahoo.com> 
>Sent: Thursday, May 16, 2013 11:34 AM
>Subject: Re: Configuring SSH - is it required? for a psedo distriburted mode?
>
>
>
>Actually, I should amend my statement -- SSH is required, but passwordless ssh 
>(i guess) you can live without if you are willing to enter your password for 
>each process that gets started.
>
>But Why wouldn't you want to implement passwordless ssh in a pseudo 
>distributed cluster ?  Its very easy to implement on a single node: 
>
>cat ~/.ssh/id_rsa.pub /root/.ssh/authorized_keys 
>
>
>
>
>
>
>
>On Thu, May 16, 2013 at 11:31 AM, Jay Vyas <jayunit...@gmail.com> wrote:
>
>Yes it is required -- in psuedodistributed node the jobtracker is not 
>necessarily aware that the task trackers  / data nodes are on the same 
>machine, and will thus attempt to ssh into them when starting the respective 
>deamons etc (i.e. start-all.sh)
>>
>>
>>
>>
>>On Thu, May 16, 2013 at 11:21 AM, kishore alajangi 
>><alajangikish...@gmail.com> wrote:
>>
>>When you start the hadoop procecess, each process will ask the password to 
>>start, to overcome this we will configure SSH if you use single node or 
>>multiple nodes for each process, if you can enter the password for each 
>>process Its not a mandatory even if you use multiple systems.
>>>
>>>Thanks,
>>>Kishore.
>>>
>>>
>>>
>>>On Thu, May 16, 2013 at 8:24 PM, Raj Hadoop <hadoop...@yahoo.com> wrote:
>>>
>>> Hi,
>>>> 
>>>>I have a dedicated user on Linux server for hadoop. I am installing it in 
>>>>psedo distributed mode on this box. I want to test my programs on this 
>>>>machine. But i see that in installation steps - they were mentioned that 
>>>>SSH needs to be configured. If it is single node, I dont require it 
>>>>...right? Please advise. 
>>>> 
>>>>I was looking at this site
>>>>http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/
>>>> 
>>>>It menionted like this - 
>>>>"Hadoop requires SSH access to manage its nodes, i.e. remote machines plus 
>>>>your local machine if you want to use Hadoop on it (which is what we want 
>>>>to do in this short tutorial). For our single-node setup of Hadoop, we 
>>>>therefore need to configure SSH access to localhost for the hduser user we 
>>>>created in the previous section.
>>>>"
>>>> 
>>>>Thanks,
>>>>Raj
>>>> 
>>>
>>
>>
>>-- 
>>Jay Vyas
>>http://jayunit100.blogspot.com/ 
>
>
>-- 
>Jay Vyas
>http://jayunit100.blogspot.com/ 
>
>

Reply via email to