Divij,
In regards to your ssh problem --
1) make sure that your authorized_keys file contains the public key (not the
private key).
2) make sure the permissions on the .ssh directory and the files within it
are correct. They should look something like this:

dvrya...@ubuntu:~$ ls -la .ssh/
total 24
drwx------  2 dvryaboy dvryaboy 4096 2009-06-18 09:28 .
drwxr-xr-x 58 dvryaboy dvryaboy 4096 2009-06-18 17:18 ..
-rw-r--r--  1 dvryaboy dvryaboy  397 2009-06-18 09:28 authorized_keys
-rw-------  1 dvryaboy dvryaboy 1675 2009-06-18 09:27 id_rsa
-rw-r--r--  1 dvryaboy dvryaboy  397 2009-06-18 09:27 id_rsa.pub
-rw-r--r--  1 dvryaboy dvryaboy 1768 2009-06-18 09:31 known_hosts

(note that the private key and the directory are restricted to my user).

-D

On Fri, Jun 19, 2009 at 9:04 AM, Divij Durve <divij.t...@gmail.com> wrote:

> Thanks for the info aaron. I think the $HADOOP_HOME does get resolved but i
> will change it anyway. I have tried all possible methods of getting the
> passwordless ssh to work even done cat <file where generated key is saved>
> >> <authorized keys file>
>
> It still asks for the pass for "ssh localhost". I moved the job tracker to
> the main  node and then the cluster started working i did a data load but
> when i sent out a query like select count(1) from <table name> it gave me
> an
> error. the query select * from <table name> worked just fine. I really cant
> figure out whats going wrong. I sent a mail out with the error a after this
> mail. Also let me know if there is any added info i need to give to help
> with a solution.
>
> Thanks
> Divij
>
>
> On Thu, Jun 18, 2009 at 4:32 PM, Aaron Kimball <aa...@cloudera.com> wrote:
>
> > Are you encountering specific problems?
> >
> > I don't think that hadoop's config files will evaluate environment
> > variables. So $HADOOP_HOME won't be interpreted correctly.
> >
> > For passwordless ssh, see
> >
> http://rcsg-gsir.imsb-dsgi.nrc-cnrc.gc.ca/documents/internet/node31.htmlor
> > just check the manpage for ssh-keygen.
> >
> > - Aaron
> >
> > On Wed, Jun 17, 2009 at 9:30 AM, Divij Durve <divij.t...@gmail.com>
> wrote:
> >
> > > Im trying to setup a cluster with 3 different machines running Fedora.
> I
> > > cant get them to log into the localhost without the password but thats
> > the
> > > least of my worries at the moment.
> > >
> > > I am posting my config files and the master and slave files let me know
> > if
> > > anyone can spot a problem with the configs...
> > >
> > >
> > > Hadoop-site.xml
> > > <?xml version="1.0"?>
> > > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
> > >
> > > <!-- Put site-specific property overrides in this file. -->
> > >
> > > <configuration>
> > > <property>
> > >    <name>dfs.data.dir</name>
> > >        <value>$HADOOP_HOME/dfs-data</value>
> > >            <final>true</final>
> > >              </property>
> > >
> > >  <property>
> > >     <name>dfs.name.dir</name>
> > >         <value>$HADOOP_HOME/dfs-name</value>
> > >             <final>true</final>
> > >               </property>
> > >
> > >
> > > <property>
> > >  <name>hadoop.tmp.dir</name>
> > >    <value>$HADOOP_HOME/hadoop-tmp</value>
> > >      <description>A base for other temporary directories.</description>
> > >      </property>
> > >
> > >
> > > <property>
> > >  <name>fs.default.name</name>
> > >    <value>hdfs://gobi.<something>.<something>:54310</value>
> > >      <description>The name of the default file system.  A URI whose
> > >        scheme and authority determine the FileSystem implementation.
>  The
> > >          uri's scheme determines the config property (fs.SCHEME.impl)
> > > naming
> > >            the FileSystem implementation class.  The uri's authority is
> > > used to
> > >              determine the host, port, etc. for a
> > FileSystem.</description>
> > >              </property>
> > >
> > > <property>
> > >  <name>mapred.job.tracker</name>
> > >    <value>kalahari.<something>.<something>:54311</value>
> > >      <description>The host and port that the MapReduce job tracker runs
> > >        at.  If "local", then jobs are run in-process as a single map
> > >          and reduce task.
> > >            </description>
> > >            </property>
> > >
> > >  <property>
> > >     <name>mapred.system.dir</name>
> > >         <value>$HADOOP_HOME/mapred-system</value>
> > >             <final>true</final>
> > >               </property>
> > >
> > > <property>
> > >  <name>dfs.replication</name>
> > >    <value>1</value>
> > >      <description>Default block replication.
> > >        The actual number of replications can be specified when the file
> > is
> > > created.
> > >          The default is used if replication is not specified in create
> > > time.
> > >            </description>
> > >            </property>
> > >
> > >
> > > <property>
> > >  <name>mapred.local.dir</name>
> > >    <value>$HADOOP_HOME/mapred-local</value>
> > >  <name>dfs.replication</name>
> > >    <value>1</value>
> > > </property>
> > >
> > >
> > > </configuration>
> > >
> > >
> > > Slave:
> > > kongur.something.something
> > >
> > > master:
> > > kalahari.something.something
> > >
> > > i execute the dfs-start.sh command from gobi.something.something.
> > >
> > > is there any other info that i should provide in order to help? Also
> > Kongur
> > > is where im running the data node the master file on kongur should have
> > > localhost in it rite? thanks for the help
> > >
> > > Divij
> > >
> >
>

Reply via email to