Thank you very much mohamad for your attention.I followed the steps but the
error is the same as the last time.
and there is my hosts file:

127.0.0.1       localhost
#127.0.0.1      ubuntu.ubuntu-domain    ubuntu


# The following lines are desirable for IPv6 capable hosts

#::1     ip6-localhost ip6-loopback
#fe00::0 ip6-localnet
#ff00::0 ip6-mcastprefix
#ff02::1 ip6-allnodes
#ff02::2 ip6-allrouters

but no effect :(

On Wed, Jun 6, 2012 at 8:25 PM, Mohammad Tariq <donta...@gmail.com> wrote:

> also change the permissions of these directories to 777.
>
> Regards,
>     Mohammad Tariq
>
>
> On Wed, Jun 6, 2012 at 11:54 PM, Mohammad Tariq <donta...@gmail.com>
> wrote:
> > create a directory "/home/username/hdfs" (or at some place of your
> > choice)..inside this hdfs directory create three sub directories -
> > name, data, and temp, then follow these steps :
> >
> > add following properties in your core-site.xml -
> >
> > <property>
> >          <name>fs.default.name</name>
> >          <value>hdfs://localhost:9000/</value>
> >        </property>
> >
> >        <property>
> >          <name>hadoop.tmp.dir</name>
> >          <value>/home/mohammad/hdfs/temp</value>
> >        </property>
> >
> > then add following two properties in your hdfs-site.xml -
> >
> > <property>
> >                <name>dfs.replication</name>
> >                <value>1</value>
> >        </property>
> >
> >        <property>
> >                <name>dfs.name.dir</name>
> >                <value>/home/mohammad/hdfs/name</value>
> >        </property>
> >
> >        <property>
> >                <name>dfs.data.dir</name>
> >                <value>/home/mohammad/hdfs/data</value>
> >        </property>
> >
> > finally add this property in your mapred-site.xml -
> >
> >       <property>
> >          <name>mapred.job.tracker</name>
> >          <value>hdfs://localhost:9001</value>
> >        </property>
> >
> > NOTE: you can give any name to these directories of your choice, just
> > keep in mind you have to give same names as values of
> >           above specified properties in your configuration files.
> > (give full path of these directories, not just the name of the
> > directory)
> >
> > After this  follow the steps provided in the previous reply.
> >
> > Regards,
> >     Mohammad Tariq
> >
> >
> > On Wed, Jun 6, 2012 at 11:42 PM, Babak Bastan <babak...@gmail.com>
> wrote:
> >> thank's Mohammad
> >>
> >> with this command:
> >>
> >> babak@ubuntu:~/Downloads/hadoop/bin$ hadoop namenode -format
> >>
> >> this is my output:
> >>
> >> 12/06/06 20:05:20 INFO namenode.NameNode: STARTUP_MSG:
> >> /************************************************************
> >> STARTUP_MSG: Starting NameNode
> >> STARTUP_MSG:   host = ubuntu/127.0.1.1
> >> STARTUP_MSG:   args = [-format]
> >> STARTUP_MSG:   version = 0.20.2
> >> STARTUP_MSG:   build =
> >> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20 -r
> >> 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010
> >> ************************************************************/
> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem:
> >> fsOwner=babak,babak,adm,dialout,cdrom,plugdev,lpadmin,admin,sambashare
> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem: supergroup=supergroup
> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem: isPermissionEnabled=true
> >> 12/06/06 20:05:20 INFO common.Storage: Image file of size 95 saved in 0
> >> seconds.
> >> 12/06/06 20:05:20 INFO common.Storage: Storage directory
> >> /tmp/hadoop-babak/dfs/name has been successfully formatted.
> >> 12/06/06 20:05:20 INFO namenode.NameNode: SHUTDOWN_MSG:
> >> /************************************************************
> >> SHUTDOWN_MSG: Shutting down NameNode at ubuntu/127.0.1.1
> >> ************************************************************/
> >>
> >> by this command:
> >>
> >> babak@ubuntu:~/Downloads/hadoop/bin$ start-dfs.sh
> >>
> >> this is the out put
> >>
> >> mkdir: kann Verzeichnis „/home/babak/Downloads/hadoop/bin/../logs“ nicht
> >> anlegen: Keine Berechtigung
> >>
> >> this out put(it's in german and it means no right to make this folder)
> >>
> >>
> >> On Wed, Jun 6, 2012 at 7:59 PM, Mohammad Tariq <donta...@gmail.com>
> wrote:
> >>>
> >>> once we are done with the configuration, we need to format the file
> >>> system..use this command to do that-
> >>> bin/hadoop namenode -format
> >>>
> >>> after this, hadoop daemon processes should be started using following
> >>> commands -
> >>> bin/start-dfs.sh (it'll start NN & DN)
> >>> bin/start-mapred.sh (it'll start JT & TT)
> >>>
> >>> after this use jps to check if everything is alright or point your
> >>> browser to localhost:50070..if you further find any problem provide us
> >>> with the error logs..:)
> >>>
> >>> Regards,
> >>>     Mohammad Tariq
> >>>
> >>>
> >>> On Wed, Jun 6, 2012 at 11:22 PM, Babak Bastan <babak...@gmail.com>
> wrote:
> >>> > were you able to format hdfs properly???
> >>> > I did'nt get your question,Do you mean HADOOP_HOME? or where did I
> >>> > install
> >>> > Hadoop?
> >>> >
> >>> > On Wed, Jun 6, 2012 at 7:49 PM, Mohammad Tariq <donta...@gmail.com>
> >>> > wrote:
> >>> >>
> >>> >> if you are getting only this, it means your hadoop is not
> >>> >> running..were you able to format hdfs properly???
> >>> >>
> >>> >> Regards,
> >>> >>     Mohammad Tariq
> >>> >>
> >>> >>
> >>> >> On Wed, Jun 6, 2012 at 11:17 PM, Babak Bastan <babak...@gmail.com>
> >>> >> wrote:
> >>> >> > Hi MohammadmI irun jps in my shel I can see this result:
> >>> >> > 2213 Jps
> >>> >> >
> >>> >> >
> >>> >> > On Wed, Jun 6, 2012 at 7:44 PM, Mohammad Tariq <
> donta...@gmail.com>
> >>> >> > wrote:
> >>> >> >>
> >>> >> >> you can also use "jps" command at your shell to see whether
> Hadoop
> >>> >> >> processes are running or not.
> >>> >> >>
> >>> >> >> Regards,
> >>> >> >>     Mohammad Tariq
> >>> >> >>
> >>> >> >>
> >>> >> >> On Wed, Jun 6, 2012 at 11:12 PM, Mohammad Tariq <
> donta...@gmail.com>
> >>> >> >> wrote:
> >>> >> >> > Hi Babak,
> >>> >> >> >
> >>> >> >> >  You have to type it in you web browser..Hadoop provides us a
> web
> >>> >> >> > GUI
> >>> >> >> > that not only allows us to browse through the file system, but
> to
> >>> >> >> > download the files as well..Apart from that it also provides a
> web
> >>> >> >> > GUI
> >>> >> >> > that can be used to see the status of Jobtracker and
> >>> >> >> > Tasktracker..When
> >>> >> >> > you run a Hive or Pig job or a Mapreduce job, you can point
> your
> >>> >> >> > browser to http://localhost:50030 to see the status and logs
> of
> >>> >> >> > your
> >>> >> >> > job.
> >>> >> >> >
> >>> >> >> > Regards,
> >>> >> >> >     Mohammad Tariq
> >>> >> >> >
> >>> >> >> >
> >>> >> >> > On Wed, Jun 6, 2012 at 8:28 PM, Babak Bastan <
> babak...@gmail.com>
> >>> >> >> > wrote:
> >>> >> >> >> Thank you shashwat for the answer,
> >>> >> >> >> where should I type http://localhost:50070?
> >>> >> >> >> I typed here: hive>http://localhost:50070 but nothing as
> result
> >>> >> >> >>
> >>> >> >> >>
> >>> >> >> >> On Wed, Jun 6, 2012 at 3:32 PM, shashwat shriparv
> >>> >> >> >> <dwivedishash...@gmail.com> wrote:
> >>> >> >> >>>
> >>> >> >> >>> first type http://localhost:50070 whether this is opening
> or not
> >>> >> >> >>> and
> >>> >> >> >>> check
> >>> >> >> >>> how many nodes are available, check some of the hadoop shell
> >>> >> >> >>> commands
> >>> >> >> >>>
> >>> >> >> >>> from
> http://hadoop.apache.org/common/docs/r0.18.3/hdfs_shell.html
> >>> >> >> >>> run
> >>> >> >> >>> example mapreduce task on hadoop take example from here
> >>> >> >> >>>
> >>> >> >> >>>
> >>> >> >> >>>
> >>> >> >> >>> :
> http://www.michael-noll.com/blog/2011/04/09/benchmarking-and-stress-testing-an-hadoop-cluster-with-terasort-testdfsio-nnbench-mrbench/
> >>> >> >> >>>
> >>> >> >> >>> if all the above you can do sucessfully means hadoop is
> >>> >> >> >>> configured
> >>> >> >> >>> correctly
> >>> >> >> >>>
> >>> >> >> >>> Regards
> >>> >> >> >>> Shashwat
> >>> >> >> >>>
> >>> >> >> >>>
> >>> >> >> >>> On Wed, Jun 6, 2012 at 1:30 AM, Babak Bastan
> >>> >> >> >>> <babak...@gmail.com>
> >>> >> >> >>> wrote:
> >>> >> >> >>>>
> >>> >> >> >>>> no I'm not working on CDH.Is there a way to test if my
> Hadoop
> >>> >> >> >>>> works
> >>> >> >> >>>> fine
> >>> >> >> >>>> or not?
> >>> >> >> >>>>
> >>> >> >> >>>>
> >>> >> >> >>>> On Tue, Jun 5, 2012 at 9:55 PM, Bejoy KS <
> bejoy...@yahoo.com>
> >>> >> >> >>>> wrote:
> >>> >> >> >>>>>
> >>> >> >> >>>>> Hi Babak
> >>> >> >> >>>>>
> >>> >> >> >>>>> You gotta follow those instructions in the apace site to
> set
> >>> >> >> >>>>> up
> >>> >> >> >>>>> hadoop
> >>> >> >> >>>>> from scratch and ensure that hdfs is working first. You
> should
> >>> >> >> >>>>> be
> >>> >> >> >>>>> able to
> >>> >> >> >>>>> read and write files to hdfs before you do your next steps.
> >>> >> >> >>>>>
> >>> >> >> >>>>> Are you on CDH or apache distribution of hadoop? If it is
> CDH
> >>> >> >> >>>>> there
> >>> >> >> >>>>> are
> >>> >> >> >>>>> detailed instructions on Cloudera web site.
> >>> >> >> >>>>>
> >>> >> >> >>>>> Regards
> >>> >> >> >>>>> Bejoy KS
> >>> >> >> >>>>>
> >>> >> >> >>>>> Sent from handheld, please excuse typos.
> >>> >> >> >>>>> ________________________________
> >>> >> >> >>>>> From: Babak Bastan <babak...@gmail.com>
> >>> >> >> >>>>> Date: Tue, 5 Jun 2012 21:30:22 +0200
> >>> >> >> >>>>> To: <user@hive.apache.org>
> >>> >> >> >>>>> ReplyTo: user@hive.apache.org
> >>> >> >> >>>>> Subject: Re: Error while Creating Table in Hive
> >>> >> >> >>>>>
> >>> >> >> >>>>> @Bejoy: I set the fs.default.name in the core-site.xml
> and I
> >>> >> >> >>>>> did
> >>> >> >> >>>>> all
> >>> >> >> >>>>> of
> >>> >> >> >>>>> thing that was mentioned in the reference but no effect
> >>> >> >> >>>>>
> >>> >> >> >>>>> On Tue, Jun 5, 2012 at 8:43 PM, Babak Bastan
> >>> >> >> >>>>> <babak...@gmail.com>
> >>> >> >> >>>>> wrote:
> >>> >> >> >>>>>>
> >>> >> >> >>>>>> Ok sorry but that was my Mistake .I thought it works but
> no.
> >>> >> >> >>>>>> I wrote the command without ; and then I think It works
> but
> >>> >> >> >>>>>> with
> >>> >> >> >>>>>> ;
> >>> >> >> >>>>>> at
> >>> >> >> >>>>>> the end of command
> >>> >> >> >>>>>>
> >>> >> >> >>>>>> CREATE TABLE pokes (foo INT, bar STRING);
> >>> >> >> >>>>>>
> >>> >> >> >>>>>> does'nt work
> >>> >> >> >>>>>>
> >>> >> >> >>>>>>
> >>> >> >> >>>>>> On Tue, Jun 5, 2012 at 8:34 PM, shashwat shriparv
> >>> >> >> >>>>>> <dwivedishash...@gmail.com> wrote:
> >>> >> >> >>>>>>>
> >>> >> >> >>>>>>> inside configuration. all properties will be inside the
> >>> >> >> >>>>>>> configuration
> >>> >> >> >>>>>>> tags
> >>> >> >> >>>>>>>
> >>> >> >> >>>>>>>
> >>> >> >> >>>>>>> On Tue, Jun 5, 2012 at 11:53 PM, Babak Bastan
> >>> >> >> >>>>>>> <babak...@gmail.com>
> >>> >> >> >>>>>>> wrote:
> >>> >> >> >>>>>>>>
> >>> >> >> >>>>>>>> Thank you so much my friend your idee works fine(no
> error)
> >>> >> >> >>>>>>>> you
> >>> >> >> >>>>>>>> are
> >>> >> >> >>>>>>>> the best :)
> >>> >> >> >>>>>>>>
> >>> >> >> >>>>>>>>
> >>> >> >> >>>>>>>> On Tue, Jun 5, 2012 at 8:20 PM, Babak Bastan
> >>> >> >> >>>>>>>> <babak...@gmail.com>
> >>> >> >> >>>>>>>> wrote:
> >>> >> >> >>>>>>>>>
> >>> >> >> >>>>>>>>> It must be inside the <configuration></configuration>
> or
> >>> >> >> >>>>>>>>> outside
> >>> >> >> >>>>>>>>> this?
> >>> >> >> >>>>>>>>>
> >>> >> >> >>>>>>>>>
> >>> >> >> >>>>>>>>> On Tue, Jun 5, 2012 at 8:15 PM, shashwat shriparv
> >>> >> >> >>>>>>>>> <dwivedishash...@gmail.com> wrote:
> >>> >> >> >>>>>>>>>>
> >>> >> >> >>>>>>>>>> It will be inside hive/conf
> >>> >> >> >>>>>>>>>>
> >>> >> >> >>>>>>>>>>
> >>> >> >> >>>>>>>>>> On Tue, Jun 5, 2012 at 11:43 PM, Babak Bastan
> >>> >> >> >>>>>>>>>> <babak...@gmail.com>
> >>> >> >> >>>>>>>>>> wrote:
> >>> >> >> >>>>>>>>>>>
> >>> >> >> >>>>>>>>>>> Thanks sShashwat, and where is this hive-site.xml
> >>> >> >> >>>>>>>>>>>
> >>> >> >> >>>>>>>>>>> On Tue, Jun 5, 2012 at 8:02 PM, shashwat shriparv
> >>> >> >> >>>>>>>>>>> <dwivedishash...@gmail.com> wrote:
> >>> >> >> >>>>>>>>>>>>
> >>> >> >> >>>>>>>>>>>> set
> >>> >> >> >>>>>>>>>>>>
> >>> >> >> >>>>>>>>>>>> hive.metastore.warehouse.dir in hive-site.xml
> >>> >> >> >>>>>>>>>>>>
> >>> >> >> >>>>>>>>>>>> <property>
> >>> >> >> >>>>>>>>>>>>   <name>hive.metastore.local</name>
> >>> >> >> >>>>>>>>>>>>   <value>true</value>
> >>> >> >> >>>>>>>>>>>> </property>
> >>> >> >> >>>>>>>>>>>>
> >>> >> >> >>>>>>>>>>>> <name>hive.metastore.warehouse.dir</name>
> >>> >> >> >>>>>>>>>>>>                <value>/home/<your
> >>> >> >> >>>>>>>>>>>> username>/hivefolder</value>
> >>> >> >> >>>>>>>>>>>>                <description>location of default
> >>> >> >> >>>>>>>>>>>> database
> >>> >> >> >>>>>>>>>>>> for
> >>> >> >> >>>>>>>>>>>> the
> >>> >> >> >>>>>>>>>>>> warehouse</description>
> >>> >> >> >>>>>>>>>>>>        </property>
> >>> >> >> >>>>>>>>>>>>
> >>> >> >> >>>>>>>>>>>>
> >>> >> >> >>>>>>>>>>>> On Tue, Jun 5, 2012 at 10:43 PM, Babak Bastan
> >>> >> >> >>>>>>>>>>>> <babak...@gmail.com> wrote:
> >>> >> >> >>>>>>>>>>>>>
> >>> >> >> >>>>>>>>>>>>> Hello Experts ,
> >>> >> >> >>>>>>>>>>>>>
> >>> >> >> >>>>>>>>>>>>> I'm new in Hive .When try to create a test Table in
> >>> >> >> >>>>>>>>>>>>> Hive
> >>> >> >> >>>>>>>>>>>>> I
> >>> >> >> >>>>>>>>>>>>> get
> >>> >> >> >>>>>>>>>>>>> an error.I want to run this command:
> >>> >> >> >>>>>>>>>>>>> CREATE TABLE Test (DateT STRING, Url STRING,
> Content
> >>> >> >> >>>>>>>>>>>>> STRING);
> >>> >> >> >>>>>>>>>>>>> but this error occured:
> >>> >> >> >>>>>>>>>>>>> FAILED: Error in metadata:
> MetaException(message:Got
> >>> >> >> >>>>>>>>>>>>> exception:
> >>> >> >> >>>>>>>>>>>>> java.io.FileNotFoundException File
> >>> >> >> >>>>>>>>>>>>> file:/user/hive/warehouse/test does not
> >>> >> >> >>>>>>>>>>>>> exist.)
> >>> >> >> >>>>>>>>>>>>> FAILED: Execution Error, return code 1 from
> >>> >> >> >>>>>>>>>>>>> org.apache.hadoop.hive.ql.exec.DDLTask
> >>> >> >> >>>>>>>>>>>>> How can I solve this Problem?
> >>> >> >> >>>>>>>>>>>>> Thank you so much
> >>> >> >> >>>>>>>>>>>>
> >>> >> >> >>>>>>>>>>>>
> >>> >> >> >>>>>>>>>>>>
> >>> >> >> >>>>>>>>>>>>
> >>> >> >> >>>>>>>>>>>> --
> >>> >> >> >>>>>>>>>>>>
> >>> >> >> >>>>>>>>>>>>
> >>> >> >> >>>>>>>>>>>> ∞
> >>> >> >> >>>>>>>>>>>>
> >>> >> >> >>>>>>>>>>>> Shashwat Shriparv
> >>> >> >> >>>>>>>>>>>>
> >>> >> >> >>>>>>>>>>>>
> >>> >> >> >>>>>>>>>>>
> >>> >> >> >>>>>>>>>>
> >>> >> >> >>>>>>>>>>
> >>> >> >> >>>>>>>>>>
> >>> >> >> >>>>>>>>>> --
> >>> >> >> >>>>>>>>>>
> >>> >> >> >>>>>>>>>>
> >>> >> >> >>>>>>>>>> ∞
> >>> >> >> >>>>>>>>>>
> >>> >> >> >>>>>>>>>> Shashwat Shriparv
> >>> >> >> >>>>>>>>>>
> >>> >> >> >>>>>>>>>>
> >>> >> >> >>>>>>>>>
> >>> >> >> >>>>>>>>
> >>> >> >> >>>>>>>
> >>> >> >> >>>>>>>
> >>> >> >> >>>>>>>
> >>> >> >> >>>>>>> --
> >>> >> >> >>>>>>>
> >>> >> >> >>>>>>>
> >>> >> >> >>>>>>> ∞
> >>> >> >> >>>>>>>
> >>> >> >> >>>>>>> Shashwat Shriparv
> >>> >> >> >>>>>>>
> >>> >> >> >>>>>>>
> >>> >> >> >>>>>>
> >>> >> >> >>>>>
> >>> >> >> >>>>
> >>> >> >> >>>
> >>> >> >> >>>
> >>> >> >> >>>
> >>> >> >> >>> --
> >>> >> >> >>>
> >>> >> >> >>>
> >>> >> >> >>> ∞
> >>> >> >> >>>
> >>> >> >> >>> Shashwat Shriparv
> >>> >> >> >>>
> >>> >> >> >>>
> >>> >> >> >>
> >>> >> >
> >>> >> >
> >>> >
> >>> >
> >>
> >>
>

Reply via email to