ok..we'll give it a final shot..then i'll email configured hadoop to
your email address..delete the hdfs directory which contains tmp, data
and name..recreate it..format hdfs again and then start the processes.

Regards,
    Mohammad Tariq


On Thu, Jun 7, 2012 at 2:22 AM, Babak Bastan <babak...@gmail.com> wrote:
> I 've performed the steps but the same error in this step as before:
> bin/start-dfs.sh
> It is about my permission to make directory
>
> On Wed, Jun 6, 2012 at 10:33 PM, Mohammad Tariq <donta...@gmail.com> wrote:
>>
>> actually this blog post explains how to install cloudera's hadoop
>> distribution...if you have followed this post and installed cloudera's
>> distribution then your logs should ideally be inside
>> /usr/lib/hadoop/logs (if everything was fine)..anyway try the steps I
>> have given and let me know.
>>
>> Regards,
>>     Mohammad Tariq
>>
>>
>> On Thu, Jun 7, 2012 at 1:52 AM, Babak Bastan <babak...@gmail.com> wrote:
>> > by the way,you are a very nice man my friend:Thank you so much :)
>> >
>> > what do you mean aboat this post in stackoverflow?
>> >
>> > I am assuming that is your first installation of hadoop.
>> >
>> > At the beginning please check if your daemons are working. To do that
>> > use
>> > (in terminal):
>> >
>> > jps
>> >
>> > If only jps appears that means all daemons are down. Please check the
>> > log
>> > files. Especially the namenode. Log folder is probably somewhere there
>> > /usr/lib/hadoop/logs
>> >
>> > If you have some permission problems. Use this guide during the
>> > installation.
>> >
>> > Good installation guide
>> >
>> > I am shooting with this explanations but these are most common problems.
>> >
>> >
>> > On Wed, Jun 6, 2012 at 10:15 PM, Babak Bastan <babak...@gmail.com>
>> > wrote:
>> >>
>> >> I checked it but no hadoop folder :(
>> >> yes you are right.I'm a student and I want to make a very very simple
>> >> programm hive but untill now hmmmmmmmmm
>> >>
>> >>
>> >> On Wed, Jun 6, 2012 at 10:12 PM, Babak Bastan <babak...@gmail.com>
>> >> wrote:
>> >>>
>> >>> no one error:
>> >>> i.e if I run this one
>> >>>
>> >>> hostname --fqdn
>> >>>
>> >>>  with the condition that I send to you :
>> >>>
>> >>> 127.0.0.1       localhost
>> >>> #127.0.0.1      ubuntu.ubuntu-domain    ubuntu
>> >>> # The following lines are desirable for IPv6 capable hosts
>> >>> #::1     ip6-localhost ip6-loopback
>> >>> #fe00::0 ip6-localnet
>> >>> #ff00::0 ip6-mcastprefix
>> >>> #ff02::1 ip6-allnodes
>> >>> #ff02::2 ip6-allrouters
>> >>>
>> >>> I get this error:
>> >>>
>> >>> hostname: Name or service not known
>> >>>
>> >>> Or in the second step by this command:
>> >>>
>> >>> babak@ubuntu:~/Downloads/hadoop/bin$ start-hdfs.sh
>> >>>
>> >>> these lines of error:
>> >>>
>> >>>
>> >>> mkdir: kann Verzeichnis „/home/babak/Downloads/hadoop/bin/../logs“
>> >>> nicht
>> >>> anlegen: Keine Berechtigung
>> >>> starting namenode, logging to
>> >>>
>> >>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-namenode-ubuntu.out
>> >>> /home/babak/Downloads/hadoop/bin/hadoop-daemon.sh: Zeile 117:
>> >>>
>> >>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-namenode-ubuntu.out:
>> >>> Datei oder Verzeichnis nicht gefunden
>> >>> head:
>> >>>
>> >>> „/home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-namenode-ubuntu.out“
>> >>> kann nicht zum Lesen geöffnet werden: Datei oder Verzeichnis nicht
>> >>> gefunden
>> >>> localhost: mkdir: kann Verzeichnis
>> >>> „/home/babak/Downloads/hadoop/bin/../logs“ nicht anlegen: Keine
>> >>> Berechtigung
>> >>> localhost: starting datanode, logging to
>> >>>
>> >>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-datanode-ubuntu.out
>> >>> localhost: /home/babak/Downloads/hadoop/bin/hadoop-daemon.sh: Zeile
>> >>> 117:
>> >>>
>> >>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-datanode-ubuntu.out:
>> >>> Datei oder Verzeichnis nicht gefunden
>> >>> localhost: head:
>> >>>
>> >>> „/home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-datanode-ubuntu.out“
>> >>> kann nicht zum Lesen geöffnet werden: Datei oder Verzeichnis nicht
>> >>> gefunden
>> >>> localhost: mkdir: kann Verzeichnis
>> >>> „/home/babak/Downloads/hadoop/bin/../logs“ nicht anlegen: Keine
>> >>> Berechtigung
>> >>> localhost: starting secondarynamenode, logging to
>> >>>
>> >>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-secondarynamenode-ubuntu.out
>> >>> localhost: /home/babak/Downloads/hadoop/bin/hadoop-daemon.sh: Zeile
>> >>> 117:
>> >>>
>> >>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-secondarynamenode-ubuntu.out:
>> >>> Datei oder Verzeichnis nicht gefunden
>> >>> localhost: head:
>> >>>
>> >>> „/home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-secondarynamenode-ubuntu.out“
>> >>> kann nicht zum Lesen geöffnet werden: Datei oder Verzeichnis nicht
>> >>> gefunden
>> >>>
>> >>> they said no permision to make logs in this
>> >>> path:/home/babak/Downloads/hadoop/bin/../logs
>> >>>
>> >>>  and generally I cant create a table in hive and get this one:
>> >>>
>> >>> FAILED: Error in metadata: MetaException(message:Got exception:
>> >>> java.io.FileNotFoundException File file:/user/hive/warehouse/test does
>> >>> not
>> >>> exist.)
>> >>> FAILED: Execution Error, return code 1 from
>> >>> org.apache.hadoop.hive.ql.exec.DDLTask
>> >>>
>> >>> On Wed, Jun 6, 2012 at 10:02 PM, shashwat shriparv
>> >>> <dwivedishash...@gmail.com> wrote:
>> >>>>
>> >>>> whats the error babak ???
>> >>>>
>> >>>>
>> >>>> On Thu, Jun 7, 2012 at 1:25 AM, Babak Bastan <babak...@gmail.com>
>> >>>> wrote:
>> >>>>>
>> >>>>> What the hell is that?I see no log folder there
>> >>>>>
>> >>>>>
>> >>>>> On Wed, Jun 6, 2012 at 9:41 PM, Mohammad Tariq <donta...@gmail.com>
>> >>>>> wrote:
>> >>>>>>
>> >>>>>> go to your HADOOP_HOME i.e your hadoop directory(that includes bin,
>> >>>>>> conf etc)..you can find logs directory there..
>> >>>>>>
>> >>>>>> Regards,
>> >>>>>>     Mohammad Tariq
>> >>>>>>
>> >>>>>>
>> >>>>>> On Thu, Jun 7, 2012 at 1:09 AM, Babak Bastan <babak...@gmail.com>
>> >>>>>> wrote:
>> >>>>>> > hoe can I get my log mohammad?
>> >>>>>> >
>> >>>>>> >
>> >>>>>> > On Wed, Jun 6, 2012 at 9:36 PM, Mohammad Tariq
>> >>>>>> > <donta...@gmail.com>
>> >>>>>> > wrote:
>> >>>>>> >>
>> >>>>>> >> could you post your logs???that would help me in understanding
>> >>>>>> >> the
>> >>>>>> >> problem properly.
>> >>>>>> >>
>> >>>>>> >> Regards,
>> >>>>>> >>     Mohammad Tariq
>> >>>>>> >>
>> >>>>>> >>
>> >>>>>> >> On Thu, Jun 7, 2012 at 1:02 AM, Babak Bastan
>> >>>>>> >> <babak...@gmail.com>
>> >>>>>> >> wrote:
>> >>>>>> >> > Thank you very much mohamad for your attention.I followed the
>> >>>>>> >> > steps but
>> >>>>>> >> > the
>> >>>>>> >> > error is the same as the last time.
>> >>>>>> >> > and there is my hosts file:
>> >>>>>> >> >
>> >>>>>> >> > 127.0.0.1       localhost
>> >>>>>> >> > #127.0.0.1      ubuntu.ubuntu-domain    ubuntu
>> >>>>>> >> >
>> >>>>>> >> >
>> >>>>>> >> > # The following lines are desirable for IPv6 capable hosts
>> >>>>>> >> >
>> >>>>>> >> > #::1     ip6-localhost ip6-loopback
>> >>>>>> >> > #fe00::0 ip6-localnet
>> >>>>>> >> > #ff00::0 ip6-mcastprefix
>> >>>>>> >> > #ff02::1 ip6-allnodes
>> >>>>>> >> > #ff02::2 ip6-allrouters
>> >>>>>> >> >
>> >>>>>> >> > but no effect :(
>> >>>>>> >> >
>> >>>>>> >> > On Wed, Jun 6, 2012 at 8:25 PM, Mohammad Tariq
>> >>>>>> >> > <donta...@gmail.com>
>> >>>>>> >> > wrote:
>> >>>>>> >> >>
>> >>>>>> >> >> also change the permissions of these directories to 777.
>> >>>>>> >> >>
>> >>>>>> >> >> Regards,
>> >>>>>> >> >>     Mohammad Tariq
>> >>>>>> >> >>
>> >>>>>> >> >>
>> >>>>>> >> >> On Wed, Jun 6, 2012 at 11:54 PM, Mohammad Tariq
>> >>>>>> >> >> <donta...@gmail.com>
>> >>>>>> >> >> wrote:
>> >>>>>> >> >> > create a directory "/home/username/hdfs" (or at some place
>> >>>>>> >> >> > of
>> >>>>>> >> >> > your
>> >>>>>> >> >> > choice)..inside this hdfs directory create three sub
>> >>>>>> >> >> > directories -
>> >>>>>> >> >> > name, data, and temp, then follow these steps :
>> >>>>>> >> >> >
>> >>>>>> >> >> > add following properties in your core-site.xml -
>> >>>>>> >> >> >
>> >>>>>> >> >> > <property>
>> >>>>>> >> >> >          <name>fs.default.name</name>
>> >>>>>> >> >> >          <value>hdfs://localhost:9000/</value>
>> >>>>>> >> >> >        </property>
>> >>>>>> >> >> >
>> >>>>>> >> >> >        <property>
>> >>>>>> >> >> >          <name>hadoop.tmp.dir</name>
>> >>>>>> >> >> >          <value>/home/mohammad/hdfs/temp</value>
>> >>>>>> >> >> >        </property>
>> >>>>>> >> >> >
>> >>>>>> >> >> > then add following two properties in your hdfs-site.xml -
>> >>>>>> >> >> >
>> >>>>>> >> >> > <property>
>> >>>>>> >> >> >                <name>dfs.replication</name>
>> >>>>>> >> >> >                <value>1</value>
>> >>>>>> >> >> >        </property>
>> >>>>>> >> >> >
>> >>>>>> >> >> >        <property>
>> >>>>>> >> >> >                <name>dfs.name.dir</name>
>> >>>>>> >> >> >                <value>/home/mohammad/hdfs/name</value>
>> >>>>>> >> >> >        </property>
>> >>>>>> >> >> >
>> >>>>>> >> >> >        <property>
>> >>>>>> >> >> >                <name>dfs.data.dir</name>
>> >>>>>> >> >> >                <value>/home/mohammad/hdfs/data</value>
>> >>>>>> >> >> >        </property>
>> >>>>>> >> >> >
>> >>>>>> >> >> > finally add this property in your mapred-site.xml -
>> >>>>>> >> >> >
>> >>>>>> >> >> >       <property>
>> >>>>>> >> >> >          <name>mapred.job.tracker</name>
>> >>>>>> >> >> >          <value>hdfs://localhost:9001</value>
>> >>>>>> >> >> >        </property>
>> >>>>>> >> >> >
>> >>>>>> >> >> > NOTE: you can give any name to these directories of your
>> >>>>>> >> >> > choice, just
>> >>>>>> >> >> > keep in mind you have to give same names as values of
>> >>>>>> >> >> >           above specified properties in your configuration
>> >>>>>> >> >> > files.
>> >>>>>> >> >> > (give full path of these directories, not just the name of
>> >>>>>> >> >> > the
>> >>>>>> >> >> > directory)
>> >>>>>> >> >> >
>> >>>>>> >> >> > After this  follow the steps provided in the previous
>> >>>>>> >> >> > reply.
>> >>>>>> >> >> >
>> >>>>>> >> >> > Regards,
>> >>>>>> >> >> >     Mohammad Tariq
>> >>>>>> >> >> >
>> >>>>>> >> >> >
>> >>>>>> >> >> > On Wed, Jun 6, 2012 at 11:42 PM, Babak Bastan
>> >>>>>> >> >> > <babak...@gmail.com>
>> >>>>>> >> >> > wrote:
>> >>>>>> >> >> >> thank's Mohammad
>> >>>>>> >> >> >>
>> >>>>>> >> >> >> with this command:
>> >>>>>> >> >> >>
>> >>>>>> >> >> >> babak@ubuntu:~/Downloads/hadoop/bin$ hadoop namenode
>> >>>>>> >> >> >> -format
>> >>>>>> >> >> >>
>> >>>>>> >> >> >> this is my output:
>> >>>>>> >> >> >>
>> >>>>>> >> >> >> 12/06/06 20:05:20 INFO namenode.NameNode: STARTUP_MSG:
>> >>>>>> >> >> >>
>> >>>>>> >> >> >> /************************************************************
>> >>>>>> >> >> >> STARTUP_MSG: Starting NameNode
>> >>>>>> >> >> >> STARTUP_MSG:   host = ubuntu/127.0.1.1
>> >>>>>> >> >> >> STARTUP_MSG:   args = [-format]
>> >>>>>> >> >> >> STARTUP_MSG:   version = 0.20.2
>> >>>>>> >> >> >> STARTUP_MSG:   build =
>> >>>>>> >> >> >>
>> >>>>>> >> >> >>
>> >>>>>> >> >> >> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20
>> >>>>>> >> >> >> -r
>> >>>>>> >> >> >> 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC
>> >>>>>> >> >> >> 2010
>> >>>>>> >> >> >>
>> >>>>>> >> >> >> ************************************************************/
>> >>>>>> >> >> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem:
>> >>>>>> >> >> >>
>> >>>>>> >> >> >>
>> >>>>>> >> >> >>
>> >>>>>> >> >> >> fsOwner=babak,babak,adm,dialout,cdrom,plugdev,lpadmin,admin,sambashare
>> >>>>>> >> >> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem:
>> >>>>>> >> >> >> supergroup=supergroup
>> >>>>>> >> >> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem:
>> >>>>>> >> >> >> isPermissionEnabled=true
>> >>>>>> >> >> >> 12/06/06 20:05:20 INFO common.Storage: Image file of size
>> >>>>>> >> >> >> 95
>> >>>>>> >> >> >> saved
>> >>>>>> >> >> >> in 0
>> >>>>>> >> >> >> seconds.
>> >>>>>> >> >> >> 12/06/06 20:05:20 INFO common.Storage: Storage directory
>> >>>>>> >> >> >> /tmp/hadoop-babak/dfs/name has been successfully
>> >>>>>> >> >> >> formatted.
>> >>>>>> >> >> >> 12/06/06 20:05:20 INFO namenode.NameNode: SHUTDOWN_MSG:
>> >>>>>> >> >> >>
>> >>>>>> >> >> >> /************************************************************
>> >>>>>> >> >> >> SHUTDOWN_MSG: Shutting down NameNode at ubuntu/127.0.1.1
>> >>>>>> >> >> >>
>> >>>>>> >> >> >> ************************************************************/
>> >>>>>> >> >> >>
>> >>>>>> >> >> >> by this command:
>> >>>>>> >> >> >>
>> >>>>>> >> >> >> babak@ubuntu:~/Downloads/hadoop/bin$ start-dfs.sh
>> >>>>>> >> >> >>
>> >>>>>> >> >> >> this is the out put
>> >>>>>> >> >> >>
>> >>>>>> >> >> >> mkdir: kann Verzeichnis
>> >>>>>> >> >> >> „/home/babak/Downloads/hadoop/bin/../logs“
>> >>>>>> >> >> >> nicht
>> >>>>>> >> >> >> anlegen: Keine Berechtigung
>> >>>>>> >> >> >>
>> >>>>>> >> >> >> this out put(it's in german and it means no right to make
>> >>>>>> >> >> >> this
>> >>>>>> >> >> >> folder)
>> >>>>>> >> >> >>
>> >>>>>> >> >> >>
>> >>>>>> >> >> >> On Wed, Jun 6, 2012 at 7:59 PM, Mohammad Tariq
>> >>>>>> >> >> >> <donta...@gmail.com>
>> >>>>>> >> >> >> wrote:
>> >>>>>> >> >> >>>
>> >>>>>> >> >> >>> once we are done with the configuration, we need to
>> >>>>>> >> >> >>> format
>> >>>>>> >> >> >>> the file
>> >>>>>> >> >> >>> system..use this command to do that-
>> >>>>>> >> >> >>> bin/hadoop namenode -format
>> >>>>>> >> >> >>>
>> >>>>>> >> >> >>> after this, hadoop daemon processes should be started
>> >>>>>> >> >> >>> using
>> >>>>>> >> >> >>> following
>> >>>>>> >> >> >>> commands -
>> >>>>>> >> >> >>> bin/start-dfs.sh (it'll start NN & DN)
>> >>>>>> >> >> >>> bin/start-mapred.sh (it'll start JT & TT)
>> >>>>>> >> >> >>>
>> >>>>>> >> >> >>> after this use jps to check if everything is alright or
>> >>>>>> >> >> >>> point your
>> >>>>>> >> >> >>> browser to localhost:50070..if you further find any
>> >>>>>> >> >> >>> problem
>> >>>>>> >> >> >>> provide
>> >>>>>> >> >> >>> us
>> >>>>>> >> >> >>> with the error logs..:)
>> >>>>>> >> >> >>>
>> >>>>>> >> >> >>> Regards,
>> >>>>>> >> >> >>>     Mohammad Tariq
>> >>>>>> >> >> >>>
>> >>>>>> >> >> >>>
>> >>>>>> >> >> >>> On Wed, Jun 6, 2012 at 11:22 PM, Babak Bastan
>> >>>>>> >> >> >>> <babak...@gmail.com>
>> >>>>>> >> >> >>> wrote:
>> >>>>>> >> >> >>> > were you able to format hdfs properly???
>> >>>>>> >> >> >>> > I did'nt get your question,Do you mean HADOOP_HOME? or
>> >>>>>> >> >> >>> > where did
>> >>>>>> >> >> >>> > I
>> >>>>>> >> >> >>> > install
>> >>>>>> >> >> >>> > Hadoop?
>> >>>>>> >> >> >>> >
>> >>>>>> >> >> >>> > On Wed, Jun 6, 2012 at 7:49 PM, Mohammad Tariq
>> >>>>>> >> >> >>> > <donta...@gmail.com>
>> >>>>>> >> >> >>> > wrote:
>> >>>>>> >> >> >>> >>
>> >>>>>> >> >> >>> >> if you are getting only this, it means your hadoop is
>> >>>>>> >> >> >>> >> not
>> >>>>>> >> >> >>> >> running..were you able to format hdfs properly???
>> >>>>>> >> >> >>> >>
>> >>>>>> >> >> >>> >> Regards,
>> >>>>>> >> >> >>> >>     Mohammad Tariq
>> >>>>>> >> >> >>> >>
>> >>>>>> >> >> >>> >>
>> >>>>>> >> >> >>> >> On Wed, Jun 6, 2012 at 11:17 PM, Babak Bastan
>> >>>>>> >> >> >>> >> <babak...@gmail.com>
>> >>>>>> >> >> >>> >> wrote:
>> >>>>>> >> >> >>> >> > Hi MohammadmI irun jps in my shel I can see this
>> >>>>>> >> >> >>> >> > result:
>> >>>>>> >> >> >>> >> > 2213 Jps
>> >>>>>> >> >> >>> >> >
>> >>>>>> >> >> >>> >> >
>> >>>>>> >> >> >>> >> > On Wed, Jun 6, 2012 at 7:44 PM, Mohammad Tariq
>> >>>>>> >> >> >>> >> > <donta...@gmail.com>
>> >>>>>> >> >> >>> >> > wrote:
>> >>>>>> >> >> >>> >> >>
>> >>>>>> >> >> >>> >> >> you can also use "jps" command at your shell to see
>> >>>>>> >> >> >>> >> >> whether
>> >>>>>> >> >> >>> >> >> Hadoop
>> >>>>>> >> >> >>> >> >> processes are running or not.
>> >>>>>> >> >> >>> >> >>
>> >>>>>> >> >> >>> >> >> Regards,
>> >>>>>> >> >> >>> >> >>     Mohammad Tariq
>> >>>>>> >> >> >>> >> >>
>> >>>>>> >> >> >>> >> >>
>> >>>>>> >> >> >>> >> >> On Wed, Jun 6, 2012 at 11:12 PM, Mohammad Tariq
>> >>>>>> >> >> >>> >> >> <donta...@gmail.com>
>> >>>>>> >> >> >>> >> >> wrote:
>> >>>>>> >> >> >>> >> >> > Hi Babak,
>> >>>>>> >> >> >>> >> >> >
>> >>>>>> >> >> >>> >> >> >  You have to type it in you web browser..Hadoop
>> >>>>>> >> >> >>> >> >> > provides us
>> >>>>>> >> >> >>> >> >> > a
>> >>>>>> >> >> >>> >> >> > web
>> >>>>>> >> >> >>> >> >> > GUI
>> >>>>>> >> >> >>> >> >> > that not only allows us to browse through the
>> >>>>>> >> >> >>> >> >> > file
>> >>>>>> >> >> >>> >> >> > system,
>> >>>>>> >> >> >>> >> >> > but
>> >>>>>> >> >> >>> >> >> > to
>> >>>>>> >> >> >>> >> >> > download the files as well..Apart from that it
>> >>>>>> >> >> >>> >> >> > also
>> >>>>>> >> >> >>> >> >> > provides a
>> >>>>>> >> >> >>> >> >> > web
>> >>>>>> >> >> >>> >> >> > GUI
>> >>>>>> >> >> >>> >> >> > that can be used to see the status of Jobtracker
>> >>>>>> >> >> >>> >> >> > and
>> >>>>>> >> >> >>> >> >> > Tasktracker..When
>> >>>>>> >> >> >>> >> >> > you run a Hive or Pig job or a Mapreduce job, you
>> >>>>>> >> >> >>> >> >> > can point
>> >>>>>> >> >> >>> >> >> > your
>> >>>>>> >> >> >>> >> >> > browser to http://localhost:50030 to see the
>> >>>>>> >> >> >>> >> >> > status
>> >>>>>> >> >> >>> >> >> > and
>> >>>>>> >> >> >>> >> >> > logs
>> >>>>>> >> >> >>> >> >> > of
>> >>>>>> >> >> >>> >> >> > your
>> >>>>>> >> >> >>> >> >> > job.
>> >>>>>> >> >> >>> >> >> >
>> >>>>>> >> >> >>> >> >> > Regards,
>> >>>>>> >> >> >>> >> >> >     Mohammad Tariq
>> >>>>>> >> >> >>> >> >> >
>> >>>>>> >> >> >>> >> >> >
>> >>>>>> >> >> >>> >> >> > On Wed, Jun 6, 2012 at 8:28 PM, Babak Bastan
>> >>>>>> >> >> >>> >> >> > <babak...@gmail.com>
>> >>>>>> >> >> >>> >> >> > wrote:
>> >>>>>> >> >> >>> >> >> >> Thank you shashwat for the answer,
>> >>>>>> >> >> >>> >> >> >> where should I type http://localhost:50070?
>> >>>>>> >> >> >>> >> >> >> I typed here: hive>http://localhost:50070 but
>> >>>>>> >> >> >>> >> >> >> nothing as
>> >>>>>> >> >> >>> >> >> >> result
>> >>>>>> >> >> >>> >> >> >>
>> >>>>>> >> >> >>> >> >> >>
>> >>>>>> >> >> >>> >> >> >> On Wed, Jun 6, 2012 at 3:32 PM, shashwat
>> >>>>>> >> >> >>> >> >> >> shriparv
>> >>>>>> >> >> >>> >> >> >> <dwivedishash...@gmail.com> wrote:
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>> first type http://localhost:50070 whether this
>> >>>>>> >> >> >>> >> >> >>> is
>> >>>>>> >> >> >>> >> >> >>> opening
>> >>>>>> >> >> >>> >> >> >>> or
>> >>>>>> >> >> >>> >> >> >>> not
>> >>>>>> >> >> >>> >> >> >>> and
>> >>>>>> >> >> >>> >> >> >>> check
>> >>>>>> >> >> >>> >> >> >>> how many nodes are available, check some of the
>> >>>>>> >> >> >>> >> >> >>> hadoop
>> >>>>>> >> >> >>> >> >> >>> shell
>> >>>>>> >> >> >>> >> >> >>> commands
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>> from http://hadoop.apache.org/common/docs/r0.18.3/hdfs_shell.html
>> >>>>>> >> >> >>> >> >> >>> run
>> >>>>>> >> >> >>> >> >> >>> example mapreduce task on hadoop take example
>> >>>>>> >> >> >>> >> >> >>> from
>> >>>>>> >> >> >>> >> >> >>> here
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>> : http://www.michael-noll.com/blog/2011/04/09/benchmarking-and-stress-testing-an-hadoop-cluster-with-terasort-testdfsio-nnbench-mrbench/
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>> if all the above you can do sucessfully means
>> >>>>>> >> >> >>> >> >> >>> hadoop is
>> >>>>>> >> >> >>> >> >> >>> configured
>> >>>>>> >> >> >>> >> >> >>> correctly
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>> Regards
>> >>>>>> >> >> >>> >> >> >>> Shashwat
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>> On Wed, Jun 6, 2012 at 1:30 AM, Babak Bastan
>> >>>>>> >> >> >>> >> >> >>> <babak...@gmail.com>
>> >>>>>> >> >> >>> >> >> >>> wrote:
>> >>>>>> >> >> >>> >> >> >>>>
>> >>>>>> >> >> >>> >> >> >>>> no I'm not working on CDH.Is there a way to
>> >>>>>> >> >> >>> >> >> >>>> test
>> >>>>>> >> >> >>> >> >> >>>> if my
>> >>>>>> >> >> >>> >> >> >>>> Hadoop
>> >>>>>> >> >> >>> >> >> >>>> works
>> >>>>>> >> >> >>> >> >> >>>> fine
>> >>>>>> >> >> >>> >> >> >>>> or not?
>> >>>>>> >> >> >>> >> >> >>>>
>> >>>>>> >> >> >>> >> >> >>>>
>> >>>>>> >> >> >>> >> >> >>>> On Tue, Jun 5, 2012 at 9:55 PM, Bejoy KS
>> >>>>>> >> >> >>> >> >> >>>> <bejoy...@yahoo.com>
>> >>>>>> >> >> >>> >> >> >>>> wrote:
>> >>>>>> >> >> >>> >> >> >>>>>
>> >>>>>> >> >> >>> >> >> >>>>> Hi Babak
>> >>>>>> >> >> >>> >> >> >>>>>
>> >>>>>> >> >> >>> >> >> >>>>> You gotta follow those instructions in the
>> >>>>>> >> >> >>> >> >> >>>>> apace
>> >>>>>> >> >> >>> >> >> >>>>> site
>> >>>>>> >> >> >>> >> >> >>>>> to
>> >>>>>> >> >> >>> >> >> >>>>> set
>> >>>>>> >> >> >>> >> >> >>>>> up
>> >>>>>> >> >> >>> >> >> >>>>> hadoop
>> >>>>>> >> >> >>> >> >> >>>>> from scratch and ensure that hdfs is working
>> >>>>>> >> >> >>> >> >> >>>>> first. You
>> >>>>>> >> >> >>> >> >> >>>>> should
>> >>>>>> >> >> >>> >> >> >>>>> be
>> >>>>>> >> >> >>> >> >> >>>>> able to
>> >>>>>> >> >> >>> >> >> >>>>> read and write files to hdfs before you do
>> >>>>>> >> >> >>> >> >> >>>>> your
>> >>>>>> >> >> >>> >> >> >>>>> next
>> >>>>>> >> >> >>> >> >> >>>>> steps.
>> >>>>>> >> >> >>> >> >> >>>>>
>> >>>>>> >> >> >>> >> >> >>>>> Are you on CDH or apache distribution of
>> >>>>>> >> >> >>> >> >> >>>>> hadoop?
>> >>>>>> >> >> >>> >> >> >>>>> If it
>> >>>>>> >> >> >>> >> >> >>>>> is
>> >>>>>> >> >> >>> >> >> >>>>> CDH
>> >>>>>> >> >> >>> >> >> >>>>> there
>> >>>>>> >> >> >>> >> >> >>>>> are
>> >>>>>> >> >> >>> >> >> >>>>> detailed instructions on Cloudera web site.
>> >>>>>> >> >> >>> >> >> >>>>>
>> >>>>>> >> >> >>> >> >> >>>>> Regards
>> >>>>>> >> >> >>> >> >> >>>>> Bejoy KS
>> >>>>>> >> >> >>> >> >> >>>>>
>> >>>>>> >> >> >>> >> >> >>>>> Sent from handheld, please excuse typos.
>> >>>>>> >> >> >>> >> >> >>>>> ________________________________
>> >>>>>> >> >> >>> >> >> >>>>> From: Babak Bastan <babak...@gmail.com>
>> >>>>>> >> >> >>> >> >> >>>>> Date: Tue, 5 Jun 2012 21:30:22 +0200
>> >>>>>> >> >> >>> >> >> >>>>> To: <user@hive.apache.org>
>> >>>>>> >> >> >>> >> >> >>>>> ReplyTo: user@hive.apache.org
>> >>>>>> >> >> >>> >> >> >>>>> Subject: Re: Error while Creating Table in
>> >>>>>> >> >> >>> >> >> >>>>> Hive
>> >>>>>> >> >> >>> >> >> >>>>>
>> >>>>>> >> >> >>> >> >> >>>>> @Bejoy: I set the fs.default.name in the
>> >>>>>> >> >> >>> >> >> >>>>> core-site.xml
>> >>>>>> >> >> >>> >> >> >>>>> and
>> >>>>>> >> >> >>> >> >> >>>>> I
>> >>>>>> >> >> >>> >> >> >>>>> did
>> >>>>>> >> >> >>> >> >> >>>>> all
>> >>>>>> >> >> >>> >> >> >>>>> of
>> >>>>>> >> >> >>> >> >> >>>>> thing that was mentioned in the reference but
>> >>>>>> >> >> >>> >> >> >>>>> no
>> >>>>>> >> >> >>> >> >> >>>>> effect
>> >>>>>> >> >> >>> >> >> >>>>>
>> >>>>>> >> >> >>> >> >> >>>>> On Tue, Jun 5, 2012 at 8:43 PM, Babak Bastan
>> >>>>>> >> >> >>> >> >> >>>>> <babak...@gmail.com>
>> >>>>>> >> >> >>> >> >> >>>>> wrote:
>> >>>>>> >> >> >>> >> >> >>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>> Ok sorry but that was my Mistake .I thought
>> >>>>>> >> >> >>> >> >> >>>>>> it
>> >>>>>> >> >> >>> >> >> >>>>>> works
>> >>>>>> >> >> >>> >> >> >>>>>> but
>> >>>>>> >> >> >>> >> >> >>>>>> no.
>> >>>>>> >> >> >>> >> >> >>>>>> I wrote the command without ; and then I
>> >>>>>> >> >> >>> >> >> >>>>>> think
>> >>>>>> >> >> >>> >> >> >>>>>> It
>> >>>>>> >> >> >>> >> >> >>>>>> works
>> >>>>>> >> >> >>> >> >> >>>>>> but
>> >>>>>> >> >> >>> >> >> >>>>>> with
>> >>>>>> >> >> >>> >> >> >>>>>> ;
>> >>>>>> >> >> >>> >> >> >>>>>> at
>> >>>>>> >> >> >>> >> >> >>>>>> the end of command
>> >>>>>> >> >> >>> >> >> >>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>> CREATE TABLE pokes (foo INT, bar STRING);
>> >>>>>> >> >> >>> >> >> >>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>> does'nt work
>> >>>>>> >> >> >>> >> >> >>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>> On Tue, Jun 5, 2012 at 8:34 PM, shashwat
>> >>>>>> >> >> >>> >> >> >>>>>> shriparv
>> >>>>>> >> >> >>> >> >> >>>>>> <dwivedishash...@gmail.com> wrote:
>> >>>>>> >> >> >>> >> >> >>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>> inside configuration. all properties will
>> >>>>>> >> >> >>> >> >> >>>>>>> be
>> >>>>>> >> >> >>> >> >> >>>>>>> inside
>> >>>>>> >> >> >>> >> >> >>>>>>> the
>> >>>>>> >> >> >>> >> >> >>>>>>> configuration
>> >>>>>> >> >> >>> >> >> >>>>>>> tags
>> >>>>>> >> >> >>> >> >> >>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>> On Tue, Jun 5, 2012 at 11:53 PM, Babak
>> >>>>>> >> >> >>> >> >> >>>>>>> Bastan
>> >>>>>> >> >> >>> >> >> >>>>>>> <babak...@gmail.com>
>> >>>>>> >> >> >>> >> >> >>>>>>> wrote:
>> >>>>>> >> >> >>> >> >> >>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>> Thank you so much my friend your idee
>> >>>>>> >> >> >>> >> >> >>>>>>>> works
>> >>>>>> >> >> >>> >> >> >>>>>>>> fine(no
>> >>>>>> >> >> >>> >> >> >>>>>>>> error)
>> >>>>>> >> >> >>> >> >> >>>>>>>> you
>> >>>>>> >> >> >>> >> >> >>>>>>>> are
>> >>>>>> >> >> >>> >> >> >>>>>>>> the best :)
>> >>>>>> >> >> >>> >> >> >>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>> On Tue, Jun 5, 2012 at 8:20 PM, Babak
>> >>>>>> >> >> >>> >> >> >>>>>>>> Bastan
>> >>>>>> >> >> >>> >> >> >>>>>>>> <babak...@gmail.com>
>> >>>>>> >> >> >>> >> >> >>>>>>>> wrote:
>> >>>>>> >> >> >>> >> >> >>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>> It must be inside the
>> >>>>>> >> >> >>> >> >> >>>>>>>>> <configuration></configuration>
>> >>>>>> >> >> >>> >> >> >>>>>>>>> or
>> >>>>>> >> >> >>> >> >> >>>>>>>>> outside
>> >>>>>> >> >> >>> >> >> >>>>>>>>> this?
>> >>>>>> >> >> >>> >> >> >>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>> On Tue, Jun 5, 2012 at 8:15 PM, shashwat
>> >>>>>> >> >> >>> >> >> >>>>>>>>> shriparv
>> >>>>>> >> >> >>> >> >> >>>>>>>>> <dwivedishash...@gmail.com> wrote:
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>> It will be inside hive/conf
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>> On Tue, Jun 5, 2012 at 11:43 PM, Babak
>> >>>>>> >> >> >>> >> >> >>>>>>>>>> Bastan
>> >>>>>> >> >> >>> >> >> >>>>>>>>>> <babak...@gmail.com>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>> wrote:
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>> Thanks sShashwat, and where is this
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>> hive-site.xml
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>> On Tue, Jun 5, 2012 at 8:02 PM,
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>> shashwat
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>> shriparv
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>> <dwivedishash...@gmail.com> wrote:
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> set
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> hive.metastore.warehouse.dir in
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> hive-site.xml
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> <property>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>   <name>hive.metastore.local</name>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>   <value>true</value>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> </property>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> <name>hive.metastore.warehouse.dir</name>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>                <value>/home/<your
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> username>/hivefolder</value>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>                <description>location
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> of
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> default
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> database
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> for
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> the
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> warehouse</description>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>        </property>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> On Tue, Jun 5, 2012 at 10:43 PM, Babak
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> Bastan
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> <babak...@gmail.com> wrote:
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> Hello Experts ,
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> I'm new in Hive .When try to create a
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> test
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> Table
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> in
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> Hive
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> I
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> get
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> an error.I want to run this command:
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> CREATE TABLE Test (DateT STRING, Url
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> STRING,
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> Content
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> STRING);
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> but this error occured:
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> FAILED: Error in metadata:
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> MetaException(message:Got
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> exception:
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> java.io.FileNotFoundException File
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> file:/user/hive/warehouse/test does
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> not
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> exist.)
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> FAILED: Execution Error, return code
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> 1
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> from
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> org.apache.hadoop.hive.ql.exec.DDLTask
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> How can I solve this Problem?
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> Thank you so much
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> --
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> ∞
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> Shashwat Shriparv
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>> --
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>> ∞
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>> Shashwat Shriparv
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>> --
>> >>>>>> >> >> >>> >> >> >>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>> ∞
>> >>>>>> >> >> >>> >> >> >>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>> Shashwat Shriparv
>> >>>>>> >> >> >>> >> >> >>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>>
>> >>>>>> >> >> >>> >> >> >>>>>
>> >>>>>> >> >> >>> >> >> >>>>
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>> --
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>> ∞
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>> Shashwat Shriparv
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>>
>> >>>>>> >> >> >>> >> >> >>
>> >>>>>> >> >> >>> >> >
>> >>>>>> >> >> >>> >> >
>> >>>>>> >> >> >>> >
>> >>>>>> >> >> >>> >
>> >>>>>> >> >> >>
>> >>>>>> >> >> >>
>> >>>>>> >> >
>> >>>>>> >> >
>> >>>>>> >
>> >>>>>> >
>> >>>>>
>> >>>>>
>> >>>>
>> >>>>
>> >>>>
>> >>>> --
>> >>>>
>> >>>>
>> >>>> ∞
>> >>>>
>> >>>> Shashwat Shriparv
>> >>>>
>> >>>>
>> >>>
>> >>
>> >
>
>

Reply via email to