I'm a little confused as to why you're putting
/pseg/local /...
as the location.

Are you sure that you've been given a folder name at the root of the
drive called /pseg/ ?
Maybe try to ssh to your server and navigate to your datastore folder,
then do "pwd".

That should give you the working directory of the datastore.  Use that
as the value for the tmp datastore location.

Sorry if that seems like a stupid suggestion.  Just trying to get a
handle on your actual problem.  My linux skillset is limited to the
basics, so I'm troubleshooting by looking for the type of mistake that
I would make.

If the above is not the issue, then I'm not sure what the issue could
be.  But, I'd be glad to continue trying to help (with my limited
knowledge) :-)


On Wed, Mar 30, 2011 at 8:37 AM, ke xie <oed...@gmail.com> wrote:
> Thank you modemide for your quick response.
>
> Sorry for not be clear...your understanding is right.
> I have a machine, called grande, and the other called pseg. Now i'm using
> grande as master (by fill the masters file by "grande") and pseg as slave.
>
> the configuration of grande is (core-site.xml)
>
> <property>
>        <name>fs.default.name</name>
>        <value>hdfs://grande:8500</value>
>        <description>The name of the default file system. A URI whose scheme
> and
>  authority determine the FileSystem implementation. </description>
> </property>
> <property>
>  <name>hadoop.tmp.dir</name>
>        <value>/grande/local/xieke-cluster/hadoop-tmp-data/</value>
> <description>. A base for other temporary directories</description>
> </property>
>
> and the configuration of pseg is:
> <property>
>        <name>fs.default.name</name>
>        <value>hdfs://grandonf/8500</value>
>        <description>The name of the default file system. A URI whose scheme
> and
>  authority determine the FileSystem implementation. </description>
> </property>
> <property>
>  <name>hadoop.tmp.dir</name>
>        <value>/pseg/local/xieke-cluster/hadoop-tmp-data/</value>
> <description> A base for other temporary directories.</description>
> </property>
>
>
> just the same as your I think?
>
> then I run ./bin/hadoop namenode -format to format nodes.
> and run ./bin/start-all.sh to start machines. but now:
>
> grande% ./bin/start-all.sh
> starting namenode, logging to
> /grande/local/hadoop/bin/../logs/hadoop-kx19-namenode-grande.out
> *pseg: /grande/local/hadoop/bin/..: No such file or directory.*
> grande: starting datanode, logging to
> /grande/local/hadoop/bin/../logs/hadoop-kx19-datanode-grande.out
> grande: starting secondarynamenode, logging to
> /grande/local/hadoop/bin/../logs/hadoop-kx19-secondarynamenode-grande.out
> starting jobtracker, logging to
> /grande/local/hadoop/bin/../logs/hadoop-kx19-jobtracker-grande.out
> pseg: /grande/local/hadoop/bin/..: No such file or directory.
> grande: starting tasktracker, logging to
> /grande/local/hadoop/bin/../logs/hadoop-kx19-tasktracker-grande.out
>
>
> Any ideas?
>
>
> On Wed, Mar 30, 2011 at 7:54 PM, modemide <modem...@gmail.com> wrote:
>
>> Ok, so if I understand correctly, you want to change the location of
>> the datastore on individual computers.
>>
>> I've tested it on my cluster, and it seems to work.  Just for the sake
>> of troubleshooting, you didn't mention the following:
>> 1) Which computer were you editing the files on
>> 2) which file were you editing?
>>
>>
>> ******************************************************************************
>> Here's my typical DataNode configuration:
>> Computer: DataNode
>> FileName: core-site.xml
>> Contents:
>> ....
>> <name>hadoop.tmp.dir</name>
>> <value>/usr/local/hadoop/datastore/hadoop-${user.name}</value>
>> ...
>>
>> ******************************************************************************
>> Here's the configuration of another DataNode I modified to test what
>> you were asking:
>> Computer: DataNode2
>> FileName: core-site.xml
>> Contents:
>> ....
>> <name>hadoop.tmp.dir</name>
>> <value>/usr/local/hadoop/ANOTHERDATASTORE/hadoop-${user.name}</value>
>> ....
>>
>> ******************************************************************************
>> Then, I moved datastore to ANOTHERDATASTORE on DataNode1.
>>
>> I started my cluster back up, and it worked perfectly.
>>
>>
>> On Wed, Mar 30, 2011 at 6:08 AM, ke xie <oed...@gmail.com> wrote:
>> > Hey guys, I'm new here, and recently I'm working on configuring a cluster
>> > with 32 nodes.
>> >
>> > However, there are some problems, I describe below
>> >
>> > The cluster consists of nodes, which I don't have "root" to configure as
>> I
>> > wish. We only have the space /localhost_name/local space to use.
>> > Thus, we only have
>> >
>> > /machine_a/local
>> > /machine_b/local
>> > ...
>> >
>> > So I guess to set hadoop.tmp.dir=/${HOSTNAME}/local will work, but sadly
>> it
>> > didn't...
>> >
>> > Almost all the tutorials online are trying to set hadoop.tmp.dir as a
>> single
>> > path, which assume on each machine the path is the same... but in my case
>> > it's not...
>> >
>> > I did some googling... like "hadoop.tmp.dir different"... but no
>> results...
>> >
>> > Anybody can help? I'll appreciate that... for i've been working on this
>> > problem for more than 30 hours...
>> >
>> > --
>> > Name: Ke Xie   Eddy
>> > Research Group of Information Retrieval
>> > State Key Laboratory of Intelligent Technology and Systems
>> > Tsinghua University
>> >
>>
>
>
>
> --
> Name: Ke Xie   Eddy
> Research Group of Information Retrieval
> State Key Laboratory of Intelligent Technology and Systems
> Tsinghua University
>

Reply via email to