Thanks!
However, in that case, how do I handle the current data that is already in NN?
Do I have to reformat HDFS?
Thanks & regards
Arko

On Wed, Feb 13, 2013 at 2:06 PM, Mohammad Tariq <donta...@gmail.com> wrote:
> Just don't run the DN daemon on that machine. Dedicate that machine to NN.
> Remove the hostname of the NN machine from the 'slaves' file.
>
> Warm Regards,
> Tariq
> https://mtariq.jux.com/
> cloudfront.blogspot.com
>
>
> On Thu, Feb 14, 2013 at 1:31 AM, Arko Provo Mukherjee
> <arkoprovomukher...@gmail.com> wrote:
>>
>> Hi,
>>
>> Thanks for the help!
>>
>> However, I am still unsure about how to "turn off" the datanode
>> feature on the NN.
>>
>> I checked the "hdfs-site.xml" file and the dfs.data.dir is pointed to
>> a directory.
>>
>> Should I just comment out that property? What would happen to the
>> current data that there is in the Master? Will it get removed
>> automatically?
>>
>> Thanks & regards
>> Arko
>>
>>
>>
>> On Wed, Feb 13, 2013 at 1:55 PM, Mohammad Tariq <donta...@gmail.com>
>> wrote:
>> > You can specify the logging level as specified by Charles. But turning
>> > logs
>> > off is never a good idea. Logs are really helpful in problem diagnosis,
>> > which are eventual.
>> >
>> > Warm Regards,
>> > Tariq
>> > https://mtariq.jux.com/
>> > cloudfront.blogspot.com
>> >
>> >
>> > On Thu, Feb 14, 2013 at 1:22 AM, Arko Provo Mukherjee
>> > <arkoprovomukher...@gmail.com> wrote:
>> >>
>> >> Hi,
>> >>
>> >> Yeah, my NameNode is also seconding as a DataNode.
>> >>
>> >> I would like to "turn off" this feature.
>> >>
>> >> Request help regarding the same.
>> >>
>> >> Thanks & regards
>> >> Arko
>> >>
>> >> On Wed, Feb 13, 2013 at 1:38 PM, Charles Baker <cba...@sdl.com> wrote:
>> >> > Hi Arko. Sounds like you may be running a DataNode on the NameNode
>> >> > which
>> >> > is
>> >> > not recommended practice. Normally, the only files the NN stores are
>> >> > the
>> >> > image and edits files. It does not store any actual HDFS data. If you
>> >> > must
>> >> > run a DN on the NN, try turning down the logging in
>> >> > /conf/log4j.properties:
>> >> >
>> >> > #hadoop.root.logger=INFO,console
>> >> > #hadoop.root.logger=WARN,console
>> >> > hadoop.root.logger=ERROR,console
>> >> >
>> >> > Depending on the logging information you require, of course.
>> >> >
>> >> > -Chuck
>> >> >
>> >> >
>> >> > -----Original Message-----
>> >> > From: Arko Provo Mukherjee [mailto:arkoprovomukher...@gmail.com]
>> >> > Sent: Wednesday, February 13, 2013 11:32 AM
>> >> > To: hdfs-user@hadoop.apache.org
>> >> > Subject: Managing space in Master Node
>> >> >
>> >> > Hello Gurus,
>> >> >
>> >> > I am managing a Hadoop Cluster to run some experiments.
>> >> >
>> >> > The issue I am continuously facing is that the Master Node runs out
>> >> > of
>> >> > disk
>> >> > space due to logs and data files.
>> >> >
>> >> > I can monitor and delete log files. However, I cannot delete the HDFS
>> >> > data.
>> >> >
>> >> > Thus, is there a way to force Hadoop not to save any HDFS data in the
>> >> > Master
>> >> > Node?
>> >> >
>> >> > Then I can use my master to handle the metadata only and store the
>> >> > logs.
>> >> >
>> >> > Thanks & regards
>> >> > Arko
>> >> > SDL Enterprise Technologies, Inc. - all rights reserved.  The
>> >> > information contained in this email may be confidential and/or
>> >> > legally
>> >> > privileged. It has been sent for the sole use of the intended
>> >> > recipient(s).
>> >> > If you are not the intended recipient of this mail, you are hereby
>> >> > notified
>> >> > that any unauthorized review, use, disclosure, dissemination,
>> >> > distribution,
>> >> > or copying of this communication, or any of its contents, is strictly
>> >> > prohibited. If you have received this communication in error, please
>> >> > reply
>> >> > to the sender and destroy all copies of the message.
>> >> > Registered address: 201 Edgewater Drive, Suite 225, Wakefield, MA
>> >> > 01880,
>> >> > USA
>> >> >
>> >
>> >
>
>

Reply via email to