Hi,
When NN is in safe mode, you get a read-only view of the hadoop file system. ( 
since NN is reconstructing its image of FS )
Use  "hadoop dfsadmin -safemode get" to check if in safe mode.
"hadoop dfsadmin -safemode leave" to leave safe mode forcefully. Or use "hadoop 
dfsadmin -safemode wait" to block till NN leaves by itself.

Amogh


On 1/19/10 10:31 AM, "prasenjit mukherjee" <prasen....@gmail.com> wrote:

Hmmm.  I am actually running it from a batch file. Is "hadoop fs -rmr"
not that stable compared to pig's rm OR hadoop's FileSystem ?

Let me try your suggestion by writing a cleanup script in pig.

-Thanks,
Prasen

On Tue, Jan 19, 2010 at 10:25 AM, Rekha Joshi <rekha...@yahoo-inc.com> wrote:
> Can you try with dfs/ without quotes?If using pig to run jobs you can use rmf 
> within your script(again w/o quotes) to force remove and avoid error if 
> file/dir not present.Or if doing this inside hadoop job, you can use 
> FileSystem/FileStatus to delete directories.HTH.
> Cheers,
> /R
>
> On 1/19/10 10:15 AM, "prasenjit mukherjee" <prasen....@gmail.com> wrote:
>
> "hadoop fs -rmr /op"
>
> That command always fails. I am trying to run sequential hadoop jobs.
> After the first run all subsequent runs fail while cleaning up ( aka
> removing the hadoop dir created by previous run ). What can I do to
> avoid this ?
>
> here is my hadoop version :
> # hadoop version
> Hadoop 0.20.0
> Subversion https://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.20
> -r 763504
> Compiled by ndaley on Thu Apr  9 05:18:40 UTC 2009
>
> Any help is greatly appreciated.
>
> -Prasen
>
>

Reply via email to