Recursive delete for an S3 directory does not actually delete files or 
subdirectories
-------------------------------------------------------------------------------------

                 Key: HADOOP-880
                 URL: https://issues.apache.org/jira/browse/HADOOP-880
             Project: Hadoop
          Issue Type: Bug
          Components: fs
    Affects Versions: 0.10.0
            Reporter: Tom White
         Assigned To: Tom White


Here is the bug report from Michael Stack:

Here I'm listing a BUCKET directory that was copied up using 'hadoop
fs', then rmr'ing it and then listing again:

[EMAIL PROTECTED]:~/checkouts/hadoop$  ./bin/hadoop fs -fs
s3://ID:[EMAIL PROTECTED] -ls /fromfile
Found 2 items
/fromfile/diff.txt      <r 1>   591
/fromfile/x.js  <r 1>   2477
[EMAIL PROTECTED]:~/checkouts/hadoop$  ./bin/hadoop fs -fs
s3://ID:[EMAIL PROTECTED] -rmr /fromfile
Deleted /fromfile
[EMAIL PROTECTED]:~/checkouts/hadoop$  ./bin/hadoop fs -fs
s3://ID:[EMAIL PROTECTED] -ls /fromfile
Found 0 items

The '0 items' is odd because, now, listing my BUCKET using a tool other
than 'hadoop fs' (i.e. hanzo webs python scripts):

[EMAIL PROTECTED]:~/checkouts/hadoop.trunk$ s3ls BUCKET
%2F
%2Ffromfile%2F.diff.txt.crc
%2Ffromfile%2F.x.js.crc
%2Ffromfile%2Fdiff.txt
%2Ffromfile%2Fx.js
block_-5013142890590722396
block_5832002498000415319
block_6889488315428893905
block_9120115089645350905

Its all still there still.  I can subsequently do the likes of the
following:

[EMAIL PROTECTED]:~/checkouts/hadoop$  ./bin/hadoop fs -fs
s3://ID:[EMAIL PROTECTED] -rmr /fromfile/diff.txt

... and the delete will succeed and looking at the bucket with alternate
tools shows that it has actually been remove, and so on up the hierarchy.

-- 
This message is automatically generated by JIRA.
-
If you think it was sent incorrectly contact one of the administrators: 
https://issues.apache.org/jira/secure/Administrators.jspa
-
For more information on JIRA, see: http://www.atlassian.com/software/jira

        

Reply via email to