[ 
https://issues.apache.org/jira/browse/JCR-2063?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12834216#action_12834216
 ] 

Thomas Mueller commented on JCR-2063:
-------------------------------------

A workaround for implementations where this is not fixed is:

gc.mark();
try {
    // sleep to ensure the last modified time is updated
    // even for file system with a lower time resolution
    Thread.sleep(5000);
} catch (Exception e) {
    // can not ignore, otherwise data that is in use may be deleted
    throw new RepositoryException("Interrupted");
}
gc.mark();



> FileDataStore: garbage collection can delete files that are still needed
> ------------------------------------------------------------------------
>
>                 Key: JCR-2063
>                 URL: https://issues.apache.org/jira/browse/JCR-2063
>             Project: Jackrabbit Content Repository
>          Issue Type: Bug
>          Components: jackrabbit-core
>            Reporter: Thomas Mueller
>            Assignee: Thomas Mueller
>             Fix For: 1.5.5
>
>
> It looks like the FileDataStore garbage collection (both regular scan and 
> persistence manager scan) can delete files that are still needed.
> Currently it looks like the reason is the last access time resolution of the 
> operating system. This is 2 seconds for FAT and Mac OS X, NTFS 100 ns, and 1 
> second for other file systems. That means file that are scanned at the very 
> beginning are sometimes deleted, because they have a later last modified time 
> then when the scan was started.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to