On Tue 20-12-16 15:23:06, Ross Zwisler wrote:
> Currently dax_mapping_entry_mkclean() fails to clean and write protect the
> pmd_t of a DAX PMD entry during an *sync operation.  This can result in
> data loss in the following sequence:
> 
> 1) mmap write to DAX PMD, dirtying PMD radix tree entry and making the
>    pmd_t dirty and writeable
> 2) fsync, flushing out PMD data and cleaning the radix tree entry. We
>    currently fail to mark the pmd_t as clean and write protected.
> 3) more mmap writes to the PMD.  These don't cause any page faults since
>    the pmd_t is dirty and writeable.  The radix tree entry remains clean.
> 4) fsync, which fails to flush the dirty PMD data because the radix tree
>    entry was clean.
> 5) crash - dirty data that should have been fsync'd as part of 4) could
>    still have been in the processor cache, and is lost.
> 
> Fix this by marking the pmd_t clean and write protected in
> dax_mapping_entry_mkclean(), which is called as part of the fsync
> operation 2).  This will cause the writes in step 3) above to generate page
> faults where we'll re-dirty the PMD radix tree entry, resulting in flushes
> in the fsync that happens in step 4).
> 
> Signed-off-by: Ross Zwisler <ross.zwis...@linux.intel.com>
> Cc: Jan Kara <j...@suse.cz>
> Fixes: 4b4bb46d00b3 ("dax: clear dirty entry tags on cache flush")

Yeah, good catch. The patch looks good. You can add:

Reviewed-by: Jan Kara <j...@suse.cz>

                                                                Honza


> ---
>  fs/dax.c           | 51 ++++++++++++++++++++++++++++++++++++---------------
>  include/linux/mm.h |  2 --
>  mm/memory.c        |  4 ++--
>  3 files changed, 38 insertions(+), 19 deletions(-)
> 
> diff --git a/fs/dax.c b/fs/dax.c
> index 5c74f60..ddcddfe 100644
> --- a/fs/dax.c
> +++ b/fs/dax.c
> @@ -691,8 +691,8 @@ static void dax_mapping_entry_mkclean(struct 
> address_space *mapping,
>                                     pgoff_t index, unsigned long pfn)
>  {
>       struct vm_area_struct *vma;
> -     pte_t *ptep;
> -     pte_t pte;
> +     pte_t pte, *ptep = NULL;
> +     pmd_t *pmdp = NULL;
>       spinlock_t *ptl;
>       bool changed;
>  
> @@ -707,21 +707,42 @@ static void dax_mapping_entry_mkclean(struct 
> address_space *mapping,
>  
>               address = pgoff_address(index, vma);
>               changed = false;
> -             if (follow_pte(vma->vm_mm, address, &ptep, &ptl))
> +             if (follow_pte_pmd(vma->vm_mm, address, &ptep, &pmdp, &ptl))
>                       continue;
> -             if (pfn != pte_pfn(*ptep))
> -                     goto unlock;
> -             if (!pte_dirty(*ptep) && !pte_write(*ptep))
> -                     goto unlock;
>  
> -             flush_cache_page(vma, address, pfn);
> -             pte = ptep_clear_flush(vma, address, ptep);
> -             pte = pte_wrprotect(pte);
> -             pte = pte_mkclean(pte);
> -             set_pte_at(vma->vm_mm, address, ptep, pte);
> -             changed = true;
> -unlock:
> -             pte_unmap_unlock(ptep, ptl);
> +             if (pmdp) {
> +#ifdef CONFIG_FS_DAX_PMD
> +                     pmd_t pmd;
> +
> +                     if (pfn != pmd_pfn(*pmdp))
> +                             goto unlock_pmd;
> +                     if (!pmd_dirty(*pmdp) && !pmd_write(*pmdp))
> +                             goto unlock_pmd;
> +
> +                     flush_cache_page(vma, address, pfn);
> +                     pmd = pmdp_huge_clear_flush(vma, address, pmdp);
> +                     pmd = pmd_wrprotect(pmd);
> +                     pmd = pmd_mkclean(pmd);
> +                     set_pmd_at(vma->vm_mm, address, pmdp, pmd);
> +                     changed = true;
> +unlock_pmd:
> +                     spin_unlock(ptl);
> +#endif
> +             } else {
> +                     if (pfn != pte_pfn(*ptep))
> +                             goto unlock_pte;
> +                     if (!pte_dirty(*ptep) && !pte_write(*ptep))
> +                             goto unlock_pte;
> +
> +                     flush_cache_page(vma, address, pfn);
> +                     pte = ptep_clear_flush(vma, address, ptep);
> +                     pte = pte_wrprotect(pte);
> +                     pte = pte_mkclean(pte);
> +                     set_pte_at(vma->vm_mm, address, ptep, pte);
> +                     changed = true;
> +unlock_pte:
> +                     pte_unmap_unlock(ptep, ptl);
> +             }
>  
>               if (changed)
>                       mmu_notifier_invalidate_page(vma->vm_mm, address);
> diff --git a/include/linux/mm.h b/include/linux/mm.h
> index ff0e1c1..f4de7fa 100644
> --- a/include/linux/mm.h
> +++ b/include/linux/mm.h
> @@ -1210,8 +1210,6 @@ int copy_page_range(struct mm_struct *dst, struct 
> mm_struct *src,
>                       struct vm_area_struct *vma);
>  void unmap_mapping_range(struct address_space *mapping,
>               loff_t const holebegin, loff_t const holelen, int even_cows);
> -int follow_pte(struct mm_struct *mm, unsigned long address, pte_t **ptepp,
> -            spinlock_t **ptlp);
>  int follow_pte_pmd(struct mm_struct *mm, unsigned long address,
>                            pte_t **ptepp, pmd_t **pmdpp, spinlock_t **ptlp);
>  int follow_pfn(struct vm_area_struct *vma, unsigned long address,
> diff --git a/mm/memory.c b/mm/memory.c
> index 29edd91..ddcf979 100644
> --- a/mm/memory.c
> +++ b/mm/memory.c
> @@ -3826,8 +3826,8 @@ static int __follow_pte_pmd(struct mm_struct *mm, 
> unsigned long address,
>       return -EINVAL;
>  }
>  
> -int follow_pte(struct mm_struct *mm, unsigned long address, pte_t **ptepp,
> -            spinlock_t **ptlp)
> +static inline int follow_pte(struct mm_struct *mm, unsigned long address,
> +                          pte_t **ptepp, spinlock_t **ptlp)
>  {
>       int res;
>  
> -- 
> 2.7.4
> 
-- 
Jan Kara <j...@suse.com>
SUSE Labs, CR

Reply via email to