> -----Original Message----- > From: Ritesh Harjani <ritesh.l...@gmail.com> > Sent: Thursday, April 1, 2021 2:40 PM > Subject: Re: [PATCH v3 05/10] fsdax: Replace mmap entry in case of CoW > > On 21/03/19 09:52AM, Shiyang Ruan wrote: > > We replace the existing entry to the newly allocated one in case of CoW. > > Also, we mark the entry as PAGECACHE_TAG_TOWRITE so writeback marks > > this entry as writeprotected. This helps us snapshots so new write > > pagefaults after snapshots trigger a CoW. > > > > Please correct me here. So the flow is like this. > 1. In case of CoW or a reflinked file, on an mmaped file if write is > attempted, > Then in DAX fault handler code, ->iomap_begin() on a given filesystem will > populate iomap and srcmap. srcmap being from where the read needs to be > attempted from and iomap on where the new write should go to. > 2. So the dax_insert_entry() code as part of the fault handling will take care > of removing the old entry and inserting the new pfn entry to xas and mark > it with PAGECACHE_TAG_TOWRITE so that dax writeback can mark the > entry as > write protected. > Is my above understanding correct?
Yes, you are right. > > > Signed-off-by: Goldwyn Rodrigues <rgold...@suse.com> > > Signed-off-by: Shiyang Ruan <ruansy.f...@fujitsu.com> > > Reviewed-by: Christoph Hellwig <h...@lst.de> > > --- > > fs/dax.c | 37 ++++++++++++++++++++++++++----------- > > 1 file changed, 26 insertions(+), 11 deletions(-) > > > > diff --git a/fs/dax.c b/fs/dax.c > > index 181aad97136a..cfe513eb111e 100644 > > --- a/fs/dax.c > > +++ b/fs/dax.c > > @@ -722,6 +722,9 @@ static int copy_cow_page_dax(struct block_device > *bdev, struct dax_device *dax_d > > return 0; > > } > > > > +#define DAX_IF_DIRTY (1 << 0) > > +#define DAX_IF_COW (1 << 1) > > + > > > small comment expalining this means DAX insert flags used in > dax_insert_entry() OK. I'll add it. > > > > > /* > > * By this point grab_mapping_entry() has ensured that we have a locked > entry > > * of the appropriate size so we don't have to worry about > > downgrading PMDs to @@ -729,16 +732,19 @@ static int > copy_cow_page_dax(struct block_device *bdev, struct dax_device *dax_d > > * already in the tree, we will skip the insertion and just dirty the PMD > > as > > * appropriate. > > */ > > -static void *dax_insert_entry(struct xa_state *xas, > > - struct address_space *mapping, struct vm_fault *vmf, > > - void *entry, pfn_t pfn, unsigned long flags, bool dirty) > > +static void *dax_insert_entry(struct xa_state *xas, struct vm_fault *vmf, > > + void *entry, pfn_t pfn, unsigned long flags, > > + unsigned int insert_flags) > > { > > + struct address_space *mapping = vmf->vma->vm_file->f_mapping; > > void *new_entry = dax_make_entry(pfn, flags); > > + bool dirty = insert_flags & DAX_IF_DIRTY; > > + bool cow = insert_flags & DAX_IF_COW; > > > > if (dirty) > > __mark_inode_dirty(mapping->host, I_DIRTY_PAGES); > > > > - if (dax_is_zero_entry(entry) && !(flags & DAX_ZERO_PAGE)) { > > + if (cow || (dax_is_zero_entry(entry) && !(flags & DAX_ZERO_PAGE))) { > > unsigned long index = xas->xa_index; > > /* we are replacing a zero page with block mapping */ > > if (dax_is_pmd_entry(entry)) > > @@ -750,7 +756,7 @@ static void *dax_insert_entry(struct xa_state > > *xas, > > > > xas_reset(xas); > > xas_lock_irq(xas); > > - if (dax_is_zero_entry(entry) || dax_is_empty_entry(entry)) { > > + if (cow || dax_is_zero_entry(entry) || dax_is_empty_entry(entry)) { > > void *old; > > > > dax_disassociate_entry(entry, mapping, false); @@ -774,6 +780,9 > @@ > > static void *dax_insert_entry(struct xa_state *xas, > > if (dirty) > > xas_set_mark(xas, PAGECACHE_TAG_DIRTY); > > > > + if (cow) > > + xas_set_mark(xas, PAGECACHE_TAG_TOWRITE); > > + > > xas_unlock_irq(xas); > > return entry; > > } > > @@ -1098,8 +1107,7 @@ static vm_fault_t dax_load_hole(struct xa_state > *xas, > > pfn_t pfn = pfn_to_pfn_t(my_zero_pfn(vaddr)); > > vm_fault_t ret; > > > > - *entry = dax_insert_entry(xas, mapping, vmf, *entry, pfn, > > - DAX_ZERO_PAGE, false); > > + *entry = dax_insert_entry(xas, vmf, *entry, pfn, DAX_ZERO_PAGE, 0); > > > > ret = vmf_insert_mixed(vmf->vma, vaddr, pfn); > > trace_dax_load_hole(inode, vmf, ret); @@ -1126,8 +1134,8 @@ static > > vm_fault_t dax_pmd_load_hole(struct xa_state *xas, struct vm_fault *vmf, > > goto fallback; > > > > pfn = page_to_pfn_t(zero_page); > > - *entry = dax_insert_entry(xas, mapping, vmf, *entry, pfn, > > - DAX_PMD | DAX_ZERO_PAGE, false); > > + *entry = dax_insert_entry(xas, vmf, *entry, pfn, > > + DAX_PMD | DAX_ZERO_PAGE, 0); > > > > if (arch_needs_pgtable_deposit()) { > > pgtable = pte_alloc_one(vma->vm_mm); @@ -1431,6 +1439,7 @@ > static > > vm_fault_t dax_fault_actor(struct vm_fault *vmf, pfn_t *pfnp, > > loff_t pos = (loff_t)xas->xa_offset << PAGE_SHIFT; > > bool write = vmf->flags & FAULT_FLAG_WRITE; > > bool sync = dax_fault_is_synchronous(flags, vmf->vma, iomap); > > + unsigned int insert_flags = 0; > > int err = 0; > > pfn_t pfn; > > void *kaddr; > > @@ -1453,8 +1462,14 @@ static vm_fault_t dax_fault_actor(struct vm_fault > *vmf, pfn_t *pfnp, > > if (err) > > return dax_fault_return(err); > > > > - entry = dax_insert_entry(xas, mapping, vmf, entry, pfn, 0, > > - write && !sync); > > + if (write) { > > + if (!sync) > > + insert_flags |= DAX_IF_DIRTY; > > + if (iomap->flags & IOMAP_F_SHARED) > > + insert_flags |= DAX_IF_COW; > > + } > > + > > + entry = dax_insert_entry(xas, vmf, entry, pfn, 0, insert_flags); > > > > if (write && srcmap->addr != iomap->addr) { > > err = dax_iomap_cow_copy(pos, size, size, srcmap, kaddr, false); > > > > Rest looks good to me. Please feel free to add > Reviewed-by: Ritesh Harjani <rite...@gmail.com> > > sorry about changing my email in between of this code review. > I am planning to use above gmail id as primary account for all upstream work > from now. > -- Thanks, Ruan Shiyang. > > -- > > 2.30.1 > > > > > >