On Thu, Jun 15, 2017 at 04:58:56PM +0200, Jan Kara wrote: > On Wed 14-06-17 11:22:11, Ross Zwisler wrote: > > @@ -216,17 +217,6 @@ static void dax_unlock_mapping_entry(struct > > address_space *mapping, > > dax_wake_mapping_entry_waiter(mapping, index, entry, false); > > } > > > > -static void put_locked_mapping_entry(struct address_space *mapping, > > - pgoff_t index, void *entry) > > -{ > > - if (!radix_tree_exceptional_entry(entry)) { > > - unlock_page(entry); > > - put_page(entry); > > - } else { > > - dax_unlock_mapping_entry(mapping, index); > > - } > > -} > > - > > The naming becomes asymetric with this. So I'd prefer keeping > put_locked_mapping_entry() as a trivial wrapper around > dax_unlock_mapping_entry() unless we can craft more sensible naming / API > for entry grabbing (and that would be a separate patch anyway).
Sure, that works for me. I'll fix for v3. > > -static int dax_load_hole(struct address_space *mapping, void **entry, > > +static int dax_load_hole(struct address_space *mapping, void *entry, > > struct vm_fault *vmf) > > { > > struct inode *inode = mapping->host; > > - struct page *page; > > - int ret; > > - > > - /* Hole page already exists? Return it... */ > > - if (!radix_tree_exceptional_entry(*entry)) { > > - page = *entry; > > - goto finish_fault; > > - } > > + unsigned long vaddr = vmf->address; > > + int ret = VM_FAULT_NOPAGE; > > + struct page *zero_page; > > + void *entry2; > > > > - /* This will replace locked radix tree entry with a hole page */ > > - page = find_or_create_page(mapping, vmf->pgoff, > > - vmf->gfp_mask | __GFP_ZERO); > > With this gone, you can also remove the special DAX handling from > mm/filemap.c: page_cache_tree_insert() and remove from dax.h > dax_wake_mapping_entry_waiter(), dax_radix_locked_entry() and RADIX_DAX > definitions. Yay! As a separate patch please. Oh, yay! :) Sure, I'll have this patch for v3.