When header struct fields use very common names like "pages" or "type", it makes it difficult to find uses of these fields with tools like grep and cscope. Add the prefix mreg_ to some fields in struct mshv_mem_region to make it easier to find them.
There is no functional change. Signed-off-by: Mukesh Rathor <[email protected]> --- drivers/hv/mshv_regions.c | 44 ++++++++++++++++++------------------- drivers/hv/mshv_root.h | 6 ++--- drivers/hv/mshv_root_main.c | 10 ++++----- 3 files changed, 30 insertions(+), 30 deletions(-) diff --git a/drivers/hv/mshv_regions.c b/drivers/hv/mshv_regions.c index 202b9d551e39..af81405f859b 100644 --- a/drivers/hv/mshv_regions.c +++ b/drivers/hv/mshv_regions.c @@ -52,7 +52,7 @@ static long mshv_region_process_chunk(struct mshv_mem_region *region, struct page *page; int ret; - page = region->pages[page_offset]; + page = region->mreg_pages[page_offset]; if (!page) return -EINVAL; @@ -65,7 +65,7 @@ static long mshv_region_process_chunk(struct mshv_mem_region *region, /* Start at stride since the first page is validated */ for (count = stride; count < page_count; count += stride) { - page = region->pages[page_offset + count]; + page = region->mreg_pages[page_offset + count]; /* Break if current page is not present */ if (!page) @@ -117,7 +117,7 @@ static int mshv_region_process_range(struct mshv_mem_region *region, while (page_count) { /* Skip non-present pages */ - if (!region->pages[page_offset]) { + if (!region->mreg_pages[page_offset]) { page_offset++; page_count--; continue; @@ -164,13 +164,13 @@ static int mshv_region_chunk_share(struct mshv_mem_region *region, u32 flags, u64 page_offset, u64 page_count) { - struct page *page = region->pages[page_offset]; + struct page *page = region->mreg_pages[page_offset]; if (PageHuge(page) || PageTransCompound(page)) flags |= HV_MODIFY_SPA_PAGE_HOST_ACCESS_LARGE_PAGE; return hv_call_modify_spa_host_access(region->partition->pt_id, - region->pages + page_offset, + region->mreg_pages + page_offset, page_count, HV_MAP_GPA_READABLE | HV_MAP_GPA_WRITABLE, @@ -190,13 +190,13 @@ static int mshv_region_chunk_unshare(struct mshv_mem_region *region, u32 flags, u64 page_offset, u64 page_count) { - struct page *page = region->pages[page_offset]; + struct page *page = region->mreg_pages[page_offset]; if (PageHuge(page) || PageTransCompound(page)) flags |= HV_MODIFY_SPA_PAGE_HOST_ACCESS_LARGE_PAGE; return hv_call_modify_spa_host_access(region->partition->pt_id, - region->pages + page_offset, + region->mreg_pages + page_offset, page_count, 0, flags, false); } @@ -214,7 +214,7 @@ static int mshv_region_chunk_remap(struct mshv_mem_region *region, u32 flags, u64 page_offset, u64 page_count) { - struct page *page = region->pages[page_offset]; + struct page *page = region->mreg_pages[page_offset]; if (PageHuge(page) || PageTransCompound(page)) flags |= HV_MAP_GPA_LARGE_PAGE; @@ -222,7 +222,7 @@ static int mshv_region_chunk_remap(struct mshv_mem_region *region, return hv_call_map_gpa_pages(region->partition->pt_id, region->start_gfn + page_offset, page_count, flags, - region->pages + page_offset); + region->mreg_pages + page_offset); } static int mshv_region_remap_pages(struct mshv_mem_region *region, @@ -245,10 +245,10 @@ int mshv_region_map(struct mshv_mem_region *region) static void mshv_region_invalidate_pages(struct mshv_mem_region *region, u64 page_offset, u64 page_count) { - if (region->type == MSHV_REGION_TYPE_MEM_PINNED) - unpin_user_pages(region->pages + page_offset, page_count); + if (region->mreg_type == MSHV_REGION_TYPE_MEM_PINNED) + unpin_user_pages(region->mreg_pages + page_offset, page_count); - memset(region->pages + page_offset, 0, + memset(region->mreg_pages + page_offset, 0, page_count * sizeof(struct page *)); } @@ -265,7 +265,7 @@ int mshv_region_pin(struct mshv_mem_region *region) int ret; for (done_count = 0; done_count < region->nr_pages; done_count += ret) { - pages = region->pages + done_count; + pages = region->mreg_pages + done_count; userspace_addr = region->start_uaddr + done_count * HV_HYP_PAGE_SIZE; nr_pages = min(region->nr_pages - done_count, @@ -297,7 +297,7 @@ static int mshv_region_chunk_unmap(struct mshv_mem_region *region, u32 flags, u64 page_offset, u64 page_count) { - struct page *page = region->pages[page_offset]; + struct page *page = region->mreg_pages[page_offset]; if (PageHuge(page) || PageTransCompound(page)) flags |= HV_UNMAP_GPA_LARGE_PAGE; @@ -321,7 +321,7 @@ static void mshv_region_destroy(struct kref *ref) struct mshv_partition *partition = region->partition; int ret; - if (region->type == MSHV_REGION_TYPE_MEM_MOVABLE) + if (region->mreg_type == MSHV_REGION_TYPE_MEM_MOVABLE) mshv_region_movable_fini(region); if (mshv_partition_encrypted(partition)) { @@ -374,9 +374,9 @@ static int mshv_region_hmm_fault_and_lock(struct mshv_mem_region *region, int ret; range->notifier_seq = mmu_interval_read_begin(range->notifier); - mmap_read_lock(region->mni.mm); + mmap_read_lock(region->mreg_mni.mm); ret = hmm_range_fault(range); - mmap_read_unlock(region->mni.mm); + mmap_read_unlock(region->mreg_mni.mm); if (ret) return ret; @@ -407,7 +407,7 @@ static int mshv_region_range_fault(struct mshv_mem_region *region, u64 page_offset, u64 page_count) { struct hmm_range range = { - .notifier = ®ion->mni, + .notifier = ®ion->mreg_mni, .default_flags = HMM_PFN_REQ_FAULT | HMM_PFN_REQ_WRITE, }; unsigned long *pfns; @@ -430,7 +430,7 @@ static int mshv_region_range_fault(struct mshv_mem_region *region, goto out; for (i = 0; i < page_count; i++) - region->pages[page_offset + i] = hmm_pfn_to_page(pfns[i]); + region->mreg_pages[page_offset + i] = hmm_pfn_to_page(pfns[i]); ret = mshv_region_remap_pages(region, region->hv_map_flags, page_offset, page_count); @@ -489,7 +489,7 @@ static bool mshv_region_interval_invalidate(struct mmu_interval_notifier *mni, { struct mshv_mem_region *region = container_of(mni, struct mshv_mem_region, - mni); + mreg_mni); u64 page_offset, page_count; unsigned long mstart, mend; int ret = -EPERM; @@ -535,14 +535,14 @@ static const struct mmu_interval_notifier_ops mshv_region_mni_ops = { void mshv_region_movable_fini(struct mshv_mem_region *region) { - mmu_interval_notifier_remove(®ion->mni); + mmu_interval_notifier_remove(®ion->mreg_mni); } bool mshv_region_movable_init(struct mshv_mem_region *region) { int ret; - ret = mmu_interval_notifier_insert(®ion->mni, current->mm, + ret = mmu_interval_notifier_insert(®ion->mreg_mni, current->mm, region->start_uaddr, region->nr_pages << HV_HYP_PAGE_SHIFT, &mshv_region_mni_ops); diff --git a/drivers/hv/mshv_root.h b/drivers/hv/mshv_root.h index 3c1d88b36741..f5b6d3979e5a 100644 --- a/drivers/hv/mshv_root.h +++ b/drivers/hv/mshv_root.h @@ -85,10 +85,10 @@ struct mshv_mem_region { u64 start_uaddr; u32 hv_map_flags; struct mshv_partition *partition; - enum mshv_region_type type; - struct mmu_interval_notifier mni; + enum mshv_region_type mreg_type; + struct mmu_interval_notifier mreg_mni; struct mutex mutex; /* protects region pages remapping */ - struct page *pages[]; + struct page *mreg_pages[]; }; struct mshv_irq_ack_notifier { diff --git a/drivers/hv/mshv_root_main.c b/drivers/hv/mshv_root_main.c index 1134a82c7881..eff1b21461dc 100644 --- a/drivers/hv/mshv_root_main.c +++ b/drivers/hv/mshv_root_main.c @@ -657,7 +657,7 @@ static bool mshv_handle_gpa_intercept(struct mshv_vp *vp) return false; /* Only movable memory ranges are supported for GPA intercepts */ - if (region->type == MSHV_REGION_TYPE_MEM_MOVABLE) + if (region->mreg_type == MSHV_REGION_TYPE_MEM_MOVABLE) ret = mshv_region_handle_gfn_fault(region, gfn); else ret = false; @@ -1175,12 +1175,12 @@ static int mshv_partition_create_region(struct mshv_partition *partition, return PTR_ERR(rg); if (is_mmio) - rg->type = MSHV_REGION_TYPE_MMIO; + rg->mreg_type = MSHV_REGION_TYPE_MMIO; else if (mshv_partition_encrypted(partition) || !mshv_region_movable_init(rg)) - rg->type = MSHV_REGION_TYPE_MEM_PINNED; + rg->mreg_type = MSHV_REGION_TYPE_MEM_PINNED; else - rg->type = MSHV_REGION_TYPE_MEM_MOVABLE; + rg->mreg_type = MSHV_REGION_TYPE_MEM_MOVABLE; rg->partition = partition; @@ -1297,7 +1297,7 @@ mshv_map_user_memory(struct mshv_partition *partition, if (ret) return ret; - switch (region->type) { + switch (region->mreg_type) { case MSHV_REGION_TYPE_MEM_PINNED: ret = mshv_prepare_pinned_region(region); break; -- 2.51.2.vfs.0.1
