Hi Jun, On 22/08/18 10:54, Jun Yao wrote: > Create the initial page table in the init_pg_dir. And before > calling kasan_early_init(), we update the init_mm.pgd by > introducing set_init_mm_pgd(). This will ensure that pgd_offset_k() > works correctly. When the final page table is created, we redirect > the init_mm.pgd to the swapper_pg_dir.
> diff --git a/arch/arm64/kernel/head.S b/arch/arm64/kernel/head.S > index c3e4b1886cde..ede2e964592b 100644 > --- a/arch/arm64/kernel/head.S > +++ b/arch/arm64/kernel/head.S > @@ -402,7 +402,6 @@ __create_page_tables: > adrp x1, init_pg_end > sub x1, x1, x0 > bl __inval_dcache_area > - > ret x28 > ENDPROC(__create_page_tables) > .ltorg Nit: spurious whitespace change. > @@ -439,6 +438,9 @@ __primary_switched: > bl __pi_memset > dsb ishst // Make zero page visible to PTW > > + adrp x0, init_pg_dir > + bl set_init_mm_pgd Having a C helper to just do a store is a bit strange, calling C code before kasan is ready is clearly causing some pain. Couldn't we just do store in assembly here?: ------------------%<------------------ diff --git a/arch/arm64/kernel/head.S b/arch/arm64/kernel/head.S index ede2e964592b..7464fb31452d 100644 --- a/arch/arm64/kernel/head.S +++ b/arch/arm64/kernel/head.S @@ -439,7 +439,8 @@ __primary_switched: dsb ishst // Make zero page visible to PTW adrp x0, init_pg_dir - bl set_init_mm_pgd + adr_l x1, init_mm + str x0, [x1, #MM_PGD] #ifdef CONFIG_KASAN bl kasan_early_init diff --git a/arch/arm64/kernel/asm-offsets.c b/arch/arm64/kernel/asm-offsets.c index 323aeb5f2fe6..43f52cfdfad4 100644 --- a/arch/arm64/kernel/asm-offsets.c +++ b/arch/arm64/kernel/asm-offsets.c @@ -82,6 +82,7 @@ int main(void) DEFINE(S_FRAME_SIZE, sizeof(struct pt_regs)); BLANK(); DEFINE(MM_CONTEXT_ID, offsetof(struct mm_struct, context.id.counter)); + DEFINE(MM_PGD, offsetof(struct mm_struct, pgd)); BLANK(); DEFINE(VMA_VM_MM, offsetof(struct vm_area_struct, vm_mm)); DEFINE(VMA_VM_FLAGS, offsetof(struct vm_area_struct, vm_flags)); ------------------%<------------------ > diff --git a/arch/arm64/mm/mmu.c b/arch/arm64/mm/mmu.c > index 65f86271f02b..f7e544f6f3eb 100644 > --- a/arch/arm64/mm/mmu.c > +++ b/arch/arm64/mm/mmu.c > @@ -623,6 +623,19 @@ static void __init map_kernel(pgd_t *pgdp) > kasan_copy_shadow(pgdp); > } > > +/* > + * set_init_mm_pgd() just updates init_mm.pgd. The purpose of using > + * assembly is to prevent KASAN instrumentation, as KASAN has not > + * been initialized when this function is called. You're hiding the store from KASAN as its shadow region hasn't been initialized yet? I think newer versions of the compiler let KASAN check stack accesses too, and the compiler may generate those all by itself. Hiding things like this gets us into an arms-race with the compiler. > +void __init set_init_mm_pgd(pgd_t *pgd) > +{ > + pgd_t **addr = &(init_mm.pgd); > + > + asm volatile("str %x0, [%1]\n" > + : : "r" (pgd), "r" (addr) : "memory"); > +} > + > /* > * paging_init() sets up the page tables, initialises the zone memory > * maps and sets up the zero page. Thanks, James