Hi,

By the way my email address is not anymore @c-s.fr but @csgroup.eu 
allthough the former still works.

Le 26/09/2022 à 08:43, Benjamin Gray a écrit :
> Adds a generic text patching mechanism for patches of 1, 2, 4, or (64-bit) 8
> bytes. The patcher conditionally syncs the icache depending on if
> the content will be executed (as opposed to, e.g., read-only data).
> 
> The `patch_instruction` function is reimplemented in terms of this
> more generic function. This generic implementation allows patching of
> arbitrary 64-bit data, whereas the original `patch_instruction` decided
> the size based on the 'instruction' opcode, so was not suitable for
> arbitrary data.

I get a lot better though still some slight degradation: I get approx 3% 
more time needed to activate and de-activate ftrace when 
STRICT_KERNEL_RWX is selected.

I get a surprising result without STRICT_KERNEL_RWX. Activation is also 
3% but the de-activation needs 25% more time.


> 
> Signed-off-by: Benjamin Gray <bg...@linux.ibm.com>
> ---
>   arch/powerpc/include/asm/code-patching.h |  7 ++
>   arch/powerpc/lib/code-patching.c         | 90 +++++++++++++++++-------
>   2 files changed, 71 insertions(+), 26 deletions(-)
> 
> diff --git a/arch/powerpc/include/asm/code-patching.h 
> b/arch/powerpc/include/asm/code-patching.h
> index 1c6316ec4b74..15efd8ab22da 100644
> --- a/arch/powerpc/include/asm/code-patching.h
> +++ b/arch/powerpc/include/asm/code-patching.h
> @@ -76,6 +76,13 @@ int create_cond_branch(ppc_inst_t *instr, const u32 *addr,
>   int patch_branch(u32 *addr, unsigned long target, int flags);
>   int patch_instruction(u32 *addr, ppc_inst_t instr);
>   int raw_patch_instruction(u32 *addr, ppc_inst_t instr);
> +int __patch_memory(void *dest, unsigned long src, size_t size);
> +
> +#define patch_memory(addr, val) \
> +({ \
> +     BUILD_BUG_ON(!__native_word(val)); \
> +     __patch_memory(addr, (unsigned long) val, sizeof(val)); \
> +})

Can you do a static __always_inline function instead of a macro here ?

> 
>   static inline unsigned long patch_site_addr(s32 *site)
>   {
> diff --git a/arch/powerpc/lib/code-patching.c 
> b/arch/powerpc/lib/code-patching.c
> index ad0cf3108dd0..9979380d55ef 100644
> --- a/arch/powerpc/lib/code-patching.c
> +++ b/arch/powerpc/lib/code-patching.c
> @@ -15,20 +15,47 @@
>   #include <asm/code-patching.h>
>   #include <asm/inst.h>
> 
> -static int __patch_instruction(u32 *exec_addr, ppc_inst_t instr, u32 
> *patch_addr)
> +static int __always_inline ___patch_memory(void *patch_addr,
> +                                        unsigned long data,
> +                                        void *prog_addr,
> +                                        size_t size)

Is it really needed in the .c file ? I would expect GCC to take the 
right decision by itself.

By the way, the __always_inline must immediately follow static.

>   {
> -     if (!ppc_inst_prefixed(instr)) {
> -             u32 val = ppc_inst_val(instr);
> +     switch (size) {
> +     case 1:
> +             __put_kernel_nofault(patch_addr, &data, u8, failed);
> +             break;
> +     case 2:
> +             __put_kernel_nofault(patch_addr, &data, u16, failed);
> +             break;
> +     case 4:
> +             __put_kernel_nofault(patch_addr, &data, u32, failed);
> +             break;
> +#ifdef CONFIG_PPC64
> +     case 8:
> +             __put_kernel_nofault(patch_addr, &data, u64, failed);
> +             break;
> +#endif
> +     default:
> +             unreachable();

A BUILD_BUG() would be better here I think.

> +     }
> 
> -             __put_kernel_nofault(patch_addr, &val, u32, failed);
> -     } else {
> -             u64 val = ppc_inst_as_ulong(instr);
> +     dcbst(patch_addr);
> +     dcbst(patch_addr + size - 1); /* Last byte of data may cross a 
> cacheline */

Or the second byte of data may cross a cacheline ...

> 
> -             __put_kernel_nofault(patch_addr, &val, u64, failed);
> -     }
> +     mb(); /* sync */
> +
> +     /* Flush on the EA that may be executed in case of a non-coherent 
> icache */
> +     icbi(prog_addr);
> +
> +     /* Also flush the last byte of the instruction if it may be a
> +      * prefixed instruction and we aren't assuming minimum 64-byte
> +      * cacheline sizes
> +      */
> +     if (IS_ENABLED(CONFIG_PPC64) && L1_CACHE_BYTES < 64)
> +             icbi(prog_addr + size - 1);
> 
> -     asm ("dcbst 0, %0; sync; icbi 0,%1; sync; isync" :: "r" (patch_addr),
> -                                                         "r" (exec_addr));
> +     mb(); /* sync */
> +     isync();
> 
>       return 0;
> 
> @@ -38,7 +65,10 @@ static int __patch_instruction(u32 *exec_addr, ppc_inst_t 
> instr, u32 *patch_addr
> 
>   int raw_patch_instruction(u32 *addr, ppc_inst_t instr)
>   {
> -     return __patch_instruction(addr, instr, addr);
> +     if (ppc_inst_prefixed(instr))
> +             return ___patch_memory(addr, ppc_inst_as_ulong(instr), addr, 
> sizeof(u64));
> +     else
> +             return ___patch_memory(addr, ppc_inst_val(instr), addr, 
> sizeof(u32));
>   }
> 
>   #ifdef CONFIG_STRICT_KERNEL_RWX
> @@ -147,24 +177,22 @@ static void unmap_patch_area(unsigned long addr)
>       flush_tlb_kernel_range(addr, addr + PAGE_SIZE);
>   }
> 
> -static int __do_patch_instruction(u32 *addr, ppc_inst_t instr)
> +static int __always_inline __do_patch_memory(void *dest, unsigned long src, 
> size_t size)
>   {

Whaou, do we really want all this to be __always_inline ? Did you check 
the text size increase ?

>       int err;
>       u32 *patch_addr;
> -     unsigned long text_poke_addr;
>       pte_t *pte;
> -     unsigned long pfn = get_patch_pfn(addr);
> -
> -     text_poke_addr = (unsigned long)__this_cpu_read(text_poke_area)->addr & 
> PAGE_MASK;
> -     patch_addr = (u32 *)(text_poke_addr + offset_in_page(addr));
> +     unsigned long text_poke_addr = (unsigned 
> long)__this_cpu_read(text_poke_area)->addr & PAGE_MASK;
> +     unsigned long pfn = get_patch_pfn(dest);
> 
> +     patch_addr = (u32 *)(text_poke_addr + offset_in_page(dest));
>       pte = virt_to_kpte(text_poke_addr);
>       __set_pte_at(&init_mm, text_poke_addr, pte, pfn_pte(pfn, PAGE_KERNEL), 
> 0);
>       /* See ptesync comment in radix__set_pte_at() */
>       if (radix_enabled())
>               asm volatile("ptesync": : :"memory");
> 
> -     err = __patch_instruction(addr, instr, patch_addr);
> +     err = ___patch_memory(patch_addr, src, dest, size);
> 
>       pte_clear(&init_mm, text_poke_addr, pte);
>       flush_tlb_kernel_range(text_poke_addr, text_poke_addr + PAGE_SIZE);
> @@ -172,7 +200,7 @@ static int __do_patch_instruction(u32 *addr, ppc_inst_t 
> instr)
>       return err;
>   }
> 
> -static int do_patch_instruction(u32 *addr, ppc_inst_t instr)
> +static int __always_inline do_patch_memory(void *dest, unsigned long src, 
> size_t size)
>   {
>       int err;
>       unsigned long flags;
> @@ -183,32 +211,42 @@ static int do_patch_instruction(u32 *addr, ppc_inst_t 
> instr)
>        * to allow patching. We just do the plain old patching
>        */
>       if (!static_branch_likely(&poking_init_done))
> -             return raw_patch_instruction(addr, instr);
> +             return ___patch_memory(dest, src, dest, size);
> 
>       local_irq_save(flags);
> -     err = __do_patch_instruction(addr, instr);
> +     err = __do_patch_memory(dest, src, size);
>       local_irq_restore(flags);
> 
>       return err;
>   }
> +
>   #else /* !CONFIG_STRICT_KERNEL_RWX */
> 
> -static int do_patch_instruction(u32 *addr, ppc_inst_t instr)
> +static int do_patch_memory(void *dest, unsigned long src, size_t size)
>   {
> -     return raw_patch_instruction(addr, instr);
> +     return ___patch_memory(dest, src, dest, size);
>   }
> 
>   #endif /* CONFIG_STRICT_KERNEL_RWX */
> 
>   __ro_after_init DEFINE_STATIC_KEY_FALSE(init_mem_is_free);
> 
> -int patch_instruction(u32 *addr, ppc_inst_t instr)
> +int __patch_memory(void *dest, unsigned long src, size_t size)
>   {
>       /* Make sure we aren't patching a freed init section */
> -     if (static_branch_likely(&init_mem_is_free) && 
> init_section_contains(addr, 4))
> +     if (static_branch_likely(&init_mem_is_free) && 
> init_section_contains(dest, 4))
>               return 0;
> 
> -     return do_patch_instruction(addr, instr);
> +     return do_patch_memory(dest, src, size);
> +}
> +NOKPROBE_SYMBOL(__patch_memory);
> +
> +int patch_instruction(u32 *addr, ppc_inst_t instr)
> +{
> +     if (ppc_inst_prefixed(instr))
> +             return patch_memory(addr, ppc_inst_as_ulong(instr));
> +     else
> +             return patch_memory(addr, ppc_inst_val(instr));
>   }
>   NOKPROBE_SYMBOL(patch_instruction);
> 
> --
> 2.37.3

Reply via email to