On Mon, Jun 29, 2015 at 12:33:43PM -0700, Andy Lutomirski wrote:
> These need to be migrated together, as the compat case used to jump
> into the middle of the 64-bit exit code.
> 
> Signed-off-by: Andy Lutomirski <l...@kernel.org>
> ---
>  arch/x86/entry/entry_64.S        | 69 
> +++++-----------------------------------
>  arch/x86/entry/entry_64_compat.S |  7 ++--
>  2 files changed, 11 insertions(+), 65 deletions(-)
> 
> diff --git a/arch/x86/entry/entry_64.S b/arch/x86/entry/entry_64.S
> index cd9cbc62159c..9bc76766aa71 100644
> --- a/arch/x86/entry/entry_64.S
> +++ b/arch/x86/entry/entry_64.S
> @@ -229,6 +229,11 @@ entry_SYSCALL_64_fastpath:
>        */
>       USERGS_SYSRET64
>  
> +GLOBAL(int_ret_from_sys_call_irqs_off)
> +     TRACE_IRQS_ON
> +     ENABLE_INTERRUPTS(CLBR_NONE)
> +     jmp int_ret_from_sys_call
> +
>       /* Do syscall entry tracing */
>  tracesys:
>       movq    %rsp, %rdi
> @@ -272,69 +277,11 @@ tracesys_phase2:
>   * Has correct iret frame.
>   */
>  GLOBAL(int_ret_from_sys_call)
> -     DISABLE_INTERRUPTS(CLBR_NONE)
> -int_ret_from_sys_call_irqs_off: /* jumps come here from the irqs-off SYSRET 
> path */
> -     TRACE_IRQS_OFF
> -     movl    $_TIF_ALLWORK_MASK, %edi
> -     /* edi: mask to check */
> -GLOBAL(int_with_check)
> -     LOCKDEP_SYS_EXIT_IRQ
> -     GET_THREAD_INFO(%rcx)
> -     movl    TI_flags(%rcx), %edx
> -     andl    %edi, %edx
> -     jnz     int_careful
> -     andl    $~TS_COMPAT, TI_status(%rcx)
> -     jmp     syscall_return
> -
> -     /*
> -      * Either reschedule or signal or syscall exit tracking needed.
> -      * First do a reschedule test.
> -      * edx: work, edi: workmask
> -      */
> -int_careful:
> -     bt      $TIF_NEED_RESCHED, %edx
> -     jnc     int_very_careful
> -     TRACE_IRQS_ON
> -     ENABLE_INTERRUPTS(CLBR_NONE)
> -     pushq   %rdi
> -     SCHEDULE_USER
> -     popq    %rdi
> -     DISABLE_INTERRUPTS(CLBR_NONE)
> -     TRACE_IRQS_OFF
> -     jmp     int_with_check
> -
> -     /* handle signals and tracing -- both require a full pt_regs */
> -int_very_careful:
> -     TRACE_IRQS_ON
> -     ENABLE_INTERRUPTS(CLBR_NONE)
>       SAVE_EXTRA_REGS
> -     /* Check for syscall exit trace */
> -     testl   $_TIF_WORK_SYSCALL_EXIT, %edx
> -     jz      int_signal
> -     pushq   %rdi
> -     leaq    8(%rsp), %rdi                   /* &ptregs -> arg1 */
> -     call    syscall_trace_leave
> -     popq    %rdi
> -     andl    $~(_TIF_WORK_SYSCALL_EXIT|_TIF_SYSCALL_EMU), %edi
> -     jmp     int_restore_rest
> -
> -int_signal:
> -     testl   $_TIF_DO_NOTIFY_MASK, %edx
> -     jz      1f
> -     movq    %rsp, %rdi                      /* &ptregs -> arg1 */
> -     xorl    %esi, %esi                      /* oldset -> arg2 */
> -     call    do_notify_resume
> -1:   movl    $_TIF_WORK_MASK, %edi
> -int_restore_rest:
> +     movq    %rsp, %rdi
> +     call    syscall_return_slowpath /* returns with IRQs disabled */
>       RESTORE_EXTRA_REGS
> -     DISABLE_INTERRUPTS(CLBR_NONE)
> -     TRACE_IRQS_OFF
> -     jmp     int_with_check
> -
> -syscall_return:
> -     /* The IRETQ could re-enable interrupts: */
> -     DISABLE_INTERRUPTS(CLBR_ANY)
> -     TRACE_IRQS_IRETQ
> +     TRACE_IRQS_IRETQ                /* we're about to change IF */
>  
>       /*
>        * Try to use SYSRET instead of IRET if we're returning to

Hallelujah!

/me luvz hunks which remove a bunch of asm :)

> diff --git a/arch/x86/entry/entry_64_compat.S 
> b/arch/x86/entry/entry_64_compat.S
> index efe0b1e499fa..ac0658142ae1 100644
> --- a/arch/x86/entry/entry_64_compat.S
> +++ b/arch/x86/entry/entry_64_compat.S
> @@ -209,10 +209,10 @@ sysexit_from_sys_call:
>       .endm
>  
>       .macro auditsys_exit exit
> -     testl   $(_TIF_ALLWORK_MASK & ~_TIF_SYSCALL_AUDIT), 
> ASM_THREAD_INFO(TI_flags, %rsp, SIZEOF_PTREGS)
> -     jnz     ia32_ret_from_sys_call
>       TRACE_IRQS_ON
>       ENABLE_INTERRUPTS(CLBR_NONE)
> +     testl $(_TIF_ALLWORK_MASK & ~_TIF_SYSCALL_AUDIT), 
> ASM_THREAD_INFO(TI_flags, %rsp, SIZEOF_PTREGS)
> +     jnz ia32_ret_from_sys_call

I guess you want to use tabs here like the rest of the macro does.

>       movl    %eax, %esi              /* second arg, syscall return value */
>       cmpl    $-MAX_ERRNO, %eax       /* is it an error ? */
>       jbe     1f
> @@ -227,11 +227,10 @@ sysexit_from_sys_call:
>       testl   %edi, ASM_THREAD_INFO(TI_flags, %rsp, SIZEOF_PTREGS)
>       jz      \exit
>       xorl    %eax, %eax              /* Do not leak kernel information */
> -     movq    %rax, R11(%rsp)

I guess that change needs at least some explanation in the commit
message. AFAIU, this is RIP we shouldn't be zeroing for we need it in
int_ret_from_sys_call...

>       movq    %rax, R10(%rsp)
>       movq    %rax, R9(%rsp)
>       movq    %rax, R8(%rsp)
> -     jmp     int_with_check
> +     jmp     int_ret_from_sys_call_irqs_off
>       .endm
>  
>  sysenter_auditsys:
> -- 
> 2.4.3
> 
> 

-- 
Regards/Gruss,
    Boris.

ECO tip #101: Trim your mails when you reply.
--
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to