On Tue, Feb 6, 2018 at 1:32 PM, Dominik Brodowski <li...@dominikbrodowski.net> wrote: > Same as is done for syscalls, interleave XOR with PUSH or MOV > instructions for exceptions/interrupts, in order to minimize > the cost of the additional instructions required for register > clearing. > > Signed-off-by: Dominik Brodowski <li...@dominikbrodowski.net>
Looks good to me. Reviewed-by: Dan Williams <dan.j.willi...@intel.com> > > diff --git a/arch/x86/entry/calling.h b/arch/x86/entry/calling.h > index 75a237c95ff7..89a906c868d8 100644 > --- a/arch/x86/entry/calling.h > +++ b/arch/x86/entry/calling.h > @@ -102,10 +102,21 @@ For 32-bit we have the following conventions - kernel > is built with > .endm > > .macro SAVE_C_REGS offset=0 > + /* > + * Save registers and sanitize registers of values that a > + * speculation attack might otherwise want to exploit. The > + * lower registers are likely clobbered well before they > + * could be put to use in a speculative execution gadget. > + * Interleave XOR with MOV for better uop scheduling: > + */ > movq %r11, 6*8+\offset(%rsp) > + xorq %r11, %r11 /* nospec r11 */ > movq %r10, 7*8+\offset(%rsp) > + xorq %r10, %r10 /* nospec r10 */ > movq %r9, 8*8+\offset(%rsp) > + xorq %r9, %r9 /* nospec r9 */ > movq %r8, 9*8+\offset(%rsp) > + xorq %r8, %r8 /* nospec r8 */ > movq %rax, 10*8+\offset(%rsp) > movq %rcx, 11*8+\offset(%rsp) > movq %rdx, 12*8+\offset(%rsp) > @@ -115,34 +126,28 @@ For 32-bit we have the following conventions - kernel > is built with > .endm > > .macro SAVE_EXTRA_REGS offset=0 > + /* > + * Save registers and sanitize registers of values that a > + * speculation attack might otherwise want to exploit. The > + * lower registers are likely clobbered well before they > + * could be put to use in a speculative execution gadget. > + * Interleave XOR with MOV for better uop scheduling: > + */ > movq %r15, 0*8+\offset(%rsp) > + xorq %r15, %r15 /* nospec r15 */ > movq %r14, 1*8+\offset(%rsp) > + xorq %r14, %r14 /* nospec r14 */ > movq %r13, 2*8+\offset(%rsp) > + xorq %r13, %r13 /* nospec r13 */ > movq %r12, 3*8+\offset(%rsp) > + xorq %r12, %r12 /* nospec r12 */ > movq %rbp, 4*8+\offset(%rsp) > + xorl %ebp, %ebp /* nospec rbp */ > movq %rbx, 5*8+\offset(%rsp) > + xorl %ebx, %ebx /* nospec rbx */ > UNWIND_HINT_REGS offset=\offset > .endm > > - /* > - * Sanitize registers of values that a speculation attack > - * might otherwise want to exploit. The lower registers are > - * likely clobbered well before they could be put to use in > - * a speculative execution gadget: > - */ > - .macro CLEAR_REGS_NOSPEC > - xorl %ebp, %ebp > - xorl %ebx, %ebx > - xorq %r8, %r8 > - xorq %r9, %r9 > - xorq %r10, %r10 > - xorq %r11, %r11 > - xorq %r12, %r12 > - xorq %r13, %r13 > - xorq %r14, %r14 > - xorq %r15, %r15 > - .endm > - > .macro POP_EXTRA_REGS > popq %r15 > popq %r14 > diff --git a/arch/x86/entry/entry_64.S b/arch/x86/entry/entry_64.S > index 9e48002b953b..903d9088bdb3 100644 > --- a/arch/x86/entry/entry_64.S > +++ b/arch/x86/entry/entry_64.S > @@ -575,7 +575,6 @@ END(irq_entries_start) > ALLOC_PT_GPREGS_ON_STACK > SAVE_C_REGS > SAVE_EXTRA_REGS > - CLEAR_REGS_NOSPEC > ENCODE_FRAME_POINTER > > testb $3, CS(%rsp) > @@ -1134,7 +1133,6 @@ ENTRY(xen_failsafe_callback) > ALLOC_PT_GPREGS_ON_STACK > SAVE_C_REGS > SAVE_EXTRA_REGS > - CLEAR_REGS_NOSPEC > ENCODE_FRAME_POINTER > jmp error_exit > END(xen_failsafe_callback) > @@ -1180,7 +1178,6 @@ ENTRY(paranoid_entry) > cld > SAVE_C_REGS 8 > SAVE_EXTRA_REGS 8 > - CLEAR_REGS_NOSPEC > ENCODE_FRAME_POINTER 8 > movl $1, %ebx > movl $MSR_GS_BASE, %ecx > @@ -1233,7 +1230,6 @@ ENTRY(error_entry) > cld > SAVE_C_REGS 8 > SAVE_EXTRA_REGS 8 > - CLEAR_REGS_NOSPEC > ENCODE_FRAME_POINTER 8 > testb $3, CS+8(%rsp) > jz .Lerror_kernelspace > @@ -1420,18 +1416,34 @@ ENTRY(nmi) > pushq (%rdx) /* pt_regs->dx */ > pushq %rcx /* pt_regs->cx */ > pushq %rax /* pt_regs->ax */ > + /* > + * Sanitize registers of values that a speculation attack > + * might otherwise want to exploit. The lower registers are > + * likely clobbered well before they could be put to use in > + * a speculative execution gadget. Interleave XOR with PUSH > + * for better uop scheduling: > + */ > pushq %r8 /* pt_regs->r8 */ > + xorq %r8, %r8 /* nospec r8 */ > pushq %r9 /* pt_regs->r9 */ > + xorq %r9, %r9 /* nospec r9 */ > pushq %r10 /* pt_regs->r10 */ > + xorq %r10, %r10 /* nospec r10 */ > pushq %r11 /* pt_regs->r11 */ > + xorq %r11, %r11 /* nospec r11*/ > pushq %rbx /* pt_regs->rbx */ > + xorl %ebx, %ebx /* nospec rbx*/ > pushq %rbp /* pt_regs->rbp */ > + xorl %ebp, %ebp /* nospec rbp*/ > pushq %r12 /* pt_regs->r12 */ > + xorq %r12, %r12 /* nospec r12*/ > pushq %r13 /* pt_regs->r13 */ > + xorq %r13, %r13 /* nospec r13*/ > pushq %r14 /* pt_regs->r14 */ > + xorq %r14, %r14 /* nospec r14*/ > pushq %r15 /* pt_regs->r15 */ > + xorq %r15, %r15 /* nospec r15*/ > UNWIND_HINT_REGS > - CLEAR_REGS_NOSPEC > ENCODE_FRAME_POINTER > > /*