On Fri, Oct 11, 2019 at 03:21:58PM +0100, Mark Rutland wrote: > On Thu, Oct 10, 2019 at 07:44:39PM +0100, Dave Martin wrote: > > Since normal execution of any non-branch instruction resets the > > PSTATE BTYPE field to 0, so do the same thing when emulating a > > trapped instruction. > > > > Branches don't trap directly, so we should never need to assign a > > non-zero value to BTYPE here. > > > > Signed-off-by: Dave Martin <dave.mar...@arm.com> > > --- > > arch/arm64/kernel/traps.c | 2 ++ > > 1 file changed, 2 insertions(+) > > > > diff --git a/arch/arm64/kernel/traps.c b/arch/arm64/kernel/traps.c > > index 3af2768..4d8ce50 100644 > > --- a/arch/arm64/kernel/traps.c > > +++ b/arch/arm64/kernel/traps.c > > @@ -331,6 +331,8 @@ void arm64_skip_faulting_instruction(struct pt_regs > > *regs, unsigned long size) > > > > if (regs->pstate & PSR_MODE32_BIT) > > advance_itstate(regs); > > + else > > + regs->pstate &= ~(u64)PSR_BTYPE_MASK; > > This looks good to me, with one nit below. > > We don't (currently) need the u64 cast here, and it's inconsistent with > what we do elsewhere. If the upper 32-bit of pstate get allocated, we'll > need to fix up all the other masking we do:
Huh, looks like I missed that. Dang. Will fix. > [mark@lakrids:~/src/linux]% git grep 'pstate &= ~' > arch/arm64/kernel/armv8_deprecated.c: regs->pstate &= > ~PSR_AA32_E_BIT; > arch/arm64/kernel/cpufeature.c: regs->pstate &= ~PSR_SSBS_BIT; > arch/arm64/kernel/debug-monitors.c: regs->pstate &= ~DBG_SPSR_SS; > arch/arm64/kernel/insn.c: pstate &= ~(pstate >> 1); /* PSR_C_BIT > &= ~PSR_Z_BIT */ > arch/arm64/kernel/insn.c: pstate &= ~(pstate >> 1); /* PSR_C_BIT > &= ~PSR_Z_BIT */ > arch/arm64/kernel/probes/kprobes.c: regs->pstate &= ~PSR_D_BIT; > arch/arm64/kernel/probes/kprobes.c: regs->pstate &= ~DAIF_MASK; > arch/arm64/kernel/ptrace.c: regs->pstate &= ~SPSR_EL1_AARCH32_RES0_BITS; > arch/arm64/kernel/ptrace.c: regs->pstate &= > ~PSR_AA32_E_BIT; > arch/arm64/kernel/ptrace.c: regs->pstate &= ~SPSR_EL1_AARCH64_RES0_BITS; > arch/arm64/kernel/ptrace.c: regs->pstate &= ~DBG_SPSR_SS; > arch/arm64/kernel/ssbd.c: task_pt_regs(task)->pstate &= ~val; > arch/arm64/kernel/traps.c: regs->pstate &= ~PSR_AA32_IT_MASK; > > ... and at that point I'd suggest we should just ensure the bit > definitions are all defined as unsigned long in the first place since > adding casts to each use is error-prone. Are we concerned about changing the types of UAPI #defines? That can cause subtle and unexpected breakage, especially when the signedness of a #define changes. Ideally, we'd just change all these to 1UL << n. Cheers ---Dave