Add aarch64 encoder/decoder implementation under tools/ as well
as the necessary arm64 headers.

Signed-off-by: Julien Thierry <jthie...@redhat.com>
---
 tools/arch/arm64/include/asm/aarch64-insn.h |  551 +++++++
 tools/arch/arm64/lib/aarch64-insn.c         | 1425 +++++++++++++++++++
 2 files changed, 1976 insertions(+)
 create mode 100644 tools/arch/arm64/include/asm/aarch64-insn.h
 create mode 100644 tools/arch/arm64/lib/aarch64-insn.c

diff --git a/tools/arch/arm64/include/asm/aarch64-insn.h 
b/tools/arch/arm64/include/asm/aarch64-insn.h
new file mode 100644
index 000000000000..b202d6e2e47e
--- /dev/null
+++ b/tools/arch/arm64/include/asm/aarch64-insn.h
@@ -0,0 +1,551 @@
+/* SPDX-License-Identifier: GPL-2.0-only */
+#ifndef        __ASM_AARCH64_INSN_H
+#define        __ASM_AARCH64_INSN_H
+
+#include <linux/build_bug.h>
+#include <linux/types.h>
+
+
+/* A64 instructions are always 32 bits. */
+#define        AARCH64_INSN_SIZE               4
+
+/*
+ * BRK instruction encoding
+ * The #imm16 value should be placed at bits[20:5] within BRK ins
+ */
+#define AARCH64_BREAK_MON      0xd4200000
+
+/*
+ * BRK instruction for provoking a fault on purpose
+ * Unlike kgdb, #imm16 value with unallocated handler is used for faulting.
+ */
+#define AARCH64_BREAK_FAULT    (AARCH64_BREAK_MON | (FAULT_BRK_IMM << 5))
+
+#ifndef __ASSEMBLY__
+/*
+ * ARM Architecture Reference Manual for ARMv8 Profile-A, Issue A.a
+ * Section C3.1 "A64 instruction index by encoding":
+ * AArch64 main encoding table
+ *  Bit position
+ *   28 27 26 25       Encoding Group
+ *   0  0  -  -                Unallocated
+ *   1  0  0  -                Data processing, immediate
+ *   1  0  1  -                Branch, exception generation and system 
instructions
+ *   -  1  -  0                Loads and stores
+ *   -  1  0  1                Data processing - register
+ *   0  1  1  1                Data processing - SIMD and floating point
+ *   1  1  1  1                Data processing - SIMD and floating point
+ * "-" means "don't care"
+ */
+enum aarch64_insn_encoding_class {
+       AARCH64_INSN_CLS_UNKNOWN,       /* UNALLOCATED */
+       AARCH64_INSN_CLS_SVE,           /* SVE instructions */
+       AARCH64_INSN_CLS_DP_IMM,        /* Data processing - immediate */
+       AARCH64_INSN_CLS_DP_REG,        /* Data processing - register */
+       AARCH64_INSN_CLS_DP_FPSIMD,     /* Data processing - SIMD and FP */
+       AARCH64_INSN_CLS_LDST,          /* Loads and stores */
+       AARCH64_INSN_CLS_BR_SYS,        /* Branch, exception generation and
+                                        * system instructions */
+};
+
+enum aarch64_insn_hint_cr_op {
+       AARCH64_INSN_HINT_NOP   = 0x0 << 5,
+       AARCH64_INSN_HINT_YIELD = 0x1 << 5,
+       AARCH64_INSN_HINT_WFE   = 0x2 << 5,
+       AARCH64_INSN_HINT_WFI   = 0x3 << 5,
+       AARCH64_INSN_HINT_SEV   = 0x4 << 5,
+       AARCH64_INSN_HINT_SEVL  = 0x5 << 5,
+
+       AARCH64_INSN_HINT_XPACLRI    = 0x07 << 5,
+       AARCH64_INSN_HINT_PACIA_1716 = 0x08 << 5,
+       AARCH64_INSN_HINT_PACIB_1716 = 0x0A << 5,
+       AARCH64_INSN_HINT_AUTIA_1716 = 0x0C << 5,
+       AARCH64_INSN_HINT_AUTIB_1716 = 0x0E << 5,
+       AARCH64_INSN_HINT_PACIAZ     = 0x18 << 5,
+       AARCH64_INSN_HINT_PACIASP    = 0x19 << 5,
+       AARCH64_INSN_HINT_PACIBZ     = 0x1A << 5,
+       AARCH64_INSN_HINT_PACIBSP    = 0x1B << 5,
+       AARCH64_INSN_HINT_AUTIAZ     = 0x1C << 5,
+       AARCH64_INSN_HINT_AUTIASP    = 0x1D << 5,
+       AARCH64_INSN_HINT_AUTIBZ     = 0x1E << 5,
+       AARCH64_INSN_HINT_AUTIBSP    = 0x1F << 5,
+
+       AARCH64_INSN_HINT_ESB  = 0x10 << 5,
+       AARCH64_INSN_HINT_PSB  = 0x11 << 5,
+       AARCH64_INSN_HINT_TSB  = 0x12 << 5,
+       AARCH64_INSN_HINT_CSDB = 0x14 << 5,
+
+       AARCH64_INSN_HINT_BTI   = 0x20 << 5,
+       AARCH64_INSN_HINT_BTIC  = 0x22 << 5,
+       AARCH64_INSN_HINT_BTIJ  = 0x24 << 5,
+       AARCH64_INSN_HINT_BTIJC = 0x26 << 5,
+};
+
+enum aarch64_insn_imm_type {
+       AARCH64_INSN_IMM_ADR,
+       AARCH64_INSN_IMM_26,
+       AARCH64_INSN_IMM_19,
+       AARCH64_INSN_IMM_16,
+       AARCH64_INSN_IMM_14,
+       AARCH64_INSN_IMM_12,
+       AARCH64_INSN_IMM_9,
+       AARCH64_INSN_IMM_7,
+       AARCH64_INSN_IMM_6,
+       AARCH64_INSN_IMM_S,
+       AARCH64_INSN_IMM_R,
+       AARCH64_INSN_IMM_N,
+       AARCH64_INSN_IMM_MAX
+};
+
+enum aarch64_insn_register_type {
+       AARCH64_INSN_REGTYPE_RT,
+       AARCH64_INSN_REGTYPE_RN,
+       AARCH64_INSN_REGTYPE_RT2,
+       AARCH64_INSN_REGTYPE_RM,
+       AARCH64_INSN_REGTYPE_RD,
+       AARCH64_INSN_REGTYPE_RA,
+       AARCH64_INSN_REGTYPE_RS,
+};
+
+enum aarch64_insn_register {
+       AARCH64_INSN_REG_0  = 0,
+       AARCH64_INSN_REG_1  = 1,
+       AARCH64_INSN_REG_2  = 2,
+       AARCH64_INSN_REG_3  = 3,
+       AARCH64_INSN_REG_4  = 4,
+       AARCH64_INSN_REG_5  = 5,
+       AARCH64_INSN_REG_6  = 6,
+       AARCH64_INSN_REG_7  = 7,
+       AARCH64_INSN_REG_8  = 8,
+       AARCH64_INSN_REG_9  = 9,
+       AARCH64_INSN_REG_10 = 10,
+       AARCH64_INSN_REG_11 = 11,
+       AARCH64_INSN_REG_12 = 12,
+       AARCH64_INSN_REG_13 = 13,
+       AARCH64_INSN_REG_14 = 14,
+       AARCH64_INSN_REG_15 = 15,
+       AARCH64_INSN_REG_16 = 16,
+       AARCH64_INSN_REG_17 = 17,
+       AARCH64_INSN_REG_18 = 18,
+       AARCH64_INSN_REG_19 = 19,
+       AARCH64_INSN_REG_20 = 20,
+       AARCH64_INSN_REG_21 = 21,
+       AARCH64_INSN_REG_22 = 22,
+       AARCH64_INSN_REG_23 = 23,
+       AARCH64_INSN_REG_24 = 24,
+       AARCH64_INSN_REG_25 = 25,
+       AARCH64_INSN_REG_26 = 26,
+       AARCH64_INSN_REG_27 = 27,
+       AARCH64_INSN_REG_28 = 28,
+       AARCH64_INSN_REG_29 = 29,
+       AARCH64_INSN_REG_FP = 29, /* Frame pointer */
+       AARCH64_INSN_REG_30 = 30,
+       AARCH64_INSN_REG_LR = 30, /* Link register */
+       AARCH64_INSN_REG_ZR = 31, /* Zero: as source register */
+       AARCH64_INSN_REG_SP = 31  /* Stack pointer: as load/store base reg */
+};
+
+enum aarch64_insn_special_register {
+       AARCH64_INSN_SPCLREG_SPSR_EL1   = 0xC200,
+       AARCH64_INSN_SPCLREG_ELR_EL1    = 0xC201,
+       AARCH64_INSN_SPCLREG_SP_EL0     = 0xC208,
+       AARCH64_INSN_SPCLREG_SPSEL      = 0xC210,
+       AARCH64_INSN_SPCLREG_CURRENTEL  = 0xC212,
+       AARCH64_INSN_SPCLREG_DAIF       = 0xDA11,
+       AARCH64_INSN_SPCLREG_NZCV       = 0xDA10,
+       AARCH64_INSN_SPCLREG_FPCR       = 0xDA20,
+       AARCH64_INSN_SPCLREG_DSPSR_EL0  = 0xDA28,
+       AARCH64_INSN_SPCLREG_DLR_EL0    = 0xDA29,
+       AARCH64_INSN_SPCLREG_SPSR_EL2   = 0xE200,
+       AARCH64_INSN_SPCLREG_ELR_EL2    = 0xE201,
+       AARCH64_INSN_SPCLREG_SP_EL1     = 0xE208,
+       AARCH64_INSN_SPCLREG_SPSR_INQ   = 0xE218,
+       AARCH64_INSN_SPCLREG_SPSR_ABT   = 0xE219,
+       AARCH64_INSN_SPCLREG_SPSR_UND   = 0xE21A,
+       AARCH64_INSN_SPCLREG_SPSR_FIQ   = 0xE21B,
+       AARCH64_INSN_SPCLREG_SPSR_EL3   = 0xF200,
+       AARCH64_INSN_SPCLREG_ELR_EL3    = 0xF201,
+       AARCH64_INSN_SPCLREG_SP_EL2     = 0xF210
+};
+
+enum aarch64_insn_variant {
+       AARCH64_INSN_VARIANT_32BIT,
+       AARCH64_INSN_VARIANT_64BIT
+};
+
+enum aarch64_insn_condition {
+       AARCH64_INSN_COND_EQ = 0x0, /* == */
+       AARCH64_INSN_COND_NE = 0x1, /* != */
+       AARCH64_INSN_COND_CS = 0x2, /* unsigned >= */
+       AARCH64_INSN_COND_CC = 0x3, /* unsigned < */
+       AARCH64_INSN_COND_MI = 0x4, /* < 0 */
+       AARCH64_INSN_COND_PL = 0x5, /* >= 0 */
+       AARCH64_INSN_COND_VS = 0x6, /* overflow */
+       AARCH64_INSN_COND_VC = 0x7, /* no overflow */
+       AARCH64_INSN_COND_HI = 0x8, /* unsigned > */
+       AARCH64_INSN_COND_LS = 0x9, /* unsigned <= */
+       AARCH64_INSN_COND_GE = 0xa, /* signed >= */
+       AARCH64_INSN_COND_LT = 0xb, /* signed < */
+       AARCH64_INSN_COND_GT = 0xc, /* signed > */
+       AARCH64_INSN_COND_LE = 0xd, /* signed <= */
+       AARCH64_INSN_COND_AL = 0xe, /* always */
+};
+
+enum aarch64_insn_branch_type {
+       AARCH64_INSN_BRANCH_NOLINK,
+       AARCH64_INSN_BRANCH_LINK,
+       AARCH64_INSN_BRANCH_RETURN,
+       AARCH64_INSN_BRANCH_COMP_ZERO,
+       AARCH64_INSN_BRANCH_COMP_NONZERO,
+};
+
+enum aarch64_insn_size_type {
+       AARCH64_INSN_SIZE_8,
+       AARCH64_INSN_SIZE_16,
+       AARCH64_INSN_SIZE_32,
+       AARCH64_INSN_SIZE_64,
+};
+
+enum aarch64_insn_ldst_type {
+       AARCH64_INSN_LDST_LOAD_REG_OFFSET,
+       AARCH64_INSN_LDST_STORE_REG_OFFSET,
+       AARCH64_INSN_LDST_LOAD_PAIR_PRE_INDEX,
+       AARCH64_INSN_LDST_STORE_PAIR_PRE_INDEX,
+       AARCH64_INSN_LDST_LOAD_PAIR_POST_INDEX,
+       AARCH64_INSN_LDST_STORE_PAIR_POST_INDEX,
+       AARCH64_INSN_LDST_LOAD_EX,
+       AARCH64_INSN_LDST_STORE_EX,
+};
+
+enum aarch64_insn_adsb_type {
+       AARCH64_INSN_ADSB_ADD,
+       AARCH64_INSN_ADSB_SUB,
+       AARCH64_INSN_ADSB_ADD_SETFLAGS,
+       AARCH64_INSN_ADSB_SUB_SETFLAGS
+};
+
+enum aarch64_insn_movewide_type {
+       AARCH64_INSN_MOVEWIDE_ZERO,
+       AARCH64_INSN_MOVEWIDE_KEEP,
+       AARCH64_INSN_MOVEWIDE_INVERSE
+};
+
+enum aarch64_insn_bitfield_type {
+       AARCH64_INSN_BITFIELD_MOVE,
+       AARCH64_INSN_BITFIELD_MOVE_UNSIGNED,
+       AARCH64_INSN_BITFIELD_MOVE_SIGNED
+};
+
+enum aarch64_insn_data1_type {
+       AARCH64_INSN_DATA1_REVERSE_16,
+       AARCH64_INSN_DATA1_REVERSE_32,
+       AARCH64_INSN_DATA1_REVERSE_64,
+};
+
+enum aarch64_insn_data2_type {
+       AARCH64_INSN_DATA2_UDIV,
+       AARCH64_INSN_DATA2_SDIV,
+       AARCH64_INSN_DATA2_LSLV,
+       AARCH64_INSN_DATA2_LSRV,
+       AARCH64_INSN_DATA2_ASRV,
+       AARCH64_INSN_DATA2_RORV,
+};
+
+enum aarch64_insn_data3_type {
+       AARCH64_INSN_DATA3_MADD,
+       AARCH64_INSN_DATA3_MSUB,
+};
+
+enum aarch64_insn_logic_type {
+       AARCH64_INSN_LOGIC_AND,
+       AARCH64_INSN_LOGIC_BIC,
+       AARCH64_INSN_LOGIC_ORR,
+       AARCH64_INSN_LOGIC_ORN,
+       AARCH64_INSN_LOGIC_EOR,
+       AARCH64_INSN_LOGIC_EON,
+       AARCH64_INSN_LOGIC_AND_SETFLAGS,
+       AARCH64_INSN_LOGIC_BIC_SETFLAGS
+};
+
+enum aarch64_insn_prfm_type {
+       AARCH64_INSN_PRFM_TYPE_PLD,
+       AARCH64_INSN_PRFM_TYPE_PLI,
+       AARCH64_INSN_PRFM_TYPE_PST,
+};
+
+enum aarch64_insn_prfm_target {
+       AARCH64_INSN_PRFM_TARGET_L1,
+       AARCH64_INSN_PRFM_TARGET_L2,
+       AARCH64_INSN_PRFM_TARGET_L3,
+};
+
+enum aarch64_insn_prfm_policy {
+       AARCH64_INSN_PRFM_POLICY_KEEP,
+       AARCH64_INSN_PRFM_POLICY_STRM,
+};
+
+enum aarch64_insn_adr_type {
+       AARCH64_INSN_ADR_TYPE_ADRP,
+       AARCH64_INSN_ADR_TYPE_ADR,
+};
+
+#define        __AARCH64_INSN_FUNCS(abbr, mask, val)                           
\
+static __always_inline bool aarch64_insn_is_##abbr(u32 code)           \
+{                                                                      \
+       BUILD_BUG_ON(~(mask) & (val));                                  \
+       return (code & (mask)) == (val);                                \
+}                                                                      \
+static __always_inline u32 aarch64_insn_get_##abbr##_value(void)       \
+{                                                                      \
+       return (val);                                                   \
+}
+
+__AARCH64_INSN_FUNCS(adr,      0x9F000000, 0x10000000)
+__AARCH64_INSN_FUNCS(adrp,     0x9F000000, 0x90000000)
+__AARCH64_INSN_FUNCS(prfm,     0x3FC00000, 0x39800000)
+__AARCH64_INSN_FUNCS(prfm_lit, 0xFF000000, 0xD8000000)
+__AARCH64_INSN_FUNCS(store_imm,        0x3FC00000, 0x39000000)
+__AARCH64_INSN_FUNCS(load_imm, 0x3FC00000, 0x39400000)
+__AARCH64_INSN_FUNCS(store_pre,        0x3FE00C00, 0x38000C00)
+__AARCH64_INSN_FUNCS(load_pre, 0x3FE00C00, 0x38400C00)
+__AARCH64_INSN_FUNCS(store_post,       0x3FE00C00, 0x38000400)
+__AARCH64_INSN_FUNCS(load_post,        0x3FE00C00, 0x38400400)
+__AARCH64_INSN_FUNCS(str_reg,  0x3FE0EC00, 0x38206800)
+__AARCH64_INSN_FUNCS(ldadd,    0x3F20FC00, 0x38200000)
+__AARCH64_INSN_FUNCS(ldr_reg,  0x3FE0EC00, 0x38606800)
+__AARCH64_INSN_FUNCS(ldr_lit,  0xBF000000, 0x18000000)
+__AARCH64_INSN_FUNCS(ldrsw_lit,        0xFF000000, 0x98000000)
+__AARCH64_INSN_FUNCS(exclusive,        0x3F800000, 0x08000000)
+__AARCH64_INSN_FUNCS(load_ex,  0x3F400000, 0x08400000)
+__AARCH64_INSN_FUNCS(store_ex, 0x3F400000, 0x08000000)
+__AARCH64_INSN_FUNCS(stp,      0x7FC00000, 0x29000000)
+__AARCH64_INSN_FUNCS(ldp,      0x7FC00000, 0x29400000)
+__AARCH64_INSN_FUNCS(stp_post, 0x7FC00000, 0x28800000)
+__AARCH64_INSN_FUNCS(ldp_post, 0x7FC00000, 0x28C00000)
+__AARCH64_INSN_FUNCS(stp_pre,  0x7FC00000, 0x29800000)
+__AARCH64_INSN_FUNCS(ldp_pre,  0x7FC00000, 0x29C00000)
+__AARCH64_INSN_FUNCS(add_imm,  0x7F000000, 0x11000000)
+__AARCH64_INSN_FUNCS(adds_imm, 0x7F000000, 0x31000000)
+__AARCH64_INSN_FUNCS(sub_imm,  0x7F000000, 0x51000000)
+__AARCH64_INSN_FUNCS(subs_imm, 0x7F000000, 0x71000000)
+__AARCH64_INSN_FUNCS(movn,     0x7F800000, 0x12800000)
+__AARCH64_INSN_FUNCS(sbfm,     0x7F800000, 0x13000000)
+__AARCH64_INSN_FUNCS(bfm,      0x7F800000, 0x33000000)
+__AARCH64_INSN_FUNCS(movz,     0x7F800000, 0x52800000)
+__AARCH64_INSN_FUNCS(ubfm,     0x7F800000, 0x53000000)
+__AARCH64_INSN_FUNCS(movk,     0x7F800000, 0x72800000)
+__AARCH64_INSN_FUNCS(add,      0x7F200000, 0x0B000000)
+__AARCH64_INSN_FUNCS(adds,     0x7F200000, 0x2B000000)
+__AARCH64_INSN_FUNCS(sub,      0x7F200000, 0x4B000000)
+__AARCH64_INSN_FUNCS(subs,     0x7F200000, 0x6B000000)
+__AARCH64_INSN_FUNCS(madd,     0x7FE08000, 0x1B000000)
+__AARCH64_INSN_FUNCS(msub,     0x7FE08000, 0x1B008000)
+__AARCH64_INSN_FUNCS(udiv,     0x7FE0FC00, 0x1AC00800)
+__AARCH64_INSN_FUNCS(sdiv,     0x7FE0FC00, 0x1AC00C00)
+__AARCH64_INSN_FUNCS(lslv,     0x7FE0FC00, 0x1AC02000)
+__AARCH64_INSN_FUNCS(lsrv,     0x7FE0FC00, 0x1AC02400)
+__AARCH64_INSN_FUNCS(asrv,     0x7FE0FC00, 0x1AC02800)
+__AARCH64_INSN_FUNCS(rorv,     0x7FE0FC00, 0x1AC02C00)
+__AARCH64_INSN_FUNCS(rev16,    0x7FFFFC00, 0x5AC00400)
+__AARCH64_INSN_FUNCS(rev32,    0x7FFFFC00, 0x5AC00800)
+__AARCH64_INSN_FUNCS(rev64,    0x7FFFFC00, 0x5AC00C00)
+__AARCH64_INSN_FUNCS(and,      0x7F200000, 0x0A000000)
+__AARCH64_INSN_FUNCS(bic,      0x7F200000, 0x0A200000)
+__AARCH64_INSN_FUNCS(orr,      0x7F200000, 0x2A000000)
+__AARCH64_INSN_FUNCS(mov_reg,  0x7FE0FFE0, 0x2A0003E0)
+__AARCH64_INSN_FUNCS(orn,      0x7F200000, 0x2A200000)
+__AARCH64_INSN_FUNCS(eor,      0x7F200000, 0x4A000000)
+__AARCH64_INSN_FUNCS(eon,      0x7F200000, 0x4A200000)
+__AARCH64_INSN_FUNCS(ands,     0x7F200000, 0x6A000000)
+__AARCH64_INSN_FUNCS(bics,     0x7F200000, 0x6A200000)
+__AARCH64_INSN_FUNCS(and_imm,  0x7F800000, 0x12000000)
+__AARCH64_INSN_FUNCS(orr_imm,  0x7F800000, 0x32000000)
+__AARCH64_INSN_FUNCS(eor_imm,  0x7F800000, 0x52000000)
+__AARCH64_INSN_FUNCS(ands_imm, 0x7F800000, 0x72000000)
+__AARCH64_INSN_FUNCS(extr,     0x7FA00000, 0x13800000)
+__AARCH64_INSN_FUNCS(b,                0xFC000000, 0x14000000)
+__AARCH64_INSN_FUNCS(bl,       0xFC000000, 0x94000000)
+__AARCH64_INSN_FUNCS(cbz,      0x7F000000, 0x34000000)
+__AARCH64_INSN_FUNCS(cbnz,     0x7F000000, 0x35000000)
+__AARCH64_INSN_FUNCS(tbz,      0x7F000000, 0x36000000)
+__AARCH64_INSN_FUNCS(tbnz,     0x7F000000, 0x37000000)
+__AARCH64_INSN_FUNCS(bcond,    0xFF000010, 0x54000000)
+__AARCH64_INSN_FUNCS(svc,      0xFFE0001F, 0xD4000001)
+__AARCH64_INSN_FUNCS(hvc,      0xFFE0001F, 0xD4000002)
+__AARCH64_INSN_FUNCS(smc,      0xFFE0001F, 0xD4000003)
+__AARCH64_INSN_FUNCS(brk,      0xFFE0001F, 0xD4200000)
+__AARCH64_INSN_FUNCS(exception,        0xFF000000, 0xD4000000)
+__AARCH64_INSN_FUNCS(hint,     0xFFFFF01F, 0xD503201F)
+__AARCH64_INSN_FUNCS(br,       0xFFFFFC1F, 0xD61F0000)
+__AARCH64_INSN_FUNCS(br_auth,  0xFEFFF800, 0xD61F0800)
+__AARCH64_INSN_FUNCS(blr,      0xFFFFFC1F, 0xD63F0000)
+__AARCH64_INSN_FUNCS(blr_auth, 0xFEFFF800, 0xD63F0800)
+__AARCH64_INSN_FUNCS(ret,      0xFFFFFC1F, 0xD65F0000)
+__AARCH64_INSN_FUNCS(ret_auth, 0xFFFFFBFF, 0xD65F0BFF)
+__AARCH64_INSN_FUNCS(eret,     0xFFFFFFFF, 0xD69F03E0)
+__AARCH64_INSN_FUNCS(eret_auth,        0xFFFFFBFF, 0xD69F0BFF)
+__AARCH64_INSN_FUNCS(mrs,      0xFFF00000, 0xD5300000)
+__AARCH64_INSN_FUNCS(msr_imm,  0xFFF8F01F, 0xD500401F)
+__AARCH64_INSN_FUNCS(msr_reg,  0xFFF00000, 0xD5100000)
+__AARCH64_INSN_FUNCS(dmb,      0xFFFFF0FF, 0xD50330BF)
+__AARCH64_INSN_FUNCS(dsb,      0xFFFFF0FF, 0xD503309F)
+__AARCH64_INSN_FUNCS(isb,      0xFFFFF0FF, 0xD50330DF)
+
+#undef __AARCH64_INSN_FUNCS
+
+bool aarch64_insn_is_steppable_hint(u32 insn);
+bool aarch64_insn_is_branch_imm(u32 insn);
+
+static inline bool aarch64_insn_is_adr_adrp(u32 insn)
+{
+       return aarch64_insn_is_adr(insn) || aarch64_insn_is_adrp(insn);
+}
+
+static inline bool aarch64_insn_is_barrier(u32 insn)
+{
+       return aarch64_insn_is_dmb(insn) || aarch64_insn_is_dsb(insn) ||
+              aarch64_insn_is_isb(insn);
+}
+
+static inline bool aarch64_insn_is_store_single(u32 insn)
+{
+       return aarch64_insn_is_store_imm(insn) ||
+              aarch64_insn_is_store_pre(insn) ||
+              aarch64_insn_is_store_post(insn);
+}
+
+static inline bool aarch64_insn_is_store_pair(u32 insn)
+{
+       return aarch64_insn_is_stp(insn) ||
+              aarch64_insn_is_stp_pre(insn) ||
+              aarch64_insn_is_stp_post(insn);
+}
+
+static inline bool aarch64_insn_is_load_single(u32 insn)
+{
+       return aarch64_insn_is_load_imm(insn) ||
+              aarch64_insn_is_load_pre(insn) ||
+              aarch64_insn_is_load_post(insn);
+}
+
+static inline bool aarch64_insn_is_load_pair(u32 insn)
+{
+       return aarch64_insn_is_ldp(insn) ||
+              aarch64_insn_is_ldp_pre(insn) ||
+              aarch64_insn_is_ldp_post(insn);
+}
+
+enum aarch64_insn_encoding_class aarch64_get_insn_class(u32 insn);
+bool aarch64_insn_uses_literal(u32 insn);
+bool aarch64_insn_is_branch(u32 insn);
+u64 aarch64_insn_decode_immediate(enum aarch64_insn_imm_type type, u32 insn);
+u32 aarch64_insn_encode_immediate(enum aarch64_insn_imm_type type,
+                                 u32 insn, u64 imm);
+u32 aarch64_insn_decode_register(enum aarch64_insn_register_type type,
+                                u32 insn);
+u32 aarch64_insn_gen_branch_imm(unsigned long pc, unsigned long addr,
+                               enum aarch64_insn_branch_type type);
+u32 aarch64_insn_gen_comp_branch_imm(unsigned long pc, unsigned long addr,
+                                    enum aarch64_insn_register reg,
+                                    enum aarch64_insn_variant variant,
+                                    enum aarch64_insn_branch_type type);
+u32 aarch64_insn_gen_cond_branch_imm(unsigned long pc, unsigned long addr,
+                                    enum aarch64_insn_condition cond);
+u32 aarch64_insn_gen_hint(enum aarch64_insn_hint_cr_op op);
+u32 aarch64_insn_gen_nop(void);
+u32 aarch64_insn_gen_branch_reg(enum aarch64_insn_register reg,
+                               enum aarch64_insn_branch_type type);
+u32 aarch64_insn_gen_load_store_reg(enum aarch64_insn_register reg,
+                                   enum aarch64_insn_register base,
+                                   enum aarch64_insn_register offset,
+                                   enum aarch64_insn_size_type size,
+                                   enum aarch64_insn_ldst_type type);
+u32 aarch64_insn_gen_load_store_pair(enum aarch64_insn_register reg1,
+                                    enum aarch64_insn_register reg2,
+                                    enum aarch64_insn_register base,
+                                    int offset,
+                                    enum aarch64_insn_variant variant,
+                                    enum aarch64_insn_ldst_type type);
+u32 aarch64_insn_gen_load_store_ex(enum aarch64_insn_register reg,
+                                  enum aarch64_insn_register base,
+                                  enum aarch64_insn_register state,
+                                  enum aarch64_insn_size_type size,
+                                  enum aarch64_insn_ldst_type type);
+u32 aarch64_insn_gen_ldadd(enum aarch64_insn_register result,
+                          enum aarch64_insn_register address,
+                          enum aarch64_insn_register value,
+                          enum aarch64_insn_size_type size);
+u32 aarch64_insn_gen_stadd(enum aarch64_insn_register address,
+                          enum aarch64_insn_register value,
+                          enum aarch64_insn_size_type size);
+u32 aarch64_insn_gen_add_sub_imm(enum aarch64_insn_register dst,
+                                enum aarch64_insn_register src,
+                                int imm, enum aarch64_insn_variant variant,
+                                enum aarch64_insn_adsb_type type);
+u32 aarch64_insn_gen_adr(unsigned long pc, unsigned long addr,
+                        enum aarch64_insn_register reg,
+                        enum aarch64_insn_adr_type type);
+u32 aarch64_insn_gen_bitfield(enum aarch64_insn_register dst,
+                             enum aarch64_insn_register src,
+                             int immr, int imms,
+                             enum aarch64_insn_variant variant,
+                             enum aarch64_insn_bitfield_type type);
+u32 aarch64_insn_gen_movewide(enum aarch64_insn_register dst,
+                             int imm, int shift,
+                             enum aarch64_insn_variant variant,
+                             enum aarch64_insn_movewide_type type);
+u32 aarch64_insn_gen_add_sub_shifted_reg(enum aarch64_insn_register dst,
+                                        enum aarch64_insn_register src,
+                                        enum aarch64_insn_register reg,
+                                        int shift,
+                                        enum aarch64_insn_variant variant,
+                                        enum aarch64_insn_adsb_type type);
+u32 aarch64_insn_gen_data1(enum aarch64_insn_register dst,
+                          enum aarch64_insn_register src,
+                          enum aarch64_insn_variant variant,
+                          enum aarch64_insn_data1_type type);
+u32 aarch64_insn_gen_data2(enum aarch64_insn_register dst,
+                          enum aarch64_insn_register src,
+                          enum aarch64_insn_register reg,
+                          enum aarch64_insn_variant variant,
+                          enum aarch64_insn_data2_type type);
+u32 aarch64_insn_gen_data3(enum aarch64_insn_register dst,
+                          enum aarch64_insn_register src,
+                          enum aarch64_insn_register reg1,
+                          enum aarch64_insn_register reg2,
+                          enum aarch64_insn_variant variant,
+                          enum aarch64_insn_data3_type type);
+u32 aarch64_insn_gen_logical_shifted_reg(enum aarch64_insn_register dst,
+                                        enum aarch64_insn_register src,
+                                        enum aarch64_insn_register reg,
+                                        int shift,
+                                        enum aarch64_insn_variant variant,
+                                        enum aarch64_insn_logic_type type);
+u32 aarch64_insn_gen_move_reg(enum aarch64_insn_register dst,
+                             enum aarch64_insn_register src,
+                             enum aarch64_insn_variant variant);
+u32 aarch64_insn_gen_logical_immediate(enum aarch64_insn_logic_type type,
+                                      enum aarch64_insn_variant variant,
+                                      enum aarch64_insn_register Rn,
+                                      enum aarch64_insn_register Rd,
+                                      u64 imm);
+u32 aarch64_insn_gen_extr(enum aarch64_insn_variant variant,
+                         enum aarch64_insn_register Rm,
+                         enum aarch64_insn_register Rn,
+                         enum aarch64_insn_register Rd,
+                         u8 lsb);
+u32 aarch64_insn_gen_prefetch(enum aarch64_insn_register base,
+                             enum aarch64_insn_prfm_type type,
+                             enum aarch64_insn_prfm_target target,
+                             enum aarch64_insn_prfm_policy policy);
+s32 aarch64_get_branch_offset(u32 insn);
+u32 aarch64_set_branch_offset(u32 insn, s32 offset);
+
+s32 aarch64_insn_adrp_get_offset(u32 insn);
+u32 aarch64_insn_adrp_set_offset(u32 insn, s32 offset);
+
+u32 aarch64_insn_extract_system_reg(u32 insn);
+
+
+#endif /* !__ASSEMBLY__ */
+#endif /* __ASM_AARCH64_INSN_H */
diff --git a/tools/arch/arm64/lib/aarch64-insn.c 
b/tools/arch/arm64/lib/aarch64-insn.c
new file mode 100644
index 000000000000..18e0b66a8284
--- /dev/null
+++ b/tools/arch/arm64/lib/aarch64-insn.c
@@ -0,0 +1,1425 @@
+// SPDX-License-Identifier: GPL-2.0-only
+
+#include <linux/bitops.h>
+#include <linux/bug.h>
+#include <linux/printk.h>
+#include <linux/sizes.h>
+#include <linux/types.h>
+
+#include <asm/aarch64-insn.h>
+#include <asm/errno.h>
+
+#define AARCH64_INSN_SF_BIT    BIT(31)
+#define AARCH64_INSN_N_BIT     BIT(22)
+#define AARCH64_INSN_LSL_12    BIT(22)
+
+static const int aarch64_insn_encoding_class[] = {
+       AARCH64_INSN_CLS_UNKNOWN,
+       AARCH64_INSN_CLS_UNKNOWN,
+       AARCH64_INSN_CLS_SVE,
+       AARCH64_INSN_CLS_UNKNOWN,
+       AARCH64_INSN_CLS_LDST,
+       AARCH64_INSN_CLS_DP_REG,
+       AARCH64_INSN_CLS_LDST,
+       AARCH64_INSN_CLS_DP_FPSIMD,
+       AARCH64_INSN_CLS_DP_IMM,
+       AARCH64_INSN_CLS_DP_IMM,
+       AARCH64_INSN_CLS_BR_SYS,
+       AARCH64_INSN_CLS_BR_SYS,
+       AARCH64_INSN_CLS_LDST,
+       AARCH64_INSN_CLS_DP_REG,
+       AARCH64_INSN_CLS_LDST,
+       AARCH64_INSN_CLS_DP_FPSIMD,
+};
+
+enum aarch64_insn_encoding_class __kprobes aarch64_get_insn_class(u32 insn)
+{
+       return aarch64_insn_encoding_class[(insn >> 25) & 0xf];
+}
+
+bool __kprobes aarch64_insn_is_steppable_hint(u32 insn)
+{
+       if (!aarch64_insn_is_hint(insn))
+               return false;
+
+       switch (insn & 0xFE0) {
+       case AARCH64_INSN_HINT_XPACLRI:
+       case AARCH64_INSN_HINT_PACIA_1716:
+       case AARCH64_INSN_HINT_PACIB_1716:
+       case AARCH64_INSN_HINT_PACIAZ:
+       case AARCH64_INSN_HINT_PACIASP:
+       case AARCH64_INSN_HINT_PACIBZ:
+       case AARCH64_INSN_HINT_PACIBSP:
+       case AARCH64_INSN_HINT_BTI:
+       case AARCH64_INSN_HINT_BTIC:
+       case AARCH64_INSN_HINT_BTIJ:
+       case AARCH64_INSN_HINT_BTIJC:
+       case AARCH64_INSN_HINT_NOP:
+               return true;
+       default:
+               return false;
+       }
+}
+
+bool aarch64_insn_is_branch_imm(u32 insn)
+{
+       return (aarch64_insn_is_b(insn) || aarch64_insn_is_bl(insn) ||
+               aarch64_insn_is_tbz(insn) || aarch64_insn_is_tbnz(insn) ||
+               aarch64_insn_is_cbz(insn) || aarch64_insn_is_cbnz(insn) ||
+               aarch64_insn_is_bcond(insn));
+}
+
+bool __kprobes aarch64_insn_uses_literal(u32 insn)
+{
+       /* ldr/ldrsw (literal), prfm */
+
+       return aarch64_insn_is_ldr_lit(insn) ||
+               aarch64_insn_is_ldrsw_lit(insn) ||
+               aarch64_insn_is_adr_adrp(insn) ||
+               aarch64_insn_is_prfm_lit(insn);
+}
+
+bool __kprobes aarch64_insn_is_branch(u32 insn)
+{
+       /* b, bl, cb*, tb*, ret*, b.cond, br*, blr* */
+
+       return aarch64_insn_is_b(insn) ||
+               aarch64_insn_is_bl(insn) ||
+               aarch64_insn_is_cbz(insn) ||
+               aarch64_insn_is_cbnz(insn) ||
+               aarch64_insn_is_tbz(insn) ||
+               aarch64_insn_is_tbnz(insn) ||
+               aarch64_insn_is_ret(insn) ||
+               aarch64_insn_is_ret_auth(insn) ||
+               aarch64_insn_is_br(insn) ||
+               aarch64_insn_is_br_auth(insn) ||
+               aarch64_insn_is_blr(insn) ||
+               aarch64_insn_is_blr_auth(insn) ||
+               aarch64_insn_is_bcond(insn);
+}
+
+static int __kprobes aarch64_get_imm_shift_mask(enum aarch64_insn_imm_type 
type,
+                                               u32 *maskp, int *shiftp)
+{
+       u32 mask;
+       int shift;
+
+       switch (type) {
+       case AARCH64_INSN_IMM_26:
+               mask = BIT(26) - 1;
+               shift = 0;
+               break;
+       case AARCH64_INSN_IMM_19:
+               mask = BIT(19) - 1;
+               shift = 5;
+               break;
+       case AARCH64_INSN_IMM_16:
+               mask = BIT(16) - 1;
+               shift = 5;
+               break;
+       case AARCH64_INSN_IMM_14:
+               mask = BIT(14) - 1;
+               shift = 5;
+               break;
+       case AARCH64_INSN_IMM_12:
+               mask = BIT(12) - 1;
+               shift = 10;
+               break;
+       case AARCH64_INSN_IMM_9:
+               mask = BIT(9) - 1;
+               shift = 12;
+               break;
+       case AARCH64_INSN_IMM_7:
+               mask = BIT(7) - 1;
+               shift = 15;
+               break;
+       case AARCH64_INSN_IMM_6:
+       case AARCH64_INSN_IMM_S:
+               mask = BIT(6) - 1;
+               shift = 10;
+               break;
+       case AARCH64_INSN_IMM_R:
+               mask = BIT(6) - 1;
+               shift = 16;
+               break;
+       case AARCH64_INSN_IMM_N:
+               mask = 1;
+               shift = 22;
+               break;
+       default:
+               return -EINVAL;
+       }
+
+       *maskp = mask;
+       *shiftp = shift;
+
+       return 0;
+}
+
+#define ADR_IMM_HILOSPLIT      2
+#define ADR_IMM_SIZE           SZ_2M
+#define ADR_IMM_LOMASK         ((1 << ADR_IMM_HILOSPLIT) - 1)
+#define ADR_IMM_HIMASK         ((ADR_IMM_SIZE >> ADR_IMM_HILOSPLIT) - 1)
+#define ADR_IMM_LOSHIFT                29
+#define ADR_IMM_HISHIFT                5
+
+u64 aarch64_insn_decode_immediate(enum aarch64_insn_imm_type type, u32 insn)
+{
+       u32 immlo, immhi, mask;
+       int shift;
+
+       switch (type) {
+       case AARCH64_INSN_IMM_ADR:
+               shift = 0;
+               immlo = (insn >> ADR_IMM_LOSHIFT) & ADR_IMM_LOMASK;
+               immhi = (insn >> ADR_IMM_HISHIFT) & ADR_IMM_HIMASK;
+               insn = (immhi << ADR_IMM_HILOSPLIT) | immlo;
+               mask = ADR_IMM_SIZE - 1;
+               break;
+       default:
+               if (aarch64_get_imm_shift_mask(type, &mask, &shift) < 0) {
+                       pr_err("aarch64_insn_decode_immediate: unknown 
immediate encoding %d\n",
+                              type);
+                       return 0;
+               }
+       }
+
+       return (insn >> shift) & mask;
+}
+
+u32 __kprobes aarch64_insn_encode_immediate(enum aarch64_insn_imm_type type,
+                                 u32 insn, u64 imm)
+{
+       u32 immlo, immhi, mask;
+       int shift;
+
+       if (insn == AARCH64_BREAK_FAULT)
+               return AARCH64_BREAK_FAULT;
+
+       switch (type) {
+       case AARCH64_INSN_IMM_ADR:
+               shift = 0;
+               immlo = (imm & ADR_IMM_LOMASK) << ADR_IMM_LOSHIFT;
+               imm >>= ADR_IMM_HILOSPLIT;
+               immhi = (imm & ADR_IMM_HIMASK) << ADR_IMM_HISHIFT;
+               imm = immlo | immhi;
+               mask = ((ADR_IMM_LOMASK << ADR_IMM_LOSHIFT) |
+                       (ADR_IMM_HIMASK << ADR_IMM_HISHIFT));
+               break;
+       default:
+               if (aarch64_get_imm_shift_mask(type, &mask, &shift) < 0) {
+                       pr_err("aarch64_insn_encode_immediate: unknown 
immediate encoding %d\n",
+                              type);
+                       return AARCH64_BREAK_FAULT;
+               }
+       }
+
+       /* Update the immediate field. */
+       insn &= ~(mask << shift);
+       insn |= (imm & mask) << shift;
+
+       return insn;
+}
+
+u32 aarch64_insn_decode_register(enum aarch64_insn_register_type type,
+                                       u32 insn)
+{
+       int shift;
+
+       switch (type) {
+       case AARCH64_INSN_REGTYPE_RT:
+       case AARCH64_INSN_REGTYPE_RD:
+               shift = 0;
+               break;
+       case AARCH64_INSN_REGTYPE_RN:
+               shift = 5;
+               break;
+       case AARCH64_INSN_REGTYPE_RT2:
+       case AARCH64_INSN_REGTYPE_RA:
+               shift = 10;
+               break;
+       case AARCH64_INSN_REGTYPE_RM:
+               shift = 16;
+               break;
+       default:
+               pr_err("%s: unknown register type encoding %d\n", __func__,
+                      type);
+               return 0;
+       }
+
+       return (insn >> shift) & GENMASK(4, 0);
+}
+
+static u32 aarch64_insn_encode_register(enum aarch64_insn_register_type type,
+                                       u32 insn,
+                                       enum aarch64_insn_register reg)
+{
+       int shift;
+
+       if (insn == AARCH64_BREAK_FAULT)
+               return AARCH64_BREAK_FAULT;
+
+       if (reg < AARCH64_INSN_REG_0 || reg > AARCH64_INSN_REG_SP) {
+               pr_err("%s: unknown register encoding %d\n", __func__, reg);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       switch (type) {
+       case AARCH64_INSN_REGTYPE_RT:
+       case AARCH64_INSN_REGTYPE_RD:
+               shift = 0;
+               break;
+       case AARCH64_INSN_REGTYPE_RN:
+               shift = 5;
+               break;
+       case AARCH64_INSN_REGTYPE_RT2:
+       case AARCH64_INSN_REGTYPE_RA:
+               shift = 10;
+               break;
+       case AARCH64_INSN_REGTYPE_RM:
+       case AARCH64_INSN_REGTYPE_RS:
+               shift = 16;
+               break;
+       default:
+               pr_err("%s: unknown register type encoding %d\n", __func__,
+                      type);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       insn &= ~(GENMASK(4, 0) << shift);
+       insn |= reg << shift;
+
+       return insn;
+}
+
+static u32 aarch64_insn_encode_ldst_size(enum aarch64_insn_size_type type,
+                                        u32 insn)
+{
+       u32 size;
+
+       switch (type) {
+       case AARCH64_INSN_SIZE_8:
+               size = 0;
+               break;
+       case AARCH64_INSN_SIZE_16:
+               size = 1;
+               break;
+       case AARCH64_INSN_SIZE_32:
+               size = 2;
+               break;
+       case AARCH64_INSN_SIZE_64:
+               size = 3;
+               break;
+       default:
+               pr_err("%s: unknown size encoding %d\n", __func__, type);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       insn &= ~GENMASK(31, 30);
+       insn |= size << 30;
+
+       return insn;
+}
+
+static inline long branch_imm_common(unsigned long pc, unsigned long addr,
+                                    long range)
+{
+       long offset;
+
+       if ((pc & 0x3) || (addr & 0x3)) {
+               pr_err("%s: A64 instructions must be word aligned\n", __func__);
+               return range;
+       }
+
+       offset = ((long)addr - (long)pc);
+
+       if (offset < -range || offset >= range) {
+               pr_err("%s: offset out of range\n", __func__);
+               return range;
+       }
+
+       return offset;
+}
+
+u32 __kprobes aarch64_insn_gen_branch_imm(unsigned long pc, unsigned long addr,
+                                         enum aarch64_insn_branch_type type)
+{
+       u32 insn;
+       long offset;
+
+       /*
+        * B/BL support [-128M, 128M) offset
+        * ARM64 virtual address arrangement guarantees all kernel and module
+        * texts are within +/-128M.
+        */
+       offset = branch_imm_common(pc, addr, SZ_128M);
+       if (offset >= SZ_128M)
+               return AARCH64_BREAK_FAULT;
+
+       switch (type) {
+       case AARCH64_INSN_BRANCH_LINK:
+               insn = aarch64_insn_get_bl_value();
+               break;
+       case AARCH64_INSN_BRANCH_NOLINK:
+               insn = aarch64_insn_get_b_value();
+               break;
+       default:
+               pr_err("%s: unknown branch encoding %d\n", __func__, type);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       return aarch64_insn_encode_immediate(AARCH64_INSN_IMM_26, insn,
+                                            offset >> 2);
+}
+
+u32 aarch64_insn_gen_comp_branch_imm(unsigned long pc, unsigned long addr,
+                                    enum aarch64_insn_register reg,
+                                    enum aarch64_insn_variant variant,
+                                    enum aarch64_insn_branch_type type)
+{
+       u32 insn;
+       long offset;
+
+       offset = branch_imm_common(pc, addr, SZ_1M);
+       if (offset >= SZ_1M)
+               return AARCH64_BREAK_FAULT;
+
+       switch (type) {
+       case AARCH64_INSN_BRANCH_COMP_ZERO:
+               insn = aarch64_insn_get_cbz_value();
+               break;
+       case AARCH64_INSN_BRANCH_COMP_NONZERO:
+               insn = aarch64_insn_get_cbnz_value();
+               break;
+       default:
+               pr_err("%s: unknown branch encoding %d\n", __func__, type);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       switch (variant) {
+       case AARCH64_INSN_VARIANT_32BIT:
+               break;
+       case AARCH64_INSN_VARIANT_64BIT:
+               insn |= AARCH64_INSN_SF_BIT;
+               break;
+       default:
+               pr_err("%s: unknown variant encoding %d\n", __func__, variant);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RT, insn, reg);
+
+       return aarch64_insn_encode_immediate(AARCH64_INSN_IMM_19, insn,
+                                            offset >> 2);
+}
+
+u32 aarch64_insn_gen_cond_branch_imm(unsigned long pc, unsigned long addr,
+                                    enum aarch64_insn_condition cond)
+{
+       u32 insn;
+       long offset;
+
+       offset = branch_imm_common(pc, addr, SZ_1M);
+
+       insn = aarch64_insn_get_bcond_value();
+
+       if (cond < AARCH64_INSN_COND_EQ || cond > AARCH64_INSN_COND_AL) {
+               pr_err("%s: unknown condition encoding %d\n", __func__, cond);
+               return AARCH64_BREAK_FAULT;
+       }
+       insn |= cond;
+
+       return aarch64_insn_encode_immediate(AARCH64_INSN_IMM_19, insn,
+                                            offset >> 2);
+}
+
+u32 __kprobes aarch64_insn_gen_hint(enum aarch64_insn_hint_cr_op op)
+{
+       return aarch64_insn_get_hint_value() | op;
+}
+
+u32 __kprobes aarch64_insn_gen_nop(void)
+{
+       return aarch64_insn_gen_hint(AARCH64_INSN_HINT_NOP);
+}
+
+u32 aarch64_insn_gen_branch_reg(enum aarch64_insn_register reg,
+                               enum aarch64_insn_branch_type type)
+{
+       u32 insn;
+
+       switch (type) {
+       case AARCH64_INSN_BRANCH_NOLINK:
+               insn = aarch64_insn_get_br_value();
+               break;
+       case AARCH64_INSN_BRANCH_LINK:
+               insn = aarch64_insn_get_blr_value();
+               break;
+       case AARCH64_INSN_BRANCH_RETURN:
+               insn = aarch64_insn_get_ret_value();
+               break;
+       default:
+               pr_err("%s: unknown branch encoding %d\n", __func__, type);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       return aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RN, insn, reg);
+}
+
+u32 aarch64_insn_gen_load_store_reg(enum aarch64_insn_register reg,
+                                   enum aarch64_insn_register base,
+                                   enum aarch64_insn_register offset,
+                                   enum aarch64_insn_size_type size,
+                                   enum aarch64_insn_ldst_type type)
+{
+       u32 insn;
+
+       switch (type) {
+       case AARCH64_INSN_LDST_LOAD_REG_OFFSET:
+               insn = aarch64_insn_get_ldr_reg_value();
+               break;
+       case AARCH64_INSN_LDST_STORE_REG_OFFSET:
+               insn = aarch64_insn_get_str_reg_value();
+               break;
+       default:
+               pr_err("%s: unknown load/store encoding %d\n", __func__, type);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       insn = aarch64_insn_encode_ldst_size(size, insn);
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RT, insn, reg);
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RN, insn,
+                                           base);
+
+       return aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RM, insn,
+                                           offset);
+}
+
+u32 aarch64_insn_gen_load_store_pair(enum aarch64_insn_register reg1,
+                                    enum aarch64_insn_register reg2,
+                                    enum aarch64_insn_register base,
+                                    int offset,
+                                    enum aarch64_insn_variant variant,
+                                    enum aarch64_insn_ldst_type type)
+{
+       u32 insn;
+       int shift;
+
+       switch (type) {
+       case AARCH64_INSN_LDST_LOAD_PAIR_PRE_INDEX:
+               insn = aarch64_insn_get_ldp_pre_value();
+               break;
+       case AARCH64_INSN_LDST_STORE_PAIR_PRE_INDEX:
+               insn = aarch64_insn_get_stp_pre_value();
+               break;
+       case AARCH64_INSN_LDST_LOAD_PAIR_POST_INDEX:
+               insn = aarch64_insn_get_ldp_post_value();
+               break;
+       case AARCH64_INSN_LDST_STORE_PAIR_POST_INDEX:
+               insn = aarch64_insn_get_stp_post_value();
+               break;
+       default:
+               pr_err("%s: unknown load/store encoding %d\n", __func__, type);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       switch (variant) {
+       case AARCH64_INSN_VARIANT_32BIT:
+               if ((offset & 0x3) || (offset < -256) || (offset > 252)) {
+                       pr_err("%s: offset must be multiples of 4 in the range 
of [-256, 252] %d\n",
+                              __func__, offset);
+                       return AARCH64_BREAK_FAULT;
+               }
+               shift = 2;
+               break;
+       case AARCH64_INSN_VARIANT_64BIT:
+               if ((offset & 0x7) || (offset < -512) || (offset > 504)) {
+                       pr_err("%s: offset must be multiples of 8 in the range 
of [-512, 504] %d\n",
+                              __func__, offset);
+                       return AARCH64_BREAK_FAULT;
+               }
+               shift = 3;
+               insn |= AARCH64_INSN_SF_BIT;
+               break;
+       default:
+               pr_err("%s: unknown variant encoding %d\n", __func__, variant);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RT, insn,
+                                           reg1);
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RT2, insn,
+                                           reg2);
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RN, insn,
+                                           base);
+
+       return aarch64_insn_encode_immediate(AARCH64_INSN_IMM_7, insn,
+                                            offset >> shift);
+}
+
+u32 aarch64_insn_gen_load_store_ex(enum aarch64_insn_register reg,
+                                  enum aarch64_insn_register base,
+                                  enum aarch64_insn_register state,
+                                  enum aarch64_insn_size_type size,
+                                  enum aarch64_insn_ldst_type type)
+{
+       u32 insn;
+
+       switch (type) {
+       case AARCH64_INSN_LDST_LOAD_EX:
+               insn = aarch64_insn_get_load_ex_value();
+               break;
+       case AARCH64_INSN_LDST_STORE_EX:
+               insn = aarch64_insn_get_store_ex_value();
+               break;
+       default:
+               pr_err("%s: unknown load/store exclusive encoding %d\n", 
__func__, type);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       insn = aarch64_insn_encode_ldst_size(size, insn);
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RT, insn,
+                                           reg);
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RN, insn,
+                                           base);
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RT2, insn,
+                                           AARCH64_INSN_REG_ZR);
+
+       return aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RS, insn,
+                                           state);
+}
+
+u32 aarch64_insn_gen_ldadd(enum aarch64_insn_register result,
+                          enum aarch64_insn_register address,
+                          enum aarch64_insn_register value,
+                          enum aarch64_insn_size_type size)
+{
+       u32 insn = aarch64_insn_get_ldadd_value();
+
+       switch (size) {
+       case AARCH64_INSN_SIZE_32:
+       case AARCH64_INSN_SIZE_64:
+               break;
+       default:
+               pr_err("%s: unimplemented size encoding %d\n", __func__, size);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       insn = aarch64_insn_encode_ldst_size(size, insn);
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RT, insn,
+                                           result);
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RN, insn,
+                                           address);
+
+       return aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RS, insn,
+                                           value);
+}
+
+u32 aarch64_insn_gen_stadd(enum aarch64_insn_register address,
+                          enum aarch64_insn_register value,
+                          enum aarch64_insn_size_type size)
+{
+       /*
+        * STADD is simply encoded as an alias for LDADD with XZR as
+        * the destination register.
+        */
+       return aarch64_insn_gen_ldadd(AARCH64_INSN_REG_ZR, address,
+                                     value, size);
+}
+
+static u32 aarch64_insn_encode_prfm_imm(enum aarch64_insn_prfm_type type,
+                                       enum aarch64_insn_prfm_target target,
+                                       enum aarch64_insn_prfm_policy policy,
+                                       u32 insn)
+{
+       u32 imm_type = 0, imm_target = 0, imm_policy = 0;
+
+       switch (type) {
+       case AARCH64_INSN_PRFM_TYPE_PLD:
+               break;
+       case AARCH64_INSN_PRFM_TYPE_PLI:
+               imm_type = BIT(0);
+               break;
+       case AARCH64_INSN_PRFM_TYPE_PST:
+               imm_type = BIT(1);
+               break;
+       default:
+               pr_err("%s: unknown prfm type encoding %d\n", __func__, type);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       switch (target) {
+       case AARCH64_INSN_PRFM_TARGET_L1:
+               break;
+       case AARCH64_INSN_PRFM_TARGET_L2:
+               imm_target = BIT(0);
+               break;
+       case AARCH64_INSN_PRFM_TARGET_L3:
+               imm_target = BIT(1);
+               break;
+       default:
+               pr_err("%s: unknown prfm target encoding %d\n", __func__, 
target);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       switch (policy) {
+       case AARCH64_INSN_PRFM_POLICY_KEEP:
+               break;
+       case AARCH64_INSN_PRFM_POLICY_STRM:
+               imm_policy = BIT(0);
+               break;
+       default:
+               pr_err("%s: unknown prfm policy encoding %d\n", __func__, 
policy);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       /* In this case, imm5 is encoded into Rt field. */
+       insn &= ~GENMASK(4, 0);
+       insn |= imm_policy | (imm_target << 1) | (imm_type << 3);
+
+       return insn;
+}
+
+u32 aarch64_insn_gen_prefetch(enum aarch64_insn_register base,
+                             enum aarch64_insn_prfm_type type,
+                             enum aarch64_insn_prfm_target target,
+                             enum aarch64_insn_prfm_policy policy)
+{
+       u32 insn = aarch64_insn_get_prfm_value();
+
+       insn = aarch64_insn_encode_ldst_size(AARCH64_INSN_SIZE_64, insn);
+
+       insn = aarch64_insn_encode_prfm_imm(type, target, policy, insn);
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RN, insn,
+                                           base);
+
+       return aarch64_insn_encode_immediate(AARCH64_INSN_IMM_12, insn, 0);
+}
+
+u32 aarch64_insn_gen_add_sub_imm(enum aarch64_insn_register dst,
+                                enum aarch64_insn_register src,
+                                int imm, enum aarch64_insn_variant variant,
+                                enum aarch64_insn_adsb_type type)
+{
+       u32 insn;
+
+       switch (type) {
+       case AARCH64_INSN_ADSB_ADD:
+               insn = aarch64_insn_get_add_imm_value();
+               break;
+       case AARCH64_INSN_ADSB_SUB:
+               insn = aarch64_insn_get_sub_imm_value();
+               break;
+       case AARCH64_INSN_ADSB_ADD_SETFLAGS:
+               insn = aarch64_insn_get_adds_imm_value();
+               break;
+       case AARCH64_INSN_ADSB_SUB_SETFLAGS:
+               insn = aarch64_insn_get_subs_imm_value();
+               break;
+       default:
+               pr_err("%s: unknown add/sub encoding %d\n", __func__, type);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       switch (variant) {
+       case AARCH64_INSN_VARIANT_32BIT:
+               break;
+       case AARCH64_INSN_VARIANT_64BIT:
+               insn |= AARCH64_INSN_SF_BIT;
+               break;
+       default:
+               pr_err("%s: unknown variant encoding %d\n", __func__, variant);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       /* We can't encode more than a 24bit value (12bit + 12bit shift) */
+       if (imm & ~(BIT(24) - 1))
+               goto out;
+
+       /* If we have something in the top 12 bits... */
+       if (imm & ~(SZ_4K - 1)) {
+               /* ... and in the low 12 bits -> error */
+               if (imm & (SZ_4K - 1))
+                       goto out;
+
+               imm >>= 12;
+               insn |= AARCH64_INSN_LSL_12;
+       }
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RD, insn, dst);
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RN, insn, src);
+
+       return aarch64_insn_encode_immediate(AARCH64_INSN_IMM_12, insn, imm);
+
+out:
+       pr_err("%s: invalid immediate encoding %d\n", __func__, imm);
+       return AARCH64_BREAK_FAULT;
+}
+
+u32 aarch64_insn_gen_bitfield(enum aarch64_insn_register dst,
+                             enum aarch64_insn_register src,
+                             int immr, int imms,
+                             enum aarch64_insn_variant variant,
+                             enum aarch64_insn_bitfield_type type)
+{
+       u32 insn;
+       u32 mask;
+
+       switch (type) {
+       case AARCH64_INSN_BITFIELD_MOVE:
+               insn = aarch64_insn_get_bfm_value();
+               break;
+       case AARCH64_INSN_BITFIELD_MOVE_UNSIGNED:
+               insn = aarch64_insn_get_ubfm_value();
+               break;
+       case AARCH64_INSN_BITFIELD_MOVE_SIGNED:
+               insn = aarch64_insn_get_sbfm_value();
+               break;
+       default:
+               pr_err("%s: unknown bitfield encoding %d\n", __func__, type);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       switch (variant) {
+       case AARCH64_INSN_VARIANT_32BIT:
+               mask = GENMASK(4, 0);
+               break;
+       case AARCH64_INSN_VARIANT_64BIT:
+               insn |= AARCH64_INSN_SF_BIT | AARCH64_INSN_N_BIT;
+               mask = GENMASK(5, 0);
+               break;
+       default:
+               pr_err("%s: unknown variant encoding %d\n", __func__, variant);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       if (immr & ~mask) {
+               pr_err("%s: invalid immr encoding %d\n", __func__, immr);
+               return AARCH64_BREAK_FAULT;
+       }
+       if (imms & ~mask) {
+               pr_err("%s: invalid imms encoding %d\n", __func__, imms);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RD, insn, dst);
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RN, insn, src);
+
+       insn = aarch64_insn_encode_immediate(AARCH64_INSN_IMM_R, insn, immr);
+
+       return aarch64_insn_encode_immediate(AARCH64_INSN_IMM_S, insn, imms);
+}
+
+u32 aarch64_insn_gen_movewide(enum aarch64_insn_register dst,
+                             int imm, int shift,
+                             enum aarch64_insn_variant variant,
+                             enum aarch64_insn_movewide_type type)
+{
+       u32 insn;
+
+       switch (type) {
+       case AARCH64_INSN_MOVEWIDE_ZERO:
+               insn = aarch64_insn_get_movz_value();
+               break;
+       case AARCH64_INSN_MOVEWIDE_KEEP:
+               insn = aarch64_insn_get_movk_value();
+               break;
+       case AARCH64_INSN_MOVEWIDE_INVERSE:
+               insn = aarch64_insn_get_movn_value();
+               break;
+       default:
+               pr_err("%s: unknown movewide encoding %d\n", __func__, type);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       if (imm & ~(SZ_64K - 1)) {
+               pr_err("%s: invalid immediate encoding %d\n", __func__, imm);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       switch (variant) {
+       case AARCH64_INSN_VARIANT_32BIT:
+               if (shift != 0 && shift != 16) {
+                       pr_err("%s: invalid shift encoding %d\n", __func__,
+                              shift);
+                       return AARCH64_BREAK_FAULT;
+               }
+               break;
+       case AARCH64_INSN_VARIANT_64BIT:
+               insn |= AARCH64_INSN_SF_BIT;
+               if (shift != 0 && shift != 16 && shift != 32 && shift != 48) {
+                       pr_err("%s: invalid shift encoding %d\n", __func__,
+                              shift);
+                       return AARCH64_BREAK_FAULT;
+               }
+               break;
+       default:
+               pr_err("%s: unknown variant encoding %d\n", __func__, variant);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       insn |= (shift >> 4) << 21;
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RD, insn, dst);
+
+       return aarch64_insn_encode_immediate(AARCH64_INSN_IMM_16, insn, imm);
+}
+
+u32 aarch64_insn_gen_add_sub_shifted_reg(enum aarch64_insn_register dst,
+                                        enum aarch64_insn_register src,
+                                        enum aarch64_insn_register reg,
+                                        int shift,
+                                        enum aarch64_insn_variant variant,
+                                        enum aarch64_insn_adsb_type type)
+{
+       u32 insn;
+
+       switch (type) {
+       case AARCH64_INSN_ADSB_ADD:
+               insn = aarch64_insn_get_add_value();
+               break;
+       case AARCH64_INSN_ADSB_SUB:
+               insn = aarch64_insn_get_sub_value();
+               break;
+       case AARCH64_INSN_ADSB_ADD_SETFLAGS:
+               insn = aarch64_insn_get_adds_value();
+               break;
+       case AARCH64_INSN_ADSB_SUB_SETFLAGS:
+               insn = aarch64_insn_get_subs_value();
+               break;
+       default:
+               pr_err("%s: unknown add/sub encoding %d\n", __func__, type);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       switch (variant) {
+       case AARCH64_INSN_VARIANT_32BIT:
+               if (shift & ~(SZ_32 - 1)) {
+                       pr_err("%s: invalid shift encoding %d\n", __func__,
+                              shift);
+                       return AARCH64_BREAK_FAULT;
+               }
+               break;
+       case AARCH64_INSN_VARIANT_64BIT:
+               insn |= AARCH64_INSN_SF_BIT;
+               if (shift & ~(SZ_64 - 1)) {
+                       pr_err("%s: invalid shift encoding %d\n", __func__,
+                              shift);
+                       return AARCH64_BREAK_FAULT;
+               }
+               break;
+       default:
+               pr_err("%s: unknown variant encoding %d\n", __func__, variant);
+               return AARCH64_BREAK_FAULT;
+       }
+
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RD, insn, dst);
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RN, insn, src);
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RM, insn, reg);
+
+       return aarch64_insn_encode_immediate(AARCH64_INSN_IMM_6, insn, shift);
+}
+
+u32 aarch64_insn_gen_data1(enum aarch64_insn_register dst,
+                          enum aarch64_insn_register src,
+                          enum aarch64_insn_variant variant,
+                          enum aarch64_insn_data1_type type)
+{
+       u32 insn;
+
+       switch (type) {
+       case AARCH64_INSN_DATA1_REVERSE_16:
+               insn = aarch64_insn_get_rev16_value();
+               break;
+       case AARCH64_INSN_DATA1_REVERSE_32:
+               insn = aarch64_insn_get_rev32_value();
+               break;
+       case AARCH64_INSN_DATA1_REVERSE_64:
+               if (variant != AARCH64_INSN_VARIANT_64BIT) {
+                       pr_err("%s: invalid variant for reverse64 %d\n",
+                              __func__, variant);
+                       return AARCH64_BREAK_FAULT;
+               }
+               insn = aarch64_insn_get_rev64_value();
+               break;
+       default:
+               pr_err("%s: unknown data1 encoding %d\n", __func__, type);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       switch (variant) {
+       case AARCH64_INSN_VARIANT_32BIT:
+               break;
+       case AARCH64_INSN_VARIANT_64BIT:
+               insn |= AARCH64_INSN_SF_BIT;
+               break;
+       default:
+               pr_err("%s: unknown variant encoding %d\n", __func__, variant);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RD, insn, dst);
+
+       return aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RN, insn, src);
+}
+
+u32 aarch64_insn_gen_data2(enum aarch64_insn_register dst,
+                          enum aarch64_insn_register src,
+                          enum aarch64_insn_register reg,
+                          enum aarch64_insn_variant variant,
+                          enum aarch64_insn_data2_type type)
+{
+       u32 insn;
+
+       switch (type) {
+       case AARCH64_INSN_DATA2_UDIV:
+               insn = aarch64_insn_get_udiv_value();
+               break;
+       case AARCH64_INSN_DATA2_SDIV:
+               insn = aarch64_insn_get_sdiv_value();
+               break;
+       case AARCH64_INSN_DATA2_LSLV:
+               insn = aarch64_insn_get_lslv_value();
+               break;
+       case AARCH64_INSN_DATA2_LSRV:
+               insn = aarch64_insn_get_lsrv_value();
+               break;
+       case AARCH64_INSN_DATA2_ASRV:
+               insn = aarch64_insn_get_asrv_value();
+               break;
+       case AARCH64_INSN_DATA2_RORV:
+               insn = aarch64_insn_get_rorv_value();
+               break;
+       default:
+               pr_err("%s: unknown data2 encoding %d\n", __func__, type);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       switch (variant) {
+       case AARCH64_INSN_VARIANT_32BIT:
+               break;
+       case AARCH64_INSN_VARIANT_64BIT:
+               insn |= AARCH64_INSN_SF_BIT;
+               break;
+       default:
+               pr_err("%s: unknown variant encoding %d\n", __func__, variant);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RD, insn, dst);
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RN, insn, src);
+
+       return aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RM, insn, reg);
+}
+
+u32 aarch64_insn_gen_data3(enum aarch64_insn_register dst,
+                          enum aarch64_insn_register src,
+                          enum aarch64_insn_register reg1,
+                          enum aarch64_insn_register reg2,
+                          enum aarch64_insn_variant variant,
+                          enum aarch64_insn_data3_type type)
+{
+       u32 insn;
+
+       switch (type) {
+       case AARCH64_INSN_DATA3_MADD:
+               insn = aarch64_insn_get_madd_value();
+               break;
+       case AARCH64_INSN_DATA3_MSUB:
+               insn = aarch64_insn_get_msub_value();
+               break;
+       default:
+               pr_err("%s: unknown data3 encoding %d\n", __func__, type);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       switch (variant) {
+       case AARCH64_INSN_VARIANT_32BIT:
+               break;
+       case AARCH64_INSN_VARIANT_64BIT:
+               insn |= AARCH64_INSN_SF_BIT;
+               break;
+       default:
+               pr_err("%s: unknown variant encoding %d\n", __func__, variant);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RD, insn, dst);
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RA, insn, src);
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RN, insn,
+                                           reg1);
+
+       return aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RM, insn,
+                                           reg2);
+}
+
+u32 aarch64_insn_gen_logical_shifted_reg(enum aarch64_insn_register dst,
+                                        enum aarch64_insn_register src,
+                                        enum aarch64_insn_register reg,
+                                        int shift,
+                                        enum aarch64_insn_variant variant,
+                                        enum aarch64_insn_logic_type type)
+{
+       u32 insn;
+
+       switch (type) {
+       case AARCH64_INSN_LOGIC_AND:
+               insn = aarch64_insn_get_and_value();
+               break;
+       case AARCH64_INSN_LOGIC_BIC:
+               insn = aarch64_insn_get_bic_value();
+               break;
+       case AARCH64_INSN_LOGIC_ORR:
+               insn = aarch64_insn_get_orr_value();
+               break;
+       case AARCH64_INSN_LOGIC_ORN:
+               insn = aarch64_insn_get_orn_value();
+               break;
+       case AARCH64_INSN_LOGIC_EOR:
+               insn = aarch64_insn_get_eor_value();
+               break;
+       case AARCH64_INSN_LOGIC_EON:
+               insn = aarch64_insn_get_eon_value();
+               break;
+       case AARCH64_INSN_LOGIC_AND_SETFLAGS:
+               insn = aarch64_insn_get_ands_value();
+               break;
+       case AARCH64_INSN_LOGIC_BIC_SETFLAGS:
+               insn = aarch64_insn_get_bics_value();
+               break;
+       default:
+               pr_err("%s: unknown logical encoding %d\n", __func__, type);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       switch (variant) {
+       case AARCH64_INSN_VARIANT_32BIT:
+               if (shift & ~(SZ_32 - 1)) {
+                       pr_err("%s: invalid shift encoding %d\n", __func__,
+                              shift);
+                       return AARCH64_BREAK_FAULT;
+               }
+               break;
+       case AARCH64_INSN_VARIANT_64BIT:
+               insn |= AARCH64_INSN_SF_BIT;
+               if (shift & ~(SZ_64 - 1)) {
+                       pr_err("%s: invalid shift encoding %d\n", __func__,
+                              shift);
+                       return AARCH64_BREAK_FAULT;
+               }
+               break;
+       default:
+               pr_err("%s: unknown variant encoding %d\n", __func__, variant);
+               return AARCH64_BREAK_FAULT;
+       }
+
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RD, insn, dst);
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RN, insn, src);
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RM, insn, reg);
+
+       return aarch64_insn_encode_immediate(AARCH64_INSN_IMM_6, insn, shift);
+}
+
+/*
+ * MOV (register) is architecturally an alias of ORR (shifted register) where
+ * MOV <*d>, <*m> is equivalent to ORR <*d>, <*ZR>, <*m>
+ */
+u32 aarch64_insn_gen_move_reg(enum aarch64_insn_register dst,
+                             enum aarch64_insn_register src,
+                             enum aarch64_insn_variant variant)
+{
+       return aarch64_insn_gen_logical_shifted_reg(dst, AARCH64_INSN_REG_ZR,
+                                                   src, 0, variant,
+                                                   AARCH64_INSN_LOGIC_ORR);
+}
+
+u32 aarch64_insn_gen_adr(unsigned long pc, unsigned long addr,
+                        enum aarch64_insn_register reg,
+                        enum aarch64_insn_adr_type type)
+{
+       u32 insn;
+       s32 offset;
+
+       switch (type) {
+       case AARCH64_INSN_ADR_TYPE_ADR:
+               insn = aarch64_insn_get_adr_value();
+               offset = addr - pc;
+               break;
+       case AARCH64_INSN_ADR_TYPE_ADRP:
+               insn = aarch64_insn_get_adrp_value();
+               offset = (addr - ALIGN_DOWN(pc, SZ_4K)) >> 12;
+               break;
+       default:
+               pr_err("%s: unknown adr encoding %d\n", __func__, type);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       if (offset < -SZ_1M || offset >= SZ_1M)
+               return AARCH64_BREAK_FAULT;
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RD, insn, reg);
+
+       return aarch64_insn_encode_immediate(AARCH64_INSN_IMM_ADR, insn, 
offset);
+}
+
+/*
+ * Decode the imm field of a branch, and return the byte offset as a
+ * signed value (so it can be used when computing a new branch
+ * target).
+ */
+s32 aarch64_get_branch_offset(u32 insn)
+{
+       s32 imm;
+
+       if (aarch64_insn_is_b(insn) || aarch64_insn_is_bl(insn)) {
+               imm = aarch64_insn_decode_immediate(AARCH64_INSN_IMM_26, insn);
+               return (imm << 6) >> 4;
+       }
+
+       if (aarch64_insn_is_cbz(insn) || aarch64_insn_is_cbnz(insn) ||
+           aarch64_insn_is_bcond(insn)) {
+               imm = aarch64_insn_decode_immediate(AARCH64_INSN_IMM_19, insn);
+               return (imm << 13) >> 11;
+       }
+
+       if (aarch64_insn_is_tbz(insn) || aarch64_insn_is_tbnz(insn)) {
+               imm = aarch64_insn_decode_immediate(AARCH64_INSN_IMM_14, insn);
+               return (imm << 18) >> 16;
+       }
+
+       /* Unhandled instruction */
+       BUG();
+}
+
+/*
+ * Encode the displacement of a branch in the imm field and return the
+ * updated instruction.
+ */
+u32 aarch64_set_branch_offset(u32 insn, s32 offset)
+{
+       if (aarch64_insn_is_b(insn) || aarch64_insn_is_bl(insn))
+               return aarch64_insn_encode_immediate(AARCH64_INSN_IMM_26, insn,
+                                                    offset >> 2);
+
+       if (aarch64_insn_is_cbz(insn) || aarch64_insn_is_cbnz(insn) ||
+           aarch64_insn_is_bcond(insn))
+               return aarch64_insn_encode_immediate(AARCH64_INSN_IMM_19, insn,
+                                                    offset >> 2);
+
+       if (aarch64_insn_is_tbz(insn) || aarch64_insn_is_tbnz(insn))
+               return aarch64_insn_encode_immediate(AARCH64_INSN_IMM_14, insn,
+                                                    offset >> 2);
+
+       /* Unhandled instruction */
+       BUG();
+}
+
+s32 aarch64_insn_adrp_get_offset(u32 insn)
+{
+       BUG_ON(!aarch64_insn_is_adrp(insn));
+       return aarch64_insn_decode_immediate(AARCH64_INSN_IMM_ADR, insn) << 12;
+}
+
+u32 aarch64_insn_adrp_set_offset(u32 insn, s32 offset)
+{
+       BUG_ON(!aarch64_insn_is_adrp(insn));
+       return aarch64_insn_encode_immediate(AARCH64_INSN_IMM_ADR, insn,
+                                               offset >> 12);
+}
+
+/*
+ * Extract the Op/CR data from a msr/mrs instruction.
+ */
+u32 aarch64_insn_extract_system_reg(u32 insn)
+{
+       return (insn & 0x1FFFE0) >> 5;
+}
+
+static bool range_of_ones(u64 val)
+{
+       /* Doesn't handle full ones or full zeroes */
+       u64 sval = val >> __ffs64(val);
+
+       /* One of Sean Eron Anderson's bithack tricks */
+       return ((sval + 1) & (sval)) == 0;
+}
+
+static u32 aarch64_encode_immediate(u64 imm,
+                                   enum aarch64_insn_variant variant,
+                                   u32 insn)
+{
+       unsigned int immr, imms, n, ones, ror, esz, tmp;
+       u64 mask;
+
+       switch (variant) {
+       case AARCH64_INSN_VARIANT_32BIT:
+               esz = 32;
+               break;
+       case AARCH64_INSN_VARIANT_64BIT:
+               insn |= AARCH64_INSN_SF_BIT;
+               esz = 64;
+               break;
+       default:
+               pr_err("%s: unknown variant encoding %d\n", __func__, variant);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       mask = GENMASK(esz - 1, 0);
+
+       /* Can't encode full zeroes, full ones, or value wider than the mask */
+       if (!imm || imm == mask || imm & ~mask)
+               return AARCH64_BREAK_FAULT;
+
+       /*
+        * Inverse of Replicate(). Try to spot a repeating pattern
+        * with a pow2 stride.
+        */
+       for (tmp = esz / 2; tmp >= 2; tmp /= 2) {
+               u64 emask = BIT(tmp) - 1;
+
+               if ((imm & emask) != ((imm >> tmp) & emask))
+                       break;
+
+               esz = tmp;
+               mask = emask;
+       }
+
+       /* N is only set if we're encoding a 64bit value */
+       n = esz == 64;
+
+       /* Trim imm to the element size */
+       imm &= mask;
+
+       /* That's how many ones we need to encode */
+       ones = hweight64(imm);
+
+       /*
+        * imms is set to (ones - 1), prefixed with a string of ones
+        * and a zero if they fit. Cap it to 6 bits.
+        */
+       imms  = ones - 1;
+       imms |= 0xf << ffs(esz);
+       imms &= BIT(6) - 1;
+
+       /* Compute the rotation */
+       if (range_of_ones(imm)) {
+               /*
+                * Pattern: 0..01..10..0
+                *
+                * Compute how many rotate we need to align it right
+                */
+               ror = __ffs64(imm);
+       } else {
+               /*
+                * Pattern: 0..01..10..01..1
+                *
+                * Fill the unused top bits with ones, and check if
+                * the result is a valid immediate (all ones with a
+                * contiguous ranges of zeroes).
+                */
+               imm |= ~mask;
+               if (!range_of_ones(~imm))
+                       return AARCH64_BREAK_FAULT;
+
+               /*
+                * Compute the rotation to get a continuous set of
+                * ones, with the first bit set at position 0
+                */
+               ror = fls(~imm);
+       }
+
+       /*
+        * immr is the number of bits we need to rotate back to the
+        * original set of ones. Note that this is relative to the
+        * element size...
+        */
+       immr = (esz - ror) % esz;
+
+       insn = aarch64_insn_encode_immediate(AARCH64_INSN_IMM_N, insn, n);
+       insn = aarch64_insn_encode_immediate(AARCH64_INSN_IMM_R, insn, immr);
+       return aarch64_insn_encode_immediate(AARCH64_INSN_IMM_S, insn, imms);
+}
+
+u32 aarch64_insn_gen_logical_immediate(enum aarch64_insn_logic_type type,
+                                      enum aarch64_insn_variant variant,
+                                      enum aarch64_insn_register Rn,
+                                      enum aarch64_insn_register Rd,
+                                      u64 imm)
+{
+       u32 insn;
+
+       switch (type) {
+       case AARCH64_INSN_LOGIC_AND:
+               insn = aarch64_insn_get_and_imm_value();
+               break;
+       case AARCH64_INSN_LOGIC_ORR:
+               insn = aarch64_insn_get_orr_imm_value();
+               break;
+       case AARCH64_INSN_LOGIC_EOR:
+               insn = aarch64_insn_get_eor_imm_value();
+               break;
+       case AARCH64_INSN_LOGIC_AND_SETFLAGS:
+               insn = aarch64_insn_get_ands_imm_value();
+               break;
+       default:
+               pr_err("%s: unknown logical encoding %d\n", __func__, type);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RD, insn, Rd);
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RN, insn, Rn);
+       return aarch64_encode_immediate(imm, variant, insn);
+}
+
+u32 aarch64_insn_gen_extr(enum aarch64_insn_variant variant,
+                         enum aarch64_insn_register Rm,
+                         enum aarch64_insn_register Rn,
+                         enum aarch64_insn_register Rd,
+                         u8 lsb)
+{
+       u32 insn;
+
+       insn = aarch64_insn_get_extr_value();
+
+       switch (variant) {
+       case AARCH64_INSN_VARIANT_32BIT:
+               if (lsb > 31)
+                       return AARCH64_BREAK_FAULT;
+               break;
+       case AARCH64_INSN_VARIANT_64BIT:
+               if (lsb > 63)
+                       return AARCH64_BREAK_FAULT;
+               insn |= AARCH64_INSN_SF_BIT;
+               insn = aarch64_insn_encode_immediate(AARCH64_INSN_IMM_N, insn, 
1);
+               break;
+       default:
+               pr_err("%s: unknown variant encoding %d\n", __func__, variant);
+               return AARCH64_BREAK_FAULT;
+       }
+
+       insn = aarch64_insn_encode_immediate(AARCH64_INSN_IMM_S, insn, lsb);
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RD, insn, Rd);
+       insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RN, insn, Rn);
+       return aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RM, insn, Rm);
+}
-- 
2.25.4

Reply via email to