This patch adds C intrinsics for Scalar Crypto Extension.
gcc/ChangeLog: * config.gcc: Add riscv_crypto.h. * config/riscv/riscv_crypto.h: New file. gcc/testsuite/ChangeLog: * gcc.target/riscv/scalar_crypto_intrinsic-1.c: New test. * gcc.target/riscv/scalar_crypto_intrinsic-2.c: New test. --- gcc/config.gcc | 2 +- gcc/config/riscv/riscv_crypto.h | 280 ++++++++++++++++++ .../riscv/scalar_crypto_intrinsic-1.c | 115 +++++++ .../riscv/scalar_crypto_intrinsic-2.c | 122 ++++++++ 4 files changed, 518 insertions(+), 1 deletion(-) create mode 100644 gcc/config/riscv/riscv_crypto.h create mode 100644 gcc/testsuite/gcc.target/riscv/scalar_crypto_intrinsic-1.c create mode 100644 gcc/testsuite/gcc.target/riscv/scalar_crypto_intrinsic-2.c diff --git a/gcc/config.gcc b/gcc/config.gcc index b88591b6fd8..d67fe8b6a6f 100644 --- a/gcc/config.gcc +++ b/gcc/config.gcc @@ -548,7 +548,7 @@ riscv*) extra_objs="${extra_objs} riscv-vector-builtins.o riscv-vector-builtins-shapes.o riscv-vector-builtins-bases.o" extra_objs="${extra_objs} thead.o riscv-target-attr.o" d_target_objs="riscv-d.o" - extra_headers="riscv_vector.h" + extra_headers="riscv_vector.h riscv_crypto.h" target_gtfiles="$target_gtfiles \$(srcdir)/config/riscv/riscv-vector-builtins.cc" target_gtfiles="$target_gtfiles \$(srcdir)/config/riscv/riscv-vector-builtins.h" ;; diff --git a/gcc/config/riscv/riscv_crypto.h b/gcc/config/riscv/riscv_crypto.h new file mode 100644 index 00000000000..9335ffda20d --- /dev/null +++ b/gcc/config/riscv/riscv_crypto.h @@ -0,0 +1,280 @@ +/* RISC-V 'Scalar Crypto' Extension intrinsics include file. + Copyright (C) 2023 Free Software Foundation, Inc. + + This file is part of GCC. + + GCC is free software; you can redistribute it and/or modify it + under the terms of the GNU General Public License as published + by the Free Software Foundation; either version 3, or (at your + option) any later version. + + GCC is distributed in the hope that it will be useful, but WITHOUT + ANY WARRANTY; without even the implied warranty of MERCHANTABILITY + or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public + License for more details. + + Under Section 7 of GPL version 3, you are granted additional + permissions described in the GCC Runtime Library Exception, version + 3.1, as published by the Free Software Foundation. + + You should have received a copy of the GNU General Public License and + a copy of the GCC Runtime Library Exception along with this program; + see the files COPYING3 and COPYING.RUNTIME respectively. If not, see + <http://www.gnu.org/licenses/>. */ + +#ifndef __RISCV_SCALAR_CRYPTO_H +#define __RISCV_SCALAR_CRYPTO_H + +#include <stdint.h> + +#ifdef __cplusplus +extern "C" { +#endif + +#if defined(__riscv_zknd) + +#if __riscv_xlen == 32 + +extern __inline uint32_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_aes32dsi (uint32_t rs1, uint32_t rs2, const int bs) +{ + return __builtin_riscv_aes32dsi (rs1,rs2,bs); +} + +extern __inline uint32_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_aes32dsmi (uint32_t rs1, uint32_t rs2, const int bs) +{ + return __builtin_riscv_aes32dsmi (rs1,rs2,bs); +} +#endif + +#if __riscv_xlen == 64 + +extern __inline uint64_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_aes64ds (uint64_t rs1, uint64_t rs2) +{ + return __builtin_riscv_aes64ds (rs1,rs2); +} + +extern __inline uint64_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_aes64dsm (uint64_t rs1, uint64_t rs2) +{ + return __builtin_riscv_aes64dsm (rs1,rs2); +} + +extern __inline uint64_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_aes64im (uint64_t rs1) +{ + return __builtin_riscv_aes64im (rs1); +} +#endif +#endif // ZKND + +#if (defined(__riscv_zknd) || defined(__riscv_zkne)) && (__riscv_xlen == 64) + +extern __inline uint64_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_aes64ks1i(uint64_t rs1, const int rnum) +{ + return __builtin_riscv_aes64ks1i(rs1,rnum); +} + +extern __inline uint64_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_aes64ks2(uint64_t rs1, uint64_t rs2) +{ + return __builtin_riscv_aes64ks2(rs1,rs2); +} + +#endif // ZKND || ZKNE + +#if defined(__riscv_zkne) + +#if __riscv_xlen == 32 + +extern __inline uint32_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_aes32esi (uint32_t rs1, uint32_t rs2, const int bs) +{ + return __builtin_riscv_aes32esi (rs1,rs2,bs); +} + +extern __inline uint32_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_aes32esmi (uint32_t rs1, uint32_t rs2, const int bs) +{ + return __builtin_riscv_aes32esmi (rs1,rs2,bs); +} +#endif + +#if __riscv_xlen == 64 + +extern __inline uint64_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_aes64es (uint64_t rs1,uint64_t rs2) +{ + return __builtin_riscv_aes64es (rs1,rs2); +} + +extern __inline uint64_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_aes64esm (uint64_t rs1,uint64_t rs2) +{ + return __builtin_riscv_aes64esm (rs1,rs2); +} +#endif +#endif // ZKNE + +#if defined(__riscv_zknh) + +extern __inline uint32_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_sha256sig0(uint32_t rs1) +{ + return __builtin_riscv_sha256sig0 (rs1); +} + +extern __inline uint32_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_sha256sig1(uint32_t rs1) +{ + return __builtin_riscv_sha256sig1 (rs1); +} + +extern __inline uint32_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_sha256sum0 (uint32_t rs1) +{ + return __builtin_riscv_sha256sum0 (rs1); +} + +extern __inline uint32_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_sha256sum1 (uint32_t rs1) +{ + return __builtin_riscv_sha256sum1 (rs1); +} + +#if __riscv_xlen == 32 + +extern __inline uint32_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_sha512sig0h(uint32_t rs1, uint32_t rs2) +{ + return __builtin_riscv_sha512sig0h (rs1,rs2); +} + +extern __inline uint32_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_sha512sig0l(uint32_t rs1, uint32_t rs2) +{ + return __builtin_riscv_sha512sig0l (rs1,rs2); +} + +extern __inline uint32_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_sha512sig1h(uint32_t rs1, uint32_t rs2) +{ + return __builtin_riscv_sha512sig1h (rs1,rs2); +} + +extern __inline uint32_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_sha512sig1l(uint32_t rs1, uint32_t rs2) +{ + return __builtin_riscv_sha512sig1l (rs1,rs2); +} + +extern __inline uint32_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_sha512sum0r(uint32_t rs1, uint32_t rs2) +{ + return __builtin_riscv_sha512sum0r (rs1,rs2); +} + +extern __inline uint32_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_sha512sum1r(uint32_t rs1, uint32_t rs2) +{ + return __builtin_riscv_sha512sum1r (rs1,rs2); +} + +#endif + +#if __riscv_xlen == 64 + +extern __inline uint64_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_sha512sig0 (uint64_t rs1) +{ + return __builtin_riscv_sha512sig0 (rs1); +} + +extern __inline uint64_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_sha512sig1 (uint64_t rs1) +{ + return __builtin_riscv_sha512sig1 (rs1); +} + +extern __inline uint64_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_sha512sum0 (uint64_t rs1) +{ + return __builtin_riscv_sha512sum0 (rs1); +} + +extern __inline uint64_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_sha512sum1 (uint64_t rs1) +{ + return __builtin_riscv_sha512sum1 (rs1); +} +#endif +#endif // ZKNH + +#if defined(__riscv_zksh) + +extern __inline uint32_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_sm3p0(uint32_t rs1) +{ + return __builtin_riscv_sm3p0 (rs1); +} + +extern __inline uint32_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_sm3p1(uint32_t rs1) +{ + return __builtin_riscv_sm3p1 (rs1); +} + +#endif // ZKSH + +#if defined(__riscv_zksed) + +extern __inline uint32_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_sm4ed (uint32_t rs1, uint32_t rs2, const int bs) +{ + return __builtin_riscv_sm4ed (rs1,rs2,bs); +} + +extern __inline uint32_t +__attribute__((__gnu_inline__, __always_inline__, __artificial__)) +__riscv_sm4ks (uint32_t rs1, uint32_t rs2, const int bs) +{ + return __builtin_riscv_sm4ks (rs1,rs2,bs); +} + +#endif // ZKSED + +#if defined(__cplusplus) +} +#endif // __cplusplus +#endif // __RISCV_SCALAR_CRYPTO_H diff --git a/gcc/testsuite/gcc.target/riscv/scalar_crypto_intrinsic-1.c b/gcc/testsuite/gcc.target/riscv/scalar_crypto_intrinsic-1.c new file mode 100644 index 00000000000..d9d6261c0fd --- /dev/null +++ b/gcc/testsuite/gcc.target/riscv/scalar_crypto_intrinsic-1.c @@ -0,0 +1,115 @@ +/* { dg-do compile } */ +/* { dg-options "-O2 -march=rv32gc_zknd_zkne_zknh_zksed_zksh -mabi=ilp32d" } */ +/* { dg-skip-if "" { *-*-* } { "-g" "-flto"} } */ + +#include "riscv_crypto.h" + +uint32_t foo1(uint32_t rs1, uint32_t rs2, const int bs) +{ + return __riscv_aes32dsi(rs1,rs2,bs); +} + +uint32_t foo2(uint32_t rs1, uint32_t rs2, const int bs) +{ + return __riscv_aes32dsmi(rs1,rs2,bs); +} + +uint32_t foo3(uint32_t rs1, uint32_t rs2, const int bs) +{ + return __riscv_aes32esi(rs1,rs2,bs); +} + +uint32_t foo4(uint32_t rs1, uint32_t rs2, const int bs) +{ + return __riscv_aes32esmi(rs1,rs2,bs); +} + +uint32_t foo5(uint32_t rs1) +{ + return __riscv_sha256sig0(rs1); +} + +uint32_t foo6(uint32_t rs1) +{ + return __riscv_sha256sig1(rs1); +} + +uint32_t foo7(uint32_t rs1) +{ + return __riscv_sha256sum0(rs1); +} + +uint32_t foo8(uint32_t rs1) +{ + return __riscv_sha256sum1(rs1); +} + +uint32_t foo9(uint32_t rs1, uint32_t rs2) +{ + return __riscv_sha512sig0h(rs1,rs2); +} + +uint32_t foo10(uint32_t rs1, uint32_t rs2) +{ + return __riscv_sha512sig0l(rs1,rs2); +} + +uint32_t foo11(uint32_t rs1, uint32_t rs2) +{ + return __riscv_sha512sig1h(rs1,rs2); +} + +uint32_t foo12(uint32_t rs1, uint32_t rs2) +{ + return __riscv_sha512sig1l(rs1,rs2); +} + +uint32_t foo13(uint32_t rs1, uint32_t rs2) +{ + return __riscv_sha512sum0r(rs1,rs2); +} + +uint32_t foo14(uint32_t rs1, uint32_t rs2) +{ + return __riscv_sha512sum1r(rs1,rs2); +} + +uint32_t foo15(uint32_t rs1) +{ + return __riscv_sm3p0(rs1); +} + +uint32_t foo16(uint32_t rs1) +{ + return __riscv_sm3p1(rs1); +} + +uint32_t foo17(uint32_t rs1, uint32_t rs2, const int bs) +{ + return __riscv_sm4ed(rs1,rs2,bs); +} + +uint32_t foo18(uint32_t rs1, uint32_t rs2, const int bs) +{ + return __riscv_sm4ks(rs1,rs2,bs); +} + +/* { dg-final { scan-assembler-times "aes32dsi" 1 } } */ +/* { dg-final { scan-assembler-times "aes32dsmi" 1 } } */ +/* { dg-final { scan-assembler-times "aes32esi" 1 } } */ +/* { dg-final { scan-assembler-times "aes32esmi" 1 } } */ +/* { dg-final { scan-assembler-times "sha256sig0" 1 } } */ +/* { dg-final { scan-assembler-times "sha256sig1" 1 } } */ +/* { dg-final { scan-assembler-times "sha256sum0" 1 } } */ +/* { dg-final { scan-assembler-times "sha256sum1" 1 } } */ +/* { dg-final { scan-assembler-times "sha512sig0h" 1 } } */ +/* { dg-final { scan-assembler-times "sha512sig0l" 1 } } */ +/* { dg-final { scan-assembler-times "sha512sig1h" 1 } } */ +/* { dg-final { scan-assembler-times "sha512sig1l" 1 } } */ +/* { dg-final { scan-assembler-times "sha512sum0r" 1 } } */ +/* { dg-final { scan-assembler-times "sha512sum1r" 1 } } */ +/* { dg-final { scan-assembler-times "sm3p0" 1 } } */ +/* { dg-final { scan-assembler-times "sm3p1" 1 } } */ +/* { dg-final { scan-assembler-times "sm4ks" 1 } } */ +/* { dg-final { scan-assembler-times "sm4ed" 1 } } */ + diff --git a/gcc/testsuite/gcc.target/riscv/scalar_crypto_intrinsic-2.c b/gcc/testsuite/gcc.target/riscv/scalar_crypto_intrinsic-2.c new file mode 100644 index 00000000000..3b7815a29e8 --- /dev/null +++ b/gcc/testsuite/gcc.target/riscv/scalar_crypto_intrinsic-2.c @@ -0,0 +1,122 @@ +/* { dg-do compile } */ +/* { dg-options "-O2 -march=rv64gc_zknd_zkne_zknh_zksed_zksh -mabi=lp64" } */ +/* { dg-skip-if "" { *-*-* } { "-g" "-flto"} } */ + +#include "riscv_crypto.h" + +uint64_t foo1(uint64_t rs1, uint64_t rs2) +{ + return __riscv_aes64ds(rs1,rs2); +} + +uint64_t foo2(uint64_t rs1, uint64_t rs2) +{ + return __riscv_aes64dsm(rs1,rs2); +} + +uint64_t foo3(uint64_t rs1) +{ + return __riscv_aes64im(rs1); +} + +uint64_t foo4(uint64_t rs1, const int rnum) +{ + return __riscv_aes64ks1i(rs1,rnum); +} + +uint64_t foo5(uint64_t rs1, uint64_t rs2) +{ + return __riscv_aes64ks2(rs1,rs2); +} + +uint64_t foo6(uint64_t rs1, uint64_t rs2) +{ + return __riscv_aes64es(rs1,rs2); +} + +uint64_t foo7(uint64_t rs1, uint64_t rs2) +{ + return __riscv_aes64esm(rs1,rs2); +} + +uint64_t foo8(uint64_t rs1) +{ + return __riscv_sha512sig0(rs1); +} + +uint64_t foo9(uint64_t rs1) +{ + return __riscv_sha512sig1(rs1); +} + +uint64_t foo10(uint64_t rs1) +{ + return __riscv_sha512sum0(rs1); +} + +uint64_t foo11(uint64_t rs1) +{ + return __riscv_sha512sum1(rs1); +} + +uint32_t foo12(uint32_t rs1) +{ + return __riscv_sha256sig0(rs1); +} + +uint32_t foo13(uint32_t rs1) +{ + return __riscv_sha256sig1(rs1); +} + +uint32_t foo14(uint32_t rs1) +{ + return __riscv_sha256sum0(rs1); +} + +uint32_t foo15(uint32_t rs1) +{ + return __riscv_sha256sum1(rs1); +} + +uint32_t foo16(uint32_t rs1) +{ + return __riscv_sm3p0(rs1); +} + +uint32_t foo17(uint32_t rs1) +{ + return __riscv_sm3p1(rs1); +} + +uint32_t foo18(uint32_t rs1, uint32_t rs2, const int bs) +{ + return __riscv_sm4ed(rs1,rs2,bs); +} + +uint32_t foo19(uint32_t rs1, uint32_t rs2, const int bs) +{ + return __riscv_sm4ks(rs1,rs2,bs); +} + +/* { dg-final { scan-assembler-times "aes64ds\t" 1 } } */ +/* { dg-final { scan-assembler-times "aes64dsm" 1 } } */ +/* { dg-final { scan-assembler-times "aes64ks1i" 1 } } */ +/* { dg-final { scan-assembler-times "aes64ks2" 1 } } */ +/* { dg-final { scan-assembler-times "aes64im" 1 } } */ +/* { dg-final { scan-assembler-times "aes64es\t" 1 } } */ +/* { dg-final { scan-assembler-times "aes64esm" 1 } } */ +/* { dg-final { scan-assembler-times "aes64ks1i" 1 } } */ +/* { dg-final { scan-assembler-times "aes64ks2" 1 } } */ +/* { dg-final { scan-assembler-times "sha512sig0" 1 } } */ +/* { dg-final { scan-assembler-times "sha512sig1" 1 } } */ +/* { dg-final { scan-assembler-times "sha512sum0" 1 } } */ +/* { dg-final { scan-assembler-times "sha512sum1" 1 } } */ +/* { dg-final { scan-assembler-times "sha256sig0" 1 } } */ +/* { dg-final { scan-assembler-times "sha256sig1" 1 } } */ +/* { dg-final { scan-assembler-times "sha256sum0" 1 } } */ +/* { dg-final { scan-assembler-times "sha256sum1" 1 } } */ +/* { dg-final { scan-assembler-times "sm3p0" 1 } } */ +/* { dg-final { scan-assembler-times "sm3p1" 1 } } */ +/* { dg-final { scan-assembler-times "sm4ks" 1 } } */ +/* { dg-final { scan-assembler-times "sm4ed" 1 } } */ -- 2.34.1