Mihail Ionescu <mihail.ione...@foss.arm.com> writes:
> Hi,
>
> On 02/17/2020 05:53 PM, Mihail Ionescu wrote:
>> Hi,
>> 
>> This patch adds support for the bf16 duplicate and reinterpret intrinsics.
>> ACLE documents are at https://developer.arm.com/docs/101028/latest
>> ISA documents are at https://developer.arm.com/docs/ddi0596/latest
>> 
>> Regression tested on aarch64-none-linux-gnu.
>> 
>> 
>> Is it ok for trunk?
>> 
>> 
>> gcc/ChangeLog:
>> 
>> 2020-02-17  Mihail Ionescu  <mihail.ione...@arm.com>
>> 
>>      * config/aarch64/iterators.md (VDQF_F16) Add V4BF and V8BF.
>>      (VALL_F16): Likewise.
>>      (VALLDI_F16): Likewise.
>>      (Vtype): Likewise.
>>      (Vetype): Likewise.
>>      (vswap_width_name): Likewise.
>>      (VSWAP_WIDTH): Likewise.
>>      (Vel): Likewise.
>>      (VEL): Likewise.
>>      (q): Likewise.
>>      * config/aarch64/aarch64-simd.md
>>      (vec_init<mode><Vel>): Add vector init pattern for bf16.
>>      (aarch64_simd_dup): Change pattern iterator to include bf16.
>>      (aarch64_dup_lane): Likewise.
>>      (aarch64_get_lane): Likewise.
>>      (vec_extract): Likewise.
>>      * config/aarch64/arm_bf16.h
>>      (vset_lane_bf16, vsetq_lane_bf16): New.
>>      (vget_lane_bf16, vgetq_lane_bf16): New.
>>      (vcreate_bf16): New.
>>      (vdup_n_bf16, vdupq_n_bf16): New.
>>      (vdup_lane_bf16, vdup_laneq_bf16): New.
>>      (vdupq_lane_bf16, vdupq_laneq_bf16): New.
>>      (vduph_lane_bf16, vduph_laneq_bf16): New.
>>      (vreinterpret_bf16_u8, vreinterpretq_bf16_u8): New.
>>      (vreinterpret_bf16_u16, vreinterpretq_bf16_u16): New.
>>      (vreinterpret_bf16_u32, vreinterpretq_bf16_u32): New.
>>      (vreinterpret_bf16_u64, vreinterpretq_bf16_u64): New.
>>      (vreinterpret_bf16_s8, vreinterpretq_bf16_s8): New.
>>      (vreinterpret_bf16_s16, vreinterpretq_bf16_s16): New.
>>      (vreinterpret_bf16_s32, vreinterpretq_bf16_s32): New.
>>      (vreinterpret_bf16_s64, vreinterpretq_bf16_s64): New.
>>      (vreinterpret_bf16_p8, vreinterpretq_bf16_p8): New.
>>      (vreinterpret_bf16_p16, vreinterpretq_bf16_p16): New.
>>      (vreinterpret_bf16_p64, vreinterpretq_bf16_p64): New
>>      (vreinterpret_bf16_f16, vreinterpretq_bf16_f16): New
>>      (vreinterpret_bf16_f32, vreinterpretq_bf16_f32): New.
>>      (vreinterpret_bf16_f64, vreinterpretq_bf16_f64): New.
>>      (vreinterpretq_bf16_p128): New.
>>      (vreinterpret_s8_bf16, vreinterpretq_s8_bf16): New.
>>      (vreinterpret_s16_bf16, vreinterpretq_s16_bf16): New.
>>      (vreinterpret_s32_bf16, vreinterpretq_s32_bf16): New.
>>      (vreinterpret_s64_bf16, vreinterpretq_s64_bf16): New.
>>      (vreinterpret_u8_bf16, vreinterpretq_u8_bf16): New.
>>      (vreinterpret_u16_bf16, vreinterpretq_u16_bf16): New.
>>      (vreinterpret_u32_bf16, vreinterpretq_u32_bf16): New.
>>      (vreinterpret_u64_bf16, vreinterpretq_u64_bf16): New.
>>      (vreinterpret_p8_bf16, vreinterpretq_p8_bf16): New.
>>      (vreinterpret_p16_bf16, vreinterpretq_p16_bf16): New.
>>      (vreinterpret_p64_bf16, vreinterpretq_p64_bf16): New.
>>      (vreinterpret_f32_bf16, vreinterpretq_f32_bf16): New.
>>      (vreinterpret_f64_bf16,vreinterpretq_f64_bf16): New.
>>      (vreinterpret_f16_bf16,vreinterpretq_f16_bf16): New.
>>      (vreinterpretq_p128_bf16): New.
>> 
>> 
>> gcc/testsuite/ChangeLog:
>> 
>> 2020-02-17  Mihail Ionescu  <mihail.ione...@arm.com>
>> 
>>      * gcc.target/aarch64/advsimd-intrinsics/bf16_dup.c: New test.
>>      * gcc.target/aarch64/advsimd-intrinsics/bf16_reinterpret.c: New test.
>> 
>> Regards,
>> Mihail
>> 
>> 
>> ###############     Attachment also inlined for ease of reply    
>> ###############
>> 
>> 
>> diff --git a/gcc/config/aarch64/arm_neon.h b/gcc/config/aarch64/arm_neon.h
>> index 
>> 7f05c3f9eca844b0e7b824a191223a4906c825b1..3cc3ace83fabf25d8e2e6e70382d335afd974290
>>  100644
>> --- a/gcc/config/aarch64/arm_neon.h
>> +++ b/gcc/config/aarch64/arm_neon.h
>> @@ -34614,6 +34614,507 @@ vrnd64xq_f64 (float64x2_t __a)
>>   #pragma GCC push_options
>>   #pragma GCC target ("arch=armv8.2-a+bf16")
>>   
>> +__extension__ extern __inline bfloat16x4_t
>> +__attribute__ ((__always_inline__, __gnu_inline__, __artificial__))
>> +vset_lane_bf16 (bfloat16_t __elem, bfloat16x4_t __vec, const int __index)
>> +{
>> +  return __aarch64_vset_lane_any (__elem, __vec, __index);
>> +}
>> +
>> +__extension__ extern __inline bfloat16x8_t
>> +__attribute__ ((__always_inline__, __gnu_inline__, __artificial__))
>> +vsetq_lane_bf16 (bfloat16_t __elem, bfloat16x8_t __vec, const int __index)
>> +{
>> +  return __aarch64_vset_lane_any (__elem, __vec, __index);
>> +}
>> +
>> +__extension__ extern __inline bfloat16_t
>> +__attribute__ ((__always_inline__, __gnu_inline__, __artificial__))
>> +vget_lane_bf16 (bfloat16x4_t __a, const int __b)
>> +{
>> +  return __aarch64_vget_lane_any (__a, __b);
>> +}
>> +
>> +__extension__ extern __inline bfloat16_t
>> +__attribute__ ((__always_inline__, __gnu_inline__, __artificial__))
>> +vgetq_lane_bf16 (bfloat16x8_t __a, const int __b)
>> +{
>> +  return __aarch64_vget_lane_any (__a, __b);
>> +}
>> +
>> +__extension__ extern __inline bfloat16x4_t
>> +__attribute__ ((__always_inline__, __gnu_inline__, __artificial__))
>> +vcreate_bf16 (uint64_t __a)
>> +{
>> +  return (bfloat16x4_t) __a;
>> +}
>> +
>> +/* vdup */
>> +
>> +__extension__ extern __inline bfloat16x4_t
>> +__attribute__ ((__always_inline__, __gnu_inline__, __artificial__))
>> +vdup_n_bf16 (bfloat16_t __a)
>> +{
>> +  return (bfloat16x4_t) {__a, __a, __a, __a};
>> +}
>> +
>> +__extension__ extern __inline bfloat16x8_t
>> +__attribute__ ((__always_inline__, __gnu_inline__, __artificial__))
>> +vdupq_n_bf16 (bfloat16_t __a)
>> +{
>> +  return (bfloat16x8_t) {__a, __a, __a, __a, __a, __a, __a, __a};
>> +}
>> +
>> +__extension__ extern __inline bfloat16x4_t
>> +__attribute__ ((__always_inline__, __gnu_inline__, __artificial__))
>> +vdup_lane_bf16 (bfloat16x4_t __a, const int __b)
>> +{
>> +  return vdup_n_bf16 (__aarch64_vget_lane_any (__a, __b));
>> +}
>> +
>> +__extension__ extern __inline bfloat16x4_t
>> +__attribute__ ((__always_inline__, __gnu_inline__, __artificial__))
>> +vdup_laneq_bf16 (bfloat16x8_t __a, const int __b)
>> +{
>> +  return vdup_n_bf16 (__aarch64_vget_lane_any (__a, __b));
>> +}
>> +
>> +__extension__ extern __inline bfloat16x8_t
>> +__attribute__ ((__always_inline__, __gnu_inline__, __artificial__))
>> +vdupq_lane_bf16 (bfloat16x4_t __a, const int __b)
>> +{
>> +  return vdupq_n_bf16 (__aarch64_vget_lane_any (__a, __b));
>> +}
>> +
>> +__extension__ extern __inline bfloat16x8_t
>> +__attribute__ ((__always_inline__, __gnu_inline__, __artificial__))
>> +vdupq_laneq_bf16 (bfloat16x8_t __a, const int __b)
>> +{
>> +  return vdupq_n_bf16 (__aarch64_vget_lane_any (__a, __b));
>> +}
>> +
>> +__extension__ extern __inline bfloat16_t
>> +__attribute__ ((__always_inline__, __gnu_inline__, __artificial__))
>> +vduph_lane_bf16 (bfloat16x4_t __a, const int __b)
>> +{
>> +  return __aarch64_vget_lane_any (__a, __b);
>> +}
>> +
>> +__extension__ extern __inline bfloat16_t
>> +__attribute__ ((__always_inline__, __gnu_inline__, __artificial__))
>> +vduph_laneq_bf16 (bfloat16x8_t __a, const int __b)
>> +{
>> +  return __aarch64_vget_lane_any (__a, __b);
>> +}
>> +
>> +/* vreinterpret */
>> +
>> +__extension__ extern __inline bfloat16x4_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_bf16_u8 (uint8x8_t __a)
>> +{
>> +  return (bfloat16x4_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x4_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_bf16_u16 (uint16x4_t __a)
>> +{
>> +  return (bfloat16x4_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x4_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_bf16_u32 (uint32x2_t __a)
>> +{
>> +  return (bfloat16x4_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x4_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_bf16_u64 (uint64x1_t __a)
>> +{
>> +  return (bfloat16x4_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x4_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_bf16_s8 (int8x8_t __a)
>> +{
>> +  return (bfloat16x4_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x4_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_bf16_s16 (int16x4_t __a)
>> +{
>> +  return (bfloat16x4_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x4_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_bf16_s32 (int32x2_t __a)
>> +{
>> +  return (bfloat16x4_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x4_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_bf16_s64 (int64x1_t __a)
>> +{
>> +  return (bfloat16x4_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x4_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_bf16_p8 (poly8x8_t __a)
>> +{
>> +  return (bfloat16x4_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x4_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_bf16_p16 (poly16x4_t __a)
>> +{
>> +  return (bfloat16x4_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x4_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_bf16_p64 (poly64x1_t __a)
>> +{
>> +  return (bfloat16x4_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x4_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_bf16_f16 (float16x4_t __a)
>> +{
>> +  return (bfloat16x4_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x4_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_bf16_f32 (float32x2_t __a)
>> +{
>> +  return (bfloat16x4_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x4_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_bf16_f64 (float64x1_t __a)
>> +{
>> +  return (bfloat16x4_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x8_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_bf16_u8 (uint8x16_t __a)
>> +{
>> +  return (bfloat16x8_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x8_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_bf16_u16 (uint16x8_t __a)
>> +{
>> +  return (bfloat16x8_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x8_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_bf16_u32 (uint32x4_t __a)
>> +{
>> +  return (bfloat16x8_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x8_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_bf16_u64 (uint64x2_t __a)
>> +{
>> +  return (bfloat16x8_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x8_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_bf16_s8 (int8x16_t __a)
>> +{
>> +  return (bfloat16x8_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x8_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_bf16_s16 (int16x8_t __a)
>> +{
>> +  return (bfloat16x8_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x8_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_bf16_s32 (int32x4_t __a)
>> +{
>> +  return (bfloat16x8_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x8_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_bf16_s64 (int64x2_t __a)
>> +{
>> +  return (bfloat16x8_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x8_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_bf16_p8 (poly8x16_t __a)
>> +{
>> +  return (bfloat16x8_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x8_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_bf16_p16 (poly16x8_t __a)
>> +{
>> +  return (bfloat16x8_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x8_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_bf16_p64 (poly64x2_t __a)
>> +{
>> +  return (bfloat16x8_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x8_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_bf16_p128 (poly128_t __a)
>> +{
>> +  return (bfloat16x8_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x8_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_bf16_f16 (float16x8_t __a)
>> +{
>> +  return (bfloat16x8_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x8_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_bf16_f32 (float32x4_t __a)
>> +{
>> +  return (bfloat16x8_t)__a;
>> +}
>> +
>> +__extension__ extern __inline bfloat16x8_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_bf16_f64 (float64x2_t __a)
>> +{
>> +  return (bfloat16x8_t)__a;
>> +}
>> +
>> +__extension__ extern __inline int8x8_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_s8_bf16 (bfloat16x4_t __a)
>> +{
>> +  return (int8x8_t)__a;
>> +}
>> +
>> +__extension__ extern __inline int16x4_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_s16_bf16 (bfloat16x4_t __a)
>> +{
>> +  return (int16x4_t)__a;
>> +}
>> +
>> +__extension__ extern __inline int32x2_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_s32_bf16 (bfloat16x4_t __a)
>> +{
>> +  return (int32x2_t)__a;
>> +}
>> +
>> +__extension__ extern __inline int64x1_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_s64_bf16 (bfloat16x4_t __a)
>> +{
>> +  return (int64x1_t)__a;
>> +}
>> +
>> +__extension__ extern __inline uint8x8_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_u8_bf16 (bfloat16x4_t __a)
>> +{
>> +  return (uint8x8_t)__a;
>> +}
>> +
>> +__extension__ extern __inline uint16x4_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_u16_bf16 (bfloat16x4_t __a)
>> +{
>> +  return (uint16x4_t)__a;
>> +}
>> +
>> +__extension__ extern __inline uint32x2_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_u32_bf16 (bfloat16x4_t __a)
>> +{
>> +  return (uint32x2_t)__a;
>> +}
>> +
>> +__extension__ extern __inline uint64x1_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_u64_bf16 (bfloat16x4_t __a)
>> +{
>> +  return (uint64x1_t)__a;
>> +}
>> +
>> +__extension__ extern __inline float16x4_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_f16_bf16 (bfloat16x4_t __a)
>> +{
>> +  return (float16x4_t)__a;
>> +}
>> +
>> +__extension__ extern __inline float32x2_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_f32_bf16 (bfloat16x4_t __a)
>> +{
>> +  return (float32x2_t)__a;
>> +}
>> +
>> +__extension__ extern __inline float32x2_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_f64_bf16 (bfloat16x4_t __a)
>> +{
>> +  return (float64x1_t)__a;
>> +}
>> +
>> +__extension__ extern __inline poly8x8_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_p8_bf16 (bfloat16x4_t __a)
>> +{
>> +  return (poly8x8_t)__a;
>> +}
>> +
>> +__extension__ extern __inline poly16x4_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_p16_bf16 (bfloat16x4_t __a)
>> +{
>> +  return (poly16x4_t)__a;
>> +}
>> +
>> +__extension__ extern __inline poly64x1_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpret_p64_bf16 (bfloat16x4_t __a)
>> +{
>> +  return (poly64x1_t)__a;
>> +}
>> +
>> +__extension__ extern __inline int8x16_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_s8_bf16 (bfloat16x8_t __a)
>> +{
>> +  return (int8x16_t)__a;
>> +}
>> +
>> +__extension__ extern __inline int16x8_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_s16_bf16 (bfloat16x8_t __a)
>> +{
>> +  return (int16x8_t)__a;
>> +}
>> +
>> +__extension__ extern __inline int32x4_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_s32_bf16 (bfloat16x8_t __a)
>> +{
>> +  return (int32x4_t)__a;
>> +}
>> +
>> +__extension__ extern __inline int64x2_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_s64_bf16 (bfloat16x8_t __a)
>> +{
>> +  return (int64x2_t)__a;
>> +}
>> +
>> +__extension__ extern __inline uint8x16_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_u8_bf16 (bfloat16x8_t __a)
>> +{
>> +  return (uint8x16_t)__a;
>> +}
>> +
>> +__extension__ extern __inline uint16x8_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_u16_bf16 (bfloat16x8_t __a)
>> +{
>> +  return (uint16x8_t)__a;
>> +}
>> +
>> +__extension__ extern __inline uint32x4_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_u32_bf16 (bfloat16x8_t __a)
>> +{
>> +  return (uint32x4_t)__a;
>> +}
>> +
>> +__extension__ extern __inline uint64x2_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_u64_bf16 (bfloat16x8_t __a)
>> +{
>> +  return (uint64x2_t)__a;
>> +}
>> +
>> +__extension__ extern __inline float16x8_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_f16_bf16 (bfloat16x8_t __a)
>> +{
>> +  return (float16x8_t)__a;
>> +}
>> +
>> +__extension__ extern __inline float32x4_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_f32_bf16 (bfloat16x8_t __a)
>> +{
>> +  return (float32x4_t)__a;
>> +}
>> +
>> +__extension__ extern __inline float64x2_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_f64_bf16 (bfloat16x8_t __a)
>> +{
>> +  return (float64x2_t)__a;
>> +}
>> +
>> +__extension__ extern __inline poly8x16_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_p8_bf16 (bfloat16x8_t __a)
>> +{
>> +  return (poly8x16_t)__a;
>> +}
>> +
>> +__extension__ extern __inline poly16x8_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_p16_bf16 (bfloat16x8_t __a)
>> +{
>> +  return (poly16x8_t)__a;
>> +}
>> +
>> +__extension__ extern __inline poly64x2_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_p64_bf16 (bfloat16x8_t __a)
>> +{
>> +  return (poly64x2_t)__a;
>> +}
>> +
>> +__extension__ extern __inline poly128_t
>> +__attribute__  ((__always_inline__, __gnu_inline__, __artificial__))
>> +vreinterpretq_p128_bf16 (bfloat16x8_t __a)
>> +{
>> +  return (poly128_t)__a;
>> +}
>> +
>>   __extension__ extern __inline float32x2_t
>>   __attribute__ ((__always_inline__, __gnu_inline__, __artificial__))
>>   vbfdot_f32 (float32x2_t __r, bfloat16x4_t __a, bfloat16x4_t __b)
>> diff --git a/gcc/config/aarch64/iterators.md 
>> b/gcc/config/aarch64/iterators.md
>> index 
>> fc7856e58c9ab65acce3fc43b18356c8c3ff6aae..76c8d9abbe79355078799a4ca227b2a352636ef4
>>  100644
>> --- a/gcc/config/aarch64/iterators.md
>> +++ b/gcc/config/aarch64/iterators.md
>> @@ -136,7 +136,8 @@
>>   (define_mode_iterator PTR [(SI "ptr_mode == SImode") (DI "ptr_mode == 
>> DImode")])
>>   
>>   ;; Advanced SIMD Float modes suitable for moving, loading and storing.
>> -(define_mode_iterator VDQF_F16 [V4HF V8HF V2SF V4SF V2DF])
>> +(define_mode_iterator VDQF_F16  [V4HF V8HF V2SF V4SF V2DF
>> +                             V4BF V8BF])
>>   
>>   ;; Advanced SIMD Float modes.
>>   (define_mode_iterator VDQF [V2SF V4SF V2DF])
>> @@ -177,7 +178,7 @@
>>   
>>   ;; All Advanced SIMD modes suitable for moving, loading, and storing.
>>   (define_mode_iterator VALL_F16 [V8QI V16QI V4HI V8HI V2SI V4SI V2DI
>> -                            V4HF V8HF V2SF V4SF V2DF])
>> +                            V4HF V8HF V4BF V8BF V2SF V4SF V2DF])
>>   
>>   ;; All Advanced SIMD modes suitable for moving, loading, and storing,
>>   ;; including special Bfloat vector types.
>> @@ -193,7 +194,7 @@
>>   
>>   ;; All Advanced SIMD modes and DI.
>>   (define_mode_iterator VALLDI_F16 [V8QI V16QI V4HI V8HI V2SI V4SI V2DI
>> -                              V4HF V8HF V2SF V4SF V2DF DI])
>> +                              V4HF V8HF V4BF V8BF V2SF V4SF V2DF DI])
>>   
>>   ;; All Advanced SIMD modes, plus DI and DF.
>>   (define_mode_iterator VALLDIF [V8QI V16QI V4HI V8HI V2SI V4SI
>> @@ -821,6 +822,7 @@
>>   
>>   (define_mode_attr Vtype [(V8QI "8b") (V16QI "16b")
>>                       (V4HI "4h") (V8HI  "8h")
>> +                     (V4BF "4h") (V8BF  "8h")
>>                            (V2SI "2s") (V4SI  "4s")
>>                            (DI   "1d") (DF    "1d")
>>                            (V2DI "2d") (V2SF "2s")
>> @@ -863,6 +865,7 @@
>>                        (VNx4SF "s") (VNx2SF "s")
>>                        (VNx2DI "d")
>>                        (VNx2DF "d")
>> +                      (BF "h") (V4BF "h") (V8BF "h")
>>                        (HF "h")
>>                        (SF "s") (DF "d")
>>                        (QI "b") (HI "h")
>> @@ -928,6 +931,7 @@
>>                     (DF   "DF") (V2DF  "DF")
>>                     (SI   "SI") (HI    "HI")
>>                     (QI   "QI")
>> +                   (V4BF "BF") (V8BF "BF")
>>                     (VNx16QI "QI") (VNx8QI "QI") (VNx4QI "QI") (VNx2QI "QI")
>>                     (VNx8HI "HI") (VNx4HI "HI") (VNx2HI "HI")
>>                     (VNx8HF "HF") (VNx4HF "HF") (VNx2HF "HF")
>> @@ -946,6 +950,7 @@
>>                     (V2DF "df") (DF   "df")
>>                     (SI   "si") (HI   "hi")
>>                     (QI   "qi")
>> +                   (V4BF "bf") (V8BF "bf")
>>                     (VNx16QI "qi") (VNx8QI "qi") (VNx4QI "qi") (VNx2QI "qi")
>>                     (VNx8HI "hi") (VNx4HI "hi") (VNx2HI "hi")
>>                     (VNx8HF "hf") (VNx4HF "hf") (VNx2HF "hf")
>> @@ -1249,6 +1254,7 @@
>>   
>>   (define_mode_attr VSWAP_WIDTH [(V8QI "V16QI") (V16QI "V8QI")
>>                              (V4HI "V8HI") (V8HI  "V4HI")
>> +                            (V8BF "V4BF") (V4BF  "V8BF")
>>                              (V2SI "V4SI") (V4SI  "V2SI")
>>                              (DI   "V2DI") (V2DI  "DI")
>>                              (V2SF "V4SF") (V4SF  "V2SF")
>> @@ -1261,6 +1267,7 @@
>>                                  (DI   "to_128") (V2DI  "to_64")
>>                                  (V4HF "to_128") (V8HF  "to_64")
>>                                  (V2SF "to_128") (V4SF  "to_64")
>> +                                (V4BF "to_128") (V8BF  "to_64")
>>                                  (DF   "to_128") (V2DF  "to_64")])
>>   
>>   ;; For certain vector-by-element multiplication instructions we must
>> @@ -1294,6 +1301,7 @@
>>   ;; Defined to '_q' for 128-bit types.
>>   (define_mode_attr q [(V8QI "") (V16QI "_q")
>>                   (V4HI "") (V8HI  "_q")
>> +                 (V4BF "") (V8BF  "_q")
>>                   (V2SI "") (V4SI  "_q")
>>                   (DI   "") (V2DI  "_q")
>>                   (V4HF "") (V8HF "_q")
>> diff --git a/gcc/testsuite/gcc.target/aarch64/advsimd-intrinsics/bf16_dup.c 
>> b/gcc/testsuite/gcc.target/aarch64/advsimd-intrinsics/bf16_dup.c
>> new file mode 100644
>> index 
>> 0000000000000000000000000000000000000000..6f9eb3a9da6b1395a60b63c0a8d21aba366dd12d
>> --- /dev/null
>> +++ b/gcc/testsuite/gcc.target/aarch64/advsimd-intrinsics/bf16_dup.c
>> @@ -0,0 +1,85 @@
>> +/* { dg-do assemble { target { aarch64*-*-* } } } */
>> +/* { dg-require-effective-target arm_v8_2a_bf16_neon_ok } */
>> +/* { dg-options "-O2" } */
>> +/* { dg-add-options arm_v8_2a_bf16_neon }  */
>> +/* { dg-additional-options "-save-temps" } */
>> +
>> +#include <arm_neon.h>
>> +
>> +float32x2_t test_vcreate (float32x2_t r, uint64_t a, uint64_t b)
>> +{
>> +  bfloat16x4_t _a = vcreate_bf16(a);
>> +  bfloat16x4_t _b = vcreate_bf16(b);
>> +
>> +  return vbfdot_f32 (r, _a, _b);
>> +}
>> +/* { dg-final { scan-assembler {bfdot\tv[0-9]+.2s, v[0-9]+.4h, v[0-9]+.4h} 
>> } } */
>> +
>> +bfloat16x4_t test_vset_lane_bf16 (bfloat16_t a, bfloat16x4_t b)
>> +{
>> +  return vset_lane_bf16 (a, b, 3);
>> +}
>> +
>> +bfloat16x8_t test_vsetq_lane_bf16 (bfloat16_t a, bfloat16x8_t b)
>> +{
>> +  return vsetq_lane_bf16 (a, b, 7);
>> +}
>> +/* { dg-final { scan-assembler-times "ins\\t" 2 } } */
>> +
>> +bfloat16x4_t vdup_test (bfloat16_t a)
>> +{
>> +  return vdup_n_bf16 (a);
>> +}
>> +/* { dg-final { scan-assembler "dup\\tv\[0-9\]+\.4h, v\[0-9\]+.h\\\[0\\\]" 
>> } } */
>> +
>> +bfloat16x8_t vdupq_test (bfloat16_t a)
>> +{
>> +  return vdupq_n_bf16 (a);
>> +}
>> +
>> +bfloat16x8_t test_vdupq_lane_bf16 (bfloat16x4_t a)
>> +{
>> +  return vdupq_lane_bf16 (a, 1);
>> +}
>> +/* { dg-final { scan-assembler-times "dup\\tv\[0-9\]+\.8h, 
>> v\[0-9\]+.h\\\[0\\\]" 2 } } */
>> +
>> +bfloat16_t test_vget_lane_bf16 (bfloat16x4_t a)
>> +{
>> +  return vget_lane_bf16 (a, 1);
>> +}
>> +/* { dg-final { scan-assembler-times "dup\\th\[0-9\]+, 
>> v\[0-9\]+\.h\\\[1\\\]" 2 } } */
>> +
>> +bfloat16x4_t test_vdup_lane_bf16 (bfloat16x4_t a)
>> +{
>> +  return vdup_lane_bf16 (a, 1);
>> +}
>> +/* { dg-final { scan-assembler "dup\\tv\[0-9\]+\.4h, v\[0-9\]+\.h\\\[1\\\]" 
>> } } */
>> +
>> +bfloat16x4_t test_vdup_laneq_bf16 (bfloat16x8_t a)
>> +{
>> +  return vdup_laneq_bf16 (a, 7);
>> +}
>> +/* { dg-final { scan-assembler "tbl\\tv\[0-9\]+\.16b, {v\[0-9\]+\.16b}, 
>> v\[0-9\]+\.16b" } } */
>> +
>> +bfloat16x8_t test_vdupq_laneq_bf16 (bfloat16x8_t a)
>> +{
>> +  return vdupq_laneq_bf16 (a, 5);
>> +}
>> +/* { dg-final { scan-assembler "dup\\tv\[0-9\]+\.8h, v\[0-9\]+\.h\\\[5\\\]" 
>> } } */
>> +
>> +bfloat16_t test_vduph_lane_bf16 (bfloat16x4_t a)
>> +{
>> +  return vduph_lane_bf16 (a, 3);
>> +}
>> +/* { dg-final { scan-assembler "dup\\th\[0-9\]+, v\[0-9\]+\.h\\\[3\\\]" } } 
>> */
>> +
>> +bfloat16_t test_vgetq_lane_bf16 (bfloat16x8_t a)
>> +{
>> +  return vgetq_lane_bf16 (a, 7);
>> +}
>> +
>> +bfloat16_t test_vduph_laneq_bf16 (bfloat16x8_t a)
>> +{
>> +  return vduph_laneq_bf16 (a, 7);
>> +}
>> +/* { dg-final { scan-assembler-times "dup\\th\[0-9\]+, 
>> v\[0-9\]+\.h\\\[7\\\]" 2 } } */
>> diff --git 
>> a/gcc/testsuite/gcc.target/aarch64/advsimd-intrinsics/bf16_reinterpret.c 
>> b/gcc/testsuite/gcc.target/aarch64/advsimd-intrinsics/bf16_reinterpret.c
>> new file mode 100644
>> index 
>> 0000000000000000000000000000000000000000..f5adf40c648e16c649ef5d68accb291b822f2936
>> --- /dev/null
>> +++ b/gcc/testsuite/gcc.target/aarch64/advsimd-intrinsics/bf16_reinterpret.c
>> @@ -0,0 +1,466 @@
>> +/* { dg-do assemble { target { aarch64*-*-* } } } */
>> +/* { dg-require-effective-target arm_v8_2a_bf16_neon_ok } */
>> +/* { dg-add-options arm_v8_2a_bf16_neon }  */
>> +/* { dg-additional-options "-save-temps" } */
>> +
>> +#include <arm_neon.h>
>> +
>> +float32x2_t
>> +test_vbfdot_f32_s8 (float32x2_t r, int8x8_t a, int8x8_t b)
>> +{
>> +  bfloat16x4_t _a = vreinterpret_bf16_s8(a);
>> +  bfloat16x4_t _b = vreinterpret_bf16_s8(b);
>> +
>> +  return vbfdot_f32 (r, _a, _b);
>> +}
>> +
>> +float32x2_t
>> +test_vbfdot_f32_s16 (float32x2_t r, int16x4_t a, int16x4_t b)
>> +{
>> +  bfloat16x4_t _a = vreinterpret_bf16_s16(a);
>> +  bfloat16x4_t _b = vreinterpret_bf16_s16(b);
>> +
>> +  return vbfdot_f32 (r, _a, _b);
>> +}
>> +
>> +float32x2_t
>> +test_vbfdot_f32_s32 (float32x2_t r, int32x2_t a, int32x2_t b)
>> +{
>> +  bfloat16x4_t _a = vreinterpret_bf16_s32(a);
>> +  bfloat16x4_t _b = vreinterpret_bf16_s32(b);
>> +
>> +  return vbfdot_f32 (r, _a, _b);
>> +}
>> +
>> +float32x2_t
>> +test_vbfdot_f32_s64 (float32x2_t r, int64x1_t a, int64x1_t b)
>> +{
>> +  bfloat16x4_t _a = vreinterpret_bf16_s64(a);
>> +  bfloat16x4_t _b = vreinterpret_bf16_s64(b);
>> +
>> +  return vbfdot_f32 (r, _a, _b);
>> +}
>> +
>> +float32x2_t
>> +test_vbfdot_f32_u8 (float32x2_t r, uint8x8_t a, uint8x8_t b)
>> +{
>> +  bfloat16x4_t _a = vreinterpret_bf16_u8(a);
>> +  bfloat16x4_t _b = vreinterpret_bf16_u8(b);
>> +
>> +  return vbfdot_f32 (r, _a, _b);
>> +}
>> +
>> +float32x2_t
>> +test_vbfdot_f32_u16 (float32x2_t r, uint16x4_t a, uint16x4_t b)
>> +{
>> +  bfloat16x4_t _a = vreinterpret_bf16_u16(a);
>> +  bfloat16x4_t _b = vreinterpret_bf16_u16(b);
>> +
>> +  return vbfdot_f32 (r, _a, _b);
>> +}
>> +
>> +float32x2_t
>> +test_vbfdot_f32_u32 (float32x2_t r, uint32x2_t a, uint32x2_t b)
>> +{
>> +  bfloat16x4_t _a = vreinterpret_bf16_u32(a);
>> +  bfloat16x4_t _b = vreinterpret_bf16_u32(b);
>> +
>> +  return vbfdot_f32 (r, _a, _b);
>> +}
>> +
>> +float32x2_t
>> +test_vbfdot_f32_u64 (float32x2_t r, uint64x1_t a, uint64x1_t b)
>> +{
>> +  bfloat16x4_t _a = vreinterpret_bf16_u64(a);
>> +  bfloat16x4_t _b = vreinterpret_bf16_u64(b);
>> +
>> +  return vbfdot_f32 (r, _a, _b);
>> +}
>> +
>> +float32x2_t
>> +test_vbfdot_f32_p8 (float32x2_t r, poly8x8_t a, poly8x8_t b)
>> +{
>> +  bfloat16x4_t _a = vreinterpret_bf16_p8(a);
>> +  bfloat16x4_t _b = vreinterpret_bf16_p8(b);
>> +
>> +  return vbfdot_f32 (r, _a, _b);
>> +}
>> +
>> +float32x2_t
>> +test_vbfdot_f32_p16 (float32x2_t r, poly16x4_t a, poly16x4_t b)
>> +{
>> +  bfloat16x4_t _a = vreinterpret_bf16_p16(a);
>> +  bfloat16x4_t _b = vreinterpret_bf16_p16(b);
>> +
>> +  return vbfdot_f32 (r, _a, _b);
>> +}
>> +
>> +float32x2_t
>> +test_vbfdot_f32_p64 (float32x2_t r, poly64x1_t a, poly64x1_t b)
>> +{
>> +  bfloat16x4_t _a = vreinterpret_bf16_p64(a);
>> +  bfloat16x4_t _b = vreinterpret_bf16_p64(b);
>> +
>> +  return vbfdot_f32 (r, _a, _b);
>> +}
>> +
>> +float32x2_t
>> +test_vbfdot_f32_f16 (float32x2_t r, float16x4_t a, float16x4_t b)
>> +{
>> +  bfloat16x4_t _a = vreinterpret_bf16_f16(a);
>> +  bfloat16x4_t _b = vreinterpret_bf16_f16(b);
>> +
>> +  return vbfdot_f32 (r, _a, _b);
>> +}
>> +
>> +float32x2_t
>> +test_vbfdot_f32_f32 (float32x2_t r, float32x2_t a, float32x2_t b)
>> +{
>> +  bfloat16x4_t _a = vreinterpret_bf16_f32(a);
>> +  bfloat16x4_t _b = vreinterpret_bf16_f32(b);
>> +
>> +  return vbfdot_f32 (r, _a, _b);
>> +}
>> +
>> +float32x2_t
>> +test_vbfdot_f32_f64 (float32x2_t r, float64x1_t a, float64x1_t b)
>> +{
>> +  bfloat16x4_t _a = vreinterpret_bf16_f64(a);
>> +  bfloat16x4_t _b = vreinterpret_bf16_f64(b);
>> +
>> +  return vbfdot_f32 (r, _a, _b);
>> +}
>> +
>> +float32x4_t
>> +test_vbfdotq_f32_s8 (float32x4_t r, int8x16_t a, int8x16_t b)
>> +{
>> +  bfloat16x8_t _a = vreinterpretq_bf16_s8(a);
>> +  bfloat16x8_t _b = vreinterpretq_bf16_s8(b);
>> +
>> +  return vbfdotq_f32 (r, _a, _b);
>> +}
>> +
>> +float32x4_t
>> +test_vbfdotq_f32_s16 (float32x4_t r, int16x8_t a, int16x8_t b)
>> +{
>> +  bfloat16x8_t _a = vreinterpretq_bf16_s16(a);
>> +  bfloat16x8_t _b = vreinterpretq_bf16_s16(b);
>> +
>> +  return vbfdotq_f32 (r, _a, _b);
>> +}
>> +
>> +float32x4_t
>> +test_vbfdotq_f32_s32 (float32x4_t r, int32x4_t a, int32x4_t b)
>> +{
>> +  bfloat16x8_t _a = vreinterpretq_bf16_s32(a);
>> +  bfloat16x8_t _b = vreinterpretq_bf16_s32(b);
>> +
>> +  return vbfdotq_f32 (r, _a, _b);
>> +}
>> +
>> +float32x4_t
>> +test_vbfdotq_f32_s64 (float32x4_t r, int64x2_t a, int64x2_t b)
>> +{
>> +  bfloat16x8_t _a = vreinterpretq_bf16_s64(a);
>> +  bfloat16x8_t _b = vreinterpretq_bf16_s64(b);
>> +
>> +  return vbfdotq_f32 (r, _a, _b);
>> +}
>> +
>> +float32x4_t
>> +test_vbfdotq_f32_u8 (float32x4_t r, uint8x16_t a, uint8x16_t b)
>> +{
>> +  bfloat16x8_t _a = vreinterpretq_bf16_u8(a);
>> +  bfloat16x8_t _b = vreinterpretq_bf16_u8(b);
>> +
>> +  return vbfdotq_f32 (r, _a, _b);
>> +}
>> +
>> +float32x4_t
>> +test_vbfdotq_f32_u16 (float32x4_t r, uint16x8_t a, uint16x8_t b)
>> +{
>> +  bfloat16x8_t _a = vreinterpretq_bf16_u16(a);
>> +  bfloat16x8_t _b = vreinterpretq_bf16_u16(b);
>> +
>> +  return vbfdotq_f32 (r, _a, _b);
>> +}
>> +
>> +float32x4_t
>> +test_vbfdotq_f32_u32 (float32x4_t r, uint32x4_t a, uint32x4_t b)
>> +{
>> +  bfloat16x8_t _a = vreinterpretq_bf16_u32(a);
>> +  bfloat16x8_t _b = vreinterpretq_bf16_u32(b);
>> +
>> +  return vbfdotq_f32 (r, _a, _b);
>> +}
>> +
>> +float32x4_t
>> +test_vbfdotq_f32_u64 (float32x4_t r, uint64x2_t a, uint64x2_t b)
>> +{
>> +  bfloat16x8_t _a = vreinterpretq_bf16_u64(a);
>> +  bfloat16x8_t _b = vreinterpretq_bf16_u64(b);
>> +
>> +  return vbfdotq_f32 (r, _a, _b);
>> +}
>> +
>> +float32x4_t
>> +test_vbfdotq_f32_p8 (float32x4_t r, poly8x16_t a, poly8x16_t b)
>> +{
>> +  bfloat16x8_t _a = vreinterpretq_bf16_p8(a);
>> +  bfloat16x8_t _b = vreinterpretq_bf16_p8(b);
>> +
>> +  return vbfdotq_f32 (r, _a, _b);
>> +}
>> +
>> +float32x4_t
>> +test_vbfdotq_f32_p16 (float32x4_t r, poly16x8_t a, poly16x8_t b)
>> +{
>> +  bfloat16x8_t _a = vreinterpretq_bf16_p16(a);
>> +  bfloat16x8_t _b = vreinterpretq_bf16_p16(b);
>> +
>> +  return vbfdotq_f32 (r, _a, _b);
>> +}
>> +
>> +float32x4_t
>> +test_vbfdotq_f32_p64 (float32x4_t r, poly64x2_t a, poly64x2_t b)
>> +{
>> +  bfloat16x8_t _a = vreinterpretq_bf16_p64(a);
>> +  bfloat16x8_t _b = vreinterpretq_bf16_p64(b);
>> +
>> +  return vbfdotq_f32 (r, _a, _b);
>> +}
>> +
>> +float32x4_t
>> +test_vbfdotq_f32_p128 (float32x4_t r, poly128_t a, poly128_t b)
>> +{
>> +  bfloat16x8_t _a = vreinterpretq_bf16_p128(a);
>> +  bfloat16x8_t _b = vreinterpretq_bf16_p128(b);
>> +
>> +  return vbfdotq_f32 (r, _a, _b);
>> +}
>> +
>> +float32x4_t
>> +test_vbfdotq_f32_f16 (float32x4_t r, float16x8_t a, float16x8_t b)
>> +{
>> +  bfloat16x8_t _a = vreinterpretq_bf16_f16(a);
>> +  bfloat16x8_t _b = vreinterpretq_bf16_f16(b);
>> +
>> +  return vbfdotq_f32 (r, _a, _b);
>> +}
>> +
>> +float32x4_t
>> +test_vbfdotq_f32_f32 (float32x4_t r, float32x4_t a, float32x4_t b)
>> +{
>> +  bfloat16x8_t _a = vreinterpretq_bf16_f32(a);
>> +  bfloat16x8_t _b = vreinterpretq_bf16_f32(b);
>> +
>> +  return vbfdotq_f32 (r, _a, _b);
>> +}
>> +
>> +float32x4_t
>> +test_vbfdotq_f32_f64 (float32x4_t r, float64x2_t a, float64x2_t b)
>> +{
>> +  bfloat16x8_t _a = vreinterpretq_bf16_f64(a);
>> +  bfloat16x8_t _b = vreinterpretq_bf16_f64(b);
>> +
>> +  return vbfdotq_f32 (r, _a, _b);
>> +}
>> +
>> +/* { dg-final { scan-assembler-times {bfdot\tv[0-9]+.2s, v[0-9]+.4h, 
>> v[0-9]+.4h} 14 } } */
>> +/* { dg-final { scan-assembler-times {bfdot\tv[0-9]+.4s, v[0-9]+.8h, 
>> v[0-9]+.8h} 15 } } */
>> +
>> +int8x8_t test_vreinterpret_s8_bf16 (bfloat16x4_t a, int8x8_t b)
>> +{
>> +  int8x8_t _a = vreinterpret_s8_bf16 (a);
>> +  return vadd_s8 (_a, b);
>> +}
>> +
>> +int16x4_t test_vreinterpret_s16_bf16 (bfloat16x4_t a, int16x4_t b)
>> +{
>> +  int16x4_t _a = vreinterpret_s16_bf16 (a);
>> +  return vadd_s16 (_a, b);
>> +}
>> +
>> +int32x2_t test_vreinterpret_s32_bf16 (bfloat16x4_t a, int32x2_t b)
>> +{
>> +  int32x2_t _a = vreinterpret_s32_bf16 (a);
>> +  return vadd_s32 (_a, b);
>> +}
>> +
>> +int64x1_t test_vreinterpret_s64_bf16 (bfloat16x4_t a, int64x1_t b)
>> +{
>> +  int64x1_t _a = vreinterpret_s64_bf16 (a);
>> +  return vrshl_s64 (_a, b);
>> +}
>> +
>> +uint8x8_t test_vreinterpret_u8_bf16 (bfloat16x4_t a, uint8x8_t b)
>> +{
>> +  uint8x8_t _a = vreinterpret_u8_bf16 (a);
>> +  return vadd_u8 (_a, b);
>> +}
>> +
>> +uint16x4_t test_vreinterpret_u16_bf16 (bfloat16x4_t a, uint16x4_t b)
>> +{
>> +  uint16x4_t _a = vreinterpret_u16_bf16 (a);
>> +  return vadd_u16 (_a, b);
>> +}
>> +
>> +uint32x2_t test_vreinterpret_u32_bf16 (bfloat16x4_t a, uint32x2_t b)
>> +{
>> +  uint32x2_t _a = vreinterpret_u32_bf16 (a);
>> +  return vadd_u32 (_a, b);
>> +}
>> +
>> +uint64x1_t test_vreinterpret_u64_bf16 (bfloat16x4_t a, int64x1_t b)
>> +{
>> +  uint64x1_t _a = vreinterpret_u64_bf16 (a);
>> +  return vrshl_u64 (_a, b);
>> +}
>> +
>> +poly8x8_t test_vreinterpret_p8_bf16 (bfloat16x4_t a, poly8x8_t b)
>> +{
>> +  poly8x8_t _a = vreinterpret_p8_bf16 (a);
>> +  return vzip1_p8 (_a, b);
>> +}
>> +
>> +poly16x4_t test_vreinterpret_p16_bf16 (bfloat16x4_t a, poly16x4_t b)
>> +{
>> +  poly16x4_t _a = vreinterpret_p16_bf16 (a);
>> +  return vzip1_p16 (_a, b);
>> +}
>> +
>> +poly64x1_t test_vreinterpret_p64_bf16 (bfloat16x4_t a, poly64x1_t b)
>> +{
>> +  poly64x1_t _a = vreinterpret_p64_bf16 (a);
>> +  return vsli_n_p64 (_a, b, 3);
>> +}
>> +
>> +float32x2_t test_vreinterpret_f32_bf16 (bfloat16x4_t a, float32x2_t b)
>> +{
>> +  float32x2_t _a = vreinterpret_f32_bf16 (a);
>> +  return vsub_f32 (_a, b);
>> +}
>> +
>> +float64x1_t test_vreinterpret_f64_bf16 (bfloat16x4_t a, float64x1_t b)
>> +{
>> +  float64x1_t _a = vreinterpret_f64_bf16 (a);
>> +  return vsub_f64 (_a, b);
>> +}
>> +
>> +int8x16_t test_vreinterpretq_s8_bf16 (bfloat16x8_t a, int8x16_t b)
>> +{
>> +  int8x16_t _a = vreinterpretq_s8_bf16 (a);
>> +  return vaddq_s8 (_a, b);
>> +}
>> +
>> +int16x8_t test_vreinterpretq_s16_bf16 (bfloat16x8_t a, int16x8_t b)
>> +{
>> +  int16x8_t _a = vreinterpretq_s16_bf16 (a);
>> +  return vaddq_s16 (_a, b);
>> +}
>> +
>> +int32x4_t test_vreinterpretq_s32_bf16 (bfloat16x8_t a, int32x4_t b)
>> +{
>> +  int32x4_t _a = vreinterpretq_s32_bf16 (a);
>> +  return vaddq_s32 (_a, b);
>> +}
>> +
>> +int64x2_t test_vreinterpretq_s64_bf16 (bfloat16x8_t a, int64x2_t b)
>> +{
>> +  int64x2_t _a = vreinterpretq_s64_bf16 (a);
>> +  return vaddq_s64 (_a, b);
>> +}
>> +
>> +uint8x16_t test_vreinterpretq_u8_bf16 (bfloat16x8_t a, uint8x16_t b)
>> +{
>> +  uint8x16_t _a = vreinterpretq_u8_bf16 (a);
>> +  return vaddq_u8 (_a, b);
>> +}
>> +
>> +uint16x8_t test_vreinterpretq_u16_bf16 (bfloat16x8_t a, uint16x8_t b)
>> +{
>> +  uint16x8_t _a = vreinterpretq_u16_bf16 (a);
>> +  return vaddq_u16 (_a, b);
>> +}
>> +
>> +uint32x4_t test_vreinterpretq_u32_bf16 (bfloat16x8_t a, uint32x4_t b)
>> +{
>> +  uint32x4_t _a = vreinterpretq_u32_bf16 (a);
>> +  return vaddq_u32 (_a, b);
>> +}
>> +
>> +uint64x2_t test_vreinterpretq_u64_bf16 (bfloat16x8_t a, uint64x2_t b)
>> +{
>> +  uint64x2_t _a = vreinterpretq_u64_bf16 (a);
>> +  return vaddq_u64 (_a, b);
>> +}
>> +
>> +poly8x16_t test_vreinterpretq_p8_bf16 (bfloat16x8_t a, poly8x16_t b)
>> +{
>> +  poly8x16_t _a = vreinterpretq_p8_bf16 (a);
>> +  return vzip1q_p8 (_a, b);
>> +}
>> +
>> +poly16x8_t test_vreinterpretq_p16_bf16 (bfloat16x8_t a, poly16x8_t b)
>> +{
>> +  poly16x8_t _a = vreinterpretq_p16_bf16 (a);
>> +  return vzip1q_p16 (_a, b);
>> +}
>> +
>> +poly64x2_t test_vreinterpretq_p64_bf16 (bfloat16x8_t a, poly64x2_t b)
>> +{
>> +  poly64x2_t _a = vreinterpretq_p64_bf16 (a);
>> +  return vsliq_n_p64 (_a, b, 3);
>> +}
>> +
>> +poly128_t test_vreinterpretq_p128_bf16 (bfloat16x8_t a, poly16x8_t b)
>> +{
>> +  poly128_t _a = vreinterpretq_p128_bf16 (a);
>> +  return _a;
>> +}
>> +
>> +float32x4_t test_vreinterpretq_f32_bf16 (bfloat16x8_t a, float32x4_t b)
>> +{
>> +  float32x4_t _a = vreinterpretq_f32_bf16 (a);
>> +  return vsubq_f32 (_a, b);
>> +}
>> +
>> +float64x2_t test_vreinterpretq_f64_bf16 (bfloat16x8_t a, float64x2_t b)
>> +{
>> +  float64x2_t _a = vreinterpretq_f64_bf16 (a);
>> +  return vsubq_f64 (_a, b);
>> +}
>> +
>> +float16x4_t test_vreinterpret_f16_bf16 (bfloat16x4_t a)
>> +{
>> +  return vreinterpret_f16_bf16 (a);
>> +}
>> +
>> +float16x8_t test_vreinterpretq_f16_bf16 (bfloat16x8_t a)
>> +{
>> +  return vreinterpretq_f16_bf16 (a);
>> +}
>> +
>> +/* { dg-final { scan-assembler-times {add\tv[0-9]+.2s, v[0-9]+.2s, 
>> v[0-9]+.2s} 2 } } */
>> +/* { dg-final { scan-assembler-times {add\tv[0-9]+.4h, v[0-9]+.4h, 
>> v[0-9]+.4h} 2 } } */
>> +/* { dg-final { scan-assembler-times {add\tv[0-9]+.8b, v[0-9]+.8b, 
>> v[0-9]+.8b} 2 } } */
>> +
>> +/* { dg-final { scan-assembler-times {add\tv[0-9]+.4s, v[0-9]+.4s, 
>> v[0-9]+.4s} 2 } } */
>> +/* { dg-final { scan-assembler-times {add\tv[0-9]+.8h, v[0-9]+.8h, 
>> v[0-9]+.8h} 2 } } */
>> +/* { dg-final { scan-assembler-times {add\tv[0-9]+.16b, v[0-9]+.16b, 
>> v[0-9]+.16b} 2 } } */
>> +
>> +/* { dg-final { scan-assembler {fsub\tv[0-9]+.2s, v[0-9]+.2s, v[0-9]+.2s} } 
>> } */
>> +/* { dg-final { scan-assembler {fsub\tv[0-9]+.4s, v[0-9]+.4s, v[0-9]+.4s} } 
>> } */
>> +/* { dg-final { scan-assembler {fsub\tv[0-9]+.2d, v[0-9]+.2d, v[0-9]+.2d} } 
>> } */
>> +/* { dg-final { scan-assembler {fsub\td[0-9]+, d[0-9]+, d[0-9]+} } } */
>> +
>> +/* { dg-final { scan-assembler {zip1\tv[0-9]+.8b, v[0-9]+.8b, v[0-9]+.8b} } 
>> } */
>> +/* { dg-final { scan-assembler {zip1\tv[0-9]+.16b, v[0-9]+.16b, 
>> v[0-9]+.16b} } } */
>> +/* { dg-final { scan-assembler {zip1\tv[0-9]+.4h, v[0-9]+.4h, v[0-9]+.4h} } 
>> } */
>> +/* { dg-final { scan-assembler {zip1\tv[0-9]+.8h, v[0-9]+.8h, v[0-9]+.8h} } 
>> } */
>> +
>> +/* { dg-final { scan-assembler {sli\tv[0-9]+.2d, v[0-9]+.2d, 3} } } */
>> +/* { dg-final { scan-assembler {sli\td[0-9]+, d[0-9]+, 3} } } */
>> +
>> +/* { dg-final { scan-assembler {urshl\td[0-9]+, d[0-9]+, d[0-9]+} } } */
>> +/* { dg-final { scan-assembler {srshl\td[0-9]+, d[0-9]+, d[0-9]+} } } */
>> 
>
> Found a small issue when rebasing on latest trunk.
> vdup_laneq_bf16 was checked against the wrong instruction. It generates
> a dup, not a tbl. I've updated the patch with the fix.
>
> Is it ok for trunk?

LGTM thanks, pushed.

Richard

Reply via email to