[PATCH 6/6] aarch64: Pass and return Neon vector-tuple types without a parallel

Richard Sandiford richard.sandiford@arm.com
Fri Oct 22 15:16:27 GMT 2021


Jonathan Wright <Jonathan.Wright@arm.com> writes:
> Hi,
>
> Neon vector-tuple types can be passed in registers on function call
> and return - there is no need to generate a parallel rtx. This patch
> adds cases to detect vector-tuple modes and generates an appropriate
> register rtx.
>
> This change greatly improves code generated when passing Neon vector-
> tuple types between functions;

Indeed.

> many new test cases are added to defend these improvements.
>
> Bootstrapped and regression tested on aarch64-none-linux-gnu and
> aarch64_be-none-linux-gnu - no issues.
>
> Ok for master?
>
> Thanks,
> Jonathan
>
> ---
>
> gcc/ChangeLog:
>
> 2021-10-07  Jonathan Wright  <jonathan.wright@arm.com>
>
>         * config/aarch64/aarch64.c (aarch64_function_value): Generate
>         a register rtx for Neon vector-tuple modes.
>         (aarch64_layout_arg): Likewise.
>
> gcc/testsuite/ChangeLog:
>
>         * gcc.target/aarch64/vector_structure_intrinsics.c: New code
>         generation tests.

OK, thanks.

Richard

> diff --git a/gcc/config/aarch64/aarch64.c b/gcc/config/aarch64/aarch64.c
> index cbfcf7efcca8e0978518b69cbeafb6812c38889a..9c2b3cb7d677a1570b32a8c9b6ee14bef156cb45 100644
> --- a/gcc/config/aarch64/aarch64.c
> +++ b/gcc/config/aarch64/aarch64.c
> @@ -6433,6 +6433,12 @@ aarch64_function_value (const_tree type, const_tree func,
>  	  gcc_assert (count == 1 && mode == ag_mode);
>  	  return gen_rtx_REG (mode, V0_REGNUM);
>  	}
> +      else if (aarch64_advsimd_full_struct_mode_p (mode)
> +	       && known_eq (GET_MODE_SIZE (ag_mode), 16))
> +	return gen_rtx_REG (mode, V0_REGNUM);
> +      else if (aarch64_advsimd_partial_struct_mode_p (mode)
> +	       && known_eq (GET_MODE_SIZE (ag_mode), 8))
> +	return gen_rtx_REG (mode, V0_REGNUM);
>        else
>  	{
>  	  int i;
> @@ -6728,6 +6734,12 @@ aarch64_layout_arg (cumulative_args_t pcum_v, const function_arg_info &arg)
>  	      gcc_assert (nregs == 1);
>  	      pcum->aapcs_reg = gen_rtx_REG (mode, V0_REGNUM + nvrn);
>  	    }
> +	  else if (aarch64_advsimd_full_struct_mode_p (mode)
> +		   && known_eq (GET_MODE_SIZE (pcum->aapcs_vfp_rmode), 16))
> +	    pcum->aapcs_reg = gen_rtx_REG (mode, V0_REGNUM + nvrn);
> +	  else if (aarch64_advsimd_partial_struct_mode_p (mode)
> +		   && known_eq (GET_MODE_SIZE (pcum->aapcs_vfp_rmode), 8))
> +	    pcum->aapcs_reg = gen_rtx_REG (mode, V0_REGNUM + nvrn);
>  	  else
>  	    {
>  	      rtx par;
> diff --git a/gcc/testsuite/gcc.target/aarch64/vector_structure_intrinsics.c b/gcc/testsuite/gcc.target/aarch64/vector_structure_intrinsics.c
> index 89e9de18a92dbc00e58261e4558b3cff38c7ca75..100739ab4e67e27a7341b8b1a4ddd9494f0e181d 100644
> --- a/gcc/testsuite/gcc.target/aarch64/vector_structure_intrinsics.c
> +++ b/gcc/testsuite/gcc.target/aarch64/vector_structure_intrinsics.c
> @@ -17,6 +17,14 @@ TEST_TBL (vqtbl2q, int8x16_t, int8x16x2_t, uint8x16_t, s8)
>  TEST_TBL (vqtbl2q, uint8x16_t, uint8x16x2_t, uint8x16_t, u8)
>  TEST_TBL (vqtbl2q, poly8x16_t, poly8x16x2_t, uint8x16_t, p8)
>  
> +TEST_TBL (vqtbl3, int8x8_t, int8x16x3_t, uint8x8_t, s8)
> +TEST_TBL (vqtbl3, uint8x8_t, uint8x16x3_t, uint8x8_t, u8)
> +TEST_TBL (vqtbl3, poly8x8_t, poly8x16x3_t, uint8x8_t, p8)
> +
> +TEST_TBL (vqtbl3q, int8x16_t, int8x16x3_t, uint8x16_t, s8)
> +TEST_TBL (vqtbl3q, uint8x16_t, uint8x16x3_t, uint8x16_t, u8)
> +TEST_TBL (vqtbl3q, poly8x16_t, poly8x16x3_t, uint8x16_t, p8)
> +
>  TEST_TBL (vqtbl4, int8x8_t, int8x16x4_t, uint8x8_t, s8)
>  TEST_TBL (vqtbl4, uint8x8_t, uint8x16x4_t, uint8x8_t, u8)
>  TEST_TBL (vqtbl4, poly8x8_t, poly8x16x4_t, uint8x8_t, p8)
> @@ -25,62 +33,35 @@ TEST_TBL (vqtbl4q, int8x16_t, int8x16x4_t, uint8x16_t, s8)
>  TEST_TBL (vqtbl4q, uint8x16_t, uint8x16x4_t, uint8x16_t, u8)
>  TEST_TBL (vqtbl4q, poly8x16_t, poly8x16x4_t, uint8x16_t, p8)
>  
> -#define TEST_TBL3(name, rettype, tbltype, idxtype, ts) \
> -  rettype test_ ## name ## _ ## ts (idxtype a, tbltype b) \
> -	{ \
> -		return name ## _ ## ts (b, a); \
> -	}
> -
> -TEST_TBL3 (vqtbl3, int8x8_t, int8x16x3_t, uint8x8_t, s8)
> -TEST_TBL3 (vqtbl3, uint8x8_t, uint8x16x3_t, uint8x8_t, u8)
> -TEST_TBL3 (vqtbl3, poly8x8_t, poly8x16x3_t, uint8x8_t, p8)
> -
> -TEST_TBL3 (vqtbl3q, int8x16_t, int8x16x3_t, uint8x16_t, s8)
> -TEST_TBL3 (vqtbl3q, uint8x16_t, uint8x16x3_t, uint8x16_t, u8)
> -TEST_TBL3 (vqtbl3q, poly8x16_t, poly8x16x3_t, uint8x16_t, p8)
> -
> -#define TEST_TBX2(name, rettype, tbltype, idxtype, ts) \
> -  rettype test_ ## name ## _ ## ts (rettype a, idxtype b, tbltype c) \
> -	{ \
> -		return name ## _ ## ts (a, c, b); \
> -	}
> -
> -TEST_TBX2 (vqtbx2, int8x8_t, int8x16x2_t, uint8x8_t, s8)
> -TEST_TBX2 (vqtbx2, uint8x8_t, uint8x16x2_t, uint8x8_t, u8)
> -TEST_TBX2 (vqtbx2, poly8x8_t, poly8x16x2_t, uint8x8_t, p8)
> -
> -TEST_TBX2 (vqtbx2q, int8x16_t, int8x16x2_t, uint8x16_t, s8)
> -TEST_TBX2 (vqtbx2q, uint8x16_t, uint8x16x2_t, uint8x16_t, u8)
> -TEST_TBX2 (vqtbx2q, poly8x16_t, poly8x16x2_t, uint8x16_t, p8)
> -
> -#define TEST_TBX3(name, rettype, tbltype, idxtype, ts) \
> +#define TEST_TBX(name, rettype, tbltype, idxtype, ts) \
>    rettype test_ ## name ## _ ## ts (rettype a, tbltype b, idxtype c) \
>  	{ \
>  		return name ## _ ## ts (a, b, c); \
>  	}
>  
> -TEST_TBX3 (vqtbx3, int8x8_t, int8x16x3_t, uint8x8_t, s8)
> -TEST_TBX3 (vqtbx3, uint8x8_t, uint8x16x3_t, uint8x8_t, u8)
> -TEST_TBX3 (vqtbx3, poly8x8_t, poly8x16x3_t, uint8x8_t, p8)
> +TEST_TBX (vqtbx2, int8x8_t, int8x16x2_t, uint8x8_t, s8)
> +TEST_TBX (vqtbx2, uint8x8_t, uint8x16x2_t, uint8x8_t, u8)
> +TEST_TBX (vqtbx2, poly8x8_t, poly8x16x2_t, uint8x8_t, p8)
>  
> -TEST_TBX3 (vqtbx3q, int8x16_t, int8x16x3_t, uint8x16_t, s8)
> -TEST_TBX3 (vqtbx3q, uint8x16_t, uint8x16x3_t, uint8x16_t, u8)
> -TEST_TBX3 (vqtbx3q, poly8x16_t, poly8x16x3_t, uint8x16_t, p8)
> +TEST_TBX (vqtbx2q, int8x16_t, int8x16x2_t, uint8x16_t, s8)
> +TEST_TBX (vqtbx2q, uint8x16_t, uint8x16x2_t, uint8x16_t, u8)
> +TEST_TBX (vqtbx2q, poly8x16_t, poly8x16x2_t, uint8x16_t, p8)
>  
> -#define TEST_TBX4(name, rettype, tbltype, idxtype, ts) \
> -  rettype test_ ## name ## _ ## ts (rettype a, idxtype b, idxtype dummy1, \
> -				    idxtype dummy2, tbltype c) \
> -	{ \
> -		return name ## _ ## ts (a, c, b); \
> -	}
> +TEST_TBX (vqtbx3, int8x8_t, int8x16x3_t, uint8x8_t, s8)
> +TEST_TBX (vqtbx3, uint8x8_t, uint8x16x3_t, uint8x8_t, u8)
> +TEST_TBX (vqtbx3, poly8x8_t, poly8x16x3_t, uint8x8_t, p8)
> +
> +TEST_TBX (vqtbx3q, int8x16_t, int8x16x3_t, uint8x16_t, s8)
> +TEST_TBX (vqtbx3q, uint8x16_t, uint8x16x3_t, uint8x16_t, u8)
> +TEST_TBX (vqtbx3q, poly8x16_t, poly8x16x3_t, uint8x16_t, p8)
>  
> -TEST_TBX4 (vqtbx4, int8x8_t, int8x16x4_t, uint8x8_t, s8)
> -TEST_TBX4 (vqtbx4, uint8x8_t, uint8x16x4_t, uint8x8_t, u8)
> -TEST_TBX4 (vqtbx4, poly8x8_t, poly8x16x4_t, uint8x8_t, p8)
> +TEST_TBX (vqtbx4, int8x8_t, int8x16x4_t, uint8x8_t, s8)
> +TEST_TBX (vqtbx4, uint8x8_t, uint8x16x4_t, uint8x8_t, u8)
> +TEST_TBX (vqtbx4, poly8x8_t, poly8x16x4_t, uint8x8_t, p8)
>  
> -TEST_TBX4 (vqtbx4q, int8x16_t, int8x16x4_t, uint8x16_t, s8)
> -TEST_TBX4 (vqtbx4q, uint8x16_t, uint8x16x4_t, uint8x16_t, u8)
> -TEST_TBX4 (vqtbx4q, poly8x16_t, poly8x16x4_t, uint8x16_t, p8)
> +TEST_TBX (vqtbx4q, int8x16_t, int8x16x4_t, uint8x16_t, s8)
> +TEST_TBX (vqtbx4q, uint8x16_t, uint8x16x4_t, uint8x16_t, u8)
> +TEST_TBX (vqtbx4q, poly8x16_t, poly8x16x4_t, uint8x16_t, p8)
>  
>  #define TEST_STX(name, tbltype, ptrtype, ts) \
>    void test_ ## name ## _ ## ts (ptrtype a, tbltype b) \
> @@ -88,21 +69,21 @@ TEST_TBX4 (vqtbx4q, poly8x16_t, poly8x16x4_t, uint8x16_t, p8)
>  		name ## _ ## ts (a, b); \
>  	}
>  
> -TEST_STX (vst4q, int8x16x4_t, int8_t*, s8);
> -TEST_STX (vst4q, uint8x16x4_t, uint8_t*, u8);
> -TEST_STX (vst4q, poly8x16x4_t, poly8_t*, p8);
> -TEST_STX (vst4q, int16x8x4_t, int16_t*, s16);
> -TEST_STX (vst4q, uint16x8x4_t, uint16_t*, u16);
> -TEST_STX (vst4q, poly16x8x4_t, poly16_t*, p16);
> -TEST_STX (vst4q, float16x8x4_t, float16_t*, f16);
> -TEST_STX (vst4q, bfloat16x8x4_t, bfloat16_t*, bf16);
> -TEST_STX (vst4q, int32x4x4_t, int32_t*, s32);
> -TEST_STX (vst4q, uint32x4x4_t, uint32_t*, u32);
> -TEST_STX (vst4q, float32x4x4_t, float32_t*, f32);
> -TEST_STX (vst4q, int64x2x4_t, int64_t*, s64);
> -TEST_STX (vst4q, uint64x2x4_t, uint64_t*, u64);
> -TEST_STX (vst4q, float64x2x4_t, float64_t*, f64);
> -TEST_STX (vst4q, poly64x2x4_t, poly64_t*, p64);
> +TEST_STX (vst2, int8x8x2_t, int8_t*, s8);
> +TEST_STX (vst2, uint8x8x2_t, uint8_t*, u8);
> +TEST_STX (vst2, poly8x8x2_t, poly8_t*, p8);
> +TEST_STX (vst2, int16x4x2_t, int16_t*, s16);
> +TEST_STX (vst2, uint16x4x2_t, uint16_t*, u16);
> +TEST_STX (vst2, poly16x4x2_t, poly16_t*, p16);
> +TEST_STX (vst2, float16x4x2_t, float16_t*, f16);
> +TEST_STX (vst2, bfloat16x4x2_t, bfloat16_t*, bf16);
> +TEST_STX (vst2, int32x2x2_t, int32_t*, s32);
> +TEST_STX (vst2, uint32x2x2_t, uint32_t*, u32);
> +TEST_STX (vst2, float32x2x2_t, float32_t*, f32);
> +TEST_STX (vst2, int64x1x2_t, int64_t*, s64);
> +TEST_STX (vst2, uint64x1x2_t, uint64_t*, u64);
> +TEST_STX (vst2, float64x1x2_t, float64_t*, f64);
> +TEST_STX (vst2, poly64x1x2_t, poly64_t*, p64);
>  
>  TEST_STX (vst2q, int8x16x2_t, int8_t*, s8);
>  TEST_STX (vst2q, uint8x16x2_t, uint8_t*, u8);
> @@ -120,49 +101,193 @@ TEST_STX (vst2q, uint64x2x2_t, uint64_t*, u64);
>  TEST_STX (vst2q, float64x2x2_t, float64_t*, f64);
>  TEST_STX (vst2q, poly64x2x2_t, poly64_t*, p64);
>  
> -#define TEST_ST3(name, tbltype, ptrtype, ts) \
> -  void test_ ## name ## _ ## ts (ptrtype a, int8x8_t dummy, tbltype b) \
> +TEST_STX (vst3, int8x8x3_t, int8_t*, s8);
> +TEST_STX (vst3, uint8x8x3_t, uint8_t*, u8);
> +TEST_STX (vst3, poly8x8x3_t, poly8_t*, p8);
> +TEST_STX (vst3, int16x4x3_t, int16_t*, s16);
> +TEST_STX (vst3, uint16x4x3_t, uint16_t*, u16);
> +TEST_STX (vst3, poly16x4x3_t, poly16_t*, p16);
> +TEST_STX (vst3, float16x4x3_t, float16_t*, f16);
> +TEST_STX (vst3, bfloat16x4x3_t, bfloat16_t*, bf16);
> +TEST_STX (vst3, int32x2x3_t, int32_t*, s32);
> +TEST_STX (vst3, uint32x2x3_t, uint32_t*, u32);
> +TEST_STX (vst3, float32x2x3_t, float32_t*, f32);
> +TEST_STX (vst3, int64x1x3_t, int64_t*, s64);
> +TEST_STX (vst3, uint64x1x3_t, uint64_t*, u64);
> +TEST_STX (vst3, float64x1x3_t, float64_t*, f64);
> +TEST_STX (vst3, poly64x1x3_t, poly64_t*, p64);
> +
> +TEST_STX (vst3q, int8x16x3_t, int8_t*, s8);
> +TEST_STX (vst3q, uint8x16x3_t, uint8_t*, u8);
> +TEST_STX (vst3q, poly8x16x3_t, poly8_t*, p8);
> +TEST_STX (vst3q, int16x8x3_t, int16_t*, s16);
> +TEST_STX (vst3q, uint16x8x3_t, uint16_t*, u16);
> +TEST_STX (vst3q, poly16x8x3_t, poly16_t*, p16);
> +TEST_STX (vst3q, float16x8x3_t, float16_t*, f16);
> +TEST_STX (vst3q, bfloat16x8x3_t, bfloat16_t*, bf16);
> +TEST_STX (vst3q, int32x4x3_t, int32_t*, s32);
> +TEST_STX (vst3q, uint32x4x3_t, uint32_t*, u32);
> +TEST_STX (vst3q, float32x4x3_t, float32_t*, f32);
> +TEST_STX (vst3q, int64x2x3_t, int64_t*, s64);
> +TEST_STX (vst3q, uint64x2x3_t, uint64_t*, u64);
> +TEST_STX (vst3q, float64x2x3_t, float64_t*, f64);
> +TEST_STX (vst3q, poly64x2x3_t, poly64_t*, p64);
> +
> +TEST_STX (vst4, int8x8x4_t, int8_t*, s8);
> +TEST_STX (vst4, uint8x8x4_t, uint8_t*, u8);
> +TEST_STX (vst4, poly8x8x4_t, poly8_t*, p8);
> +TEST_STX (vst4, int16x4x4_t, int16_t*, s16);
> +TEST_STX (vst4, uint16x4x4_t, uint16_t*, u16);
> +TEST_STX (vst4, poly16x4x4_t, poly16_t*, p16);
> +TEST_STX (vst4, float16x4x4_t, float16_t*, f16);
> +TEST_STX (vst4, bfloat16x4x4_t, bfloat16_t*, bf16);
> +TEST_STX (vst4, int32x2x4_t, int32_t*, s32);
> +TEST_STX (vst4, uint32x2x4_t, uint32_t*, u32);
> +TEST_STX (vst4, float32x2x4_t, float32_t*, f32);
> +TEST_STX (vst4, int64x1x4_t, int64_t*, s64);
> +TEST_STX (vst4, uint64x1x4_t, uint64_t*, u64);
> +TEST_STX (vst4, float64x1x4_t, float64_t*, f64);
> +TEST_STX (vst4, poly64x1x4_t, poly64_t*, p64);
> +
> +TEST_STX (vst4q, int8x16x4_t, int8_t*, s8);
> +TEST_STX (vst4q, uint8x16x4_t, uint8_t*, u8);
> +TEST_STX (vst4q, poly8x16x4_t, poly8_t*, p8);
> +TEST_STX (vst4q, int16x8x4_t, int16_t*, s16);
> +TEST_STX (vst4q, uint16x8x4_t, uint16_t*, u16);
> +TEST_STX (vst4q, poly16x8x4_t, poly16_t*, p16);
> +TEST_STX (vst4q, float16x8x4_t, float16_t*, f16);
> +TEST_STX (vst4q, bfloat16x8x4_t, bfloat16_t*, bf16);
> +TEST_STX (vst4q, int32x4x4_t, int32_t*, s32);
> +TEST_STX (vst4q, uint32x4x4_t, uint32_t*, u32);
> +TEST_STX (vst4q, float32x4x4_t, float32_t*, f32);
> +TEST_STX (vst4q, int64x2x4_t, int64_t*, s64);
> +TEST_STX (vst4q, uint64x2x4_t, uint64_t*, u64);
> +TEST_STX (vst4q, float64x2x4_t, float64_t*, f64);
> +TEST_STX (vst4q, poly64x2x4_t, poly64_t*, p64);
> +
> +#define TEST_LDX(name, rettype, ptrtype, ts) \
> +  rettype test_ ## name ## _ ## ts (ptrtype a) \
>  	{ \
> -		name ## _ ## ts (a, b); \
> +		return name ## _ ## ts (a); \
>  	}
>  
> -TEST_ST3 (vst3q, int8x16x3_t, int8_t*, s8);
> -TEST_ST3 (vst3q, uint8x16x3_t, uint8_t*, u8);
> -TEST_ST3 (vst3q, poly8x16x3_t, poly8_t*, p8);
> -TEST_ST3 (vst3q, int16x8x3_t, int16_t*, s16);
> -TEST_ST3 (vst3q, uint16x8x3_t, uint16_t*, u16);
> -TEST_ST3 (vst3q, poly16x8x3_t, poly16_t*, p16);
> -TEST_ST3 (vst3q, float16x8x3_t, float16_t*, f16);
> -TEST_ST3 (vst3q, bfloat16x8x3_t, bfloat16_t*, bf16);
> -TEST_ST3 (vst3q, int32x4x3_t, int32_t*, s32);
> -TEST_ST3 (vst3q, uint32x4x3_t, uint32_t*, u32);
> -TEST_ST3 (vst3q, float32x4x3_t, float32_t*, f32);
> -TEST_ST3 (vst3q, int64x2x3_t, int64_t*, s64);
> -TEST_ST3 (vst3q, uint64x2x3_t, uint64_t*, u64);
> -TEST_ST3 (vst3q, float64x2x3_t, float64_t*, f64);
> -TEST_ST3 (vst3q, poly64x2x3_t, poly64_t*, p64);
> +TEST_LDX (vld2, int8x8x2_t, int8_t*, s8);
> +TEST_LDX (vld2, uint8x8x2_t, uint8_t*, u8);
> +TEST_LDX (vld2, poly8x8x2_t, poly8_t*, p8);
> +TEST_LDX (vld2, int16x4x2_t, int16_t*, s16);
> +TEST_LDX (vld2, uint16x4x2_t, uint16_t*, u16);
> +TEST_LDX (vld2, poly16x4x2_t, poly16_t*, p16);
> +TEST_LDX (vld2, float16x4x2_t, float16_t*, f16);
> +TEST_LDX (vld2, bfloat16x4x2_t, bfloat16_t*, bf16);
> +TEST_LDX (vld2, int32x2x2_t, int32_t*, s32);
> +TEST_LDX (vld2, uint32x2x2_t, uint32_t*, u32);
> +TEST_LDX (vld2, float32x2x2_t, float32_t*, f32);
> +TEST_LDX (vld2, int64x1x2_t, int64_t*, s64);
> +TEST_LDX (vld2, uint64x1x2_t, uint64_t*, u64);
> +TEST_LDX (vld2, float64x1x2_t, float64_t*, f64);
> +TEST_LDX (vld2, poly64x1x2_t, poly64_t*, p64);
> +
> +TEST_LDX (vld2q, int8x16x2_t, int8_t*, s8);
> +TEST_LDX (vld2q, uint8x16x2_t, uint8_t*, u8);
> +TEST_LDX (vld2q, poly8x16x2_t, poly8_t*, p8);
> +TEST_LDX (vld2q, int16x8x2_t, int16_t*, s16);
> +TEST_LDX (vld2q, uint16x8x2_t, uint16_t*, u16);
> +TEST_LDX (vld2q, poly16x8x2_t, poly16_t*, p16);
> +TEST_LDX (vld2q, float16x8x2_t, float16_t*, f16);
> +TEST_LDX (vld2q, bfloat16x8x2_t, bfloat16_t*, bf16);
> +TEST_LDX (vld2q, int32x4x2_t, int32_t*, s32);
> +TEST_LDX (vld2q, uint32x4x2_t, uint32_t*, u32);
> +TEST_LDX (vld2q, float32x4x2_t, float32_t*, f32);
> +TEST_LDX (vld2q, int64x2x2_t, int64_t*, s64);
> +TEST_LDX (vld2q, uint64x2x2_t, uint64_t*, u64);
> +TEST_LDX (vld2q, float64x2x2_t, float64_t*, f64);
> +TEST_LDX (vld2q, poly64x2x2_t, poly64_t*, p64);
> +
> +TEST_LDX (vld3, int8x8x3_t, int8_t*, s8);
> +TEST_LDX (vld3, uint8x8x3_t, uint8_t*, u8);
> +TEST_LDX (vld3, poly8x8x3_t, poly8_t*, p8);
> +TEST_LDX (vld3, int16x4x3_t, int16_t*, s16);
> +TEST_LDX (vld3, uint16x4x3_t, uint16_t*, u16);
> +TEST_LDX (vld3, poly16x4x3_t, poly16_t*, p16);
> +TEST_LDX (vld3, float16x4x3_t, float16_t*, f16);
> +TEST_LDX (vld3, bfloat16x4x3_t, bfloat16_t*, bf16);
> +TEST_LDX (vld3, int32x2x3_t, int32_t*, s32);
> +TEST_LDX (vld3, uint32x2x3_t, uint32_t*, u32);
> +TEST_LDX (vld3, float32x2x3_t, float32_t*, f32);
> +TEST_LDX (vld3, int64x1x3_t, int64_t*, s64);
> +TEST_LDX (vld3, uint64x1x3_t, uint64_t*, u64);
> +TEST_LDX (vld3, float64x1x3_t, float64_t*, f64);
> +TEST_LDX (vld3, poly64x1x3_t, poly64_t*, p64);
> +
> +TEST_LDX (vld3q, int8x16x3_t, int8_t*, s8);
> +TEST_LDX (vld3q, uint8x16x3_t, uint8_t*, u8);
> +TEST_LDX (vld3q, poly8x16x3_t, poly8_t*, p8);
> +TEST_LDX (vld3q, int16x8x3_t, int16_t*, s16);
> +TEST_LDX (vld3q, uint16x8x3_t, uint16_t*, u16);
> +TEST_LDX (vld3q, poly16x8x3_t, poly16_t*, p16);
> +TEST_LDX (vld3q, float16x8x3_t, float16_t*, f16);
> +TEST_LDX (vld3q, bfloat16x8x3_t, bfloat16_t*, bf16);
> +TEST_LDX (vld3q, int32x4x3_t, int32_t*, s32);
> +TEST_LDX (vld3q, uint32x4x3_t, uint32_t*, u32);
> +TEST_LDX (vld3q, float32x4x3_t, float32_t*, f32);
> +TEST_LDX (vld3q, int64x2x3_t, int64_t*, s64);
> +TEST_LDX (vld3q, uint64x2x3_t, uint64_t*, u64);
> +TEST_LDX (vld3q, float64x2x3_t, float64_t*, f64);
> +TEST_LDX (vld3q, poly64x2x3_t, poly64_t*, p64);
> +
> +TEST_LDX (vld4, int8x8x4_t, int8_t*, s8);
> +TEST_LDX (vld4, uint8x8x4_t, uint8_t*, u8);
> +TEST_LDX (vld4, poly8x8x4_t, poly8_t*, p8);
> +TEST_LDX (vld4, int16x4x4_t, int16_t*, s16);
> +TEST_LDX (vld4, uint16x4x4_t, uint16_t*, u16);
> +TEST_LDX (vld4, poly16x4x4_t, poly16_t*, p16);
> +TEST_LDX (vld4, float16x4x4_t, float16_t*, f16);
> +TEST_LDX (vld4, bfloat16x4x4_t, bfloat16_t*, bf16);
> +TEST_LDX (vld4, int32x2x4_t, int32_t*, s32);
> +TEST_LDX (vld4, uint32x2x4_t, uint32_t*, u32);
> +TEST_LDX (vld4, float32x2x4_t, float32_t*, f32);
> +TEST_LDX (vld4, int64x1x4_t, int64_t*, s64);
> +TEST_LDX (vld4, uint64x1x4_t, uint64_t*, u64);
> +TEST_LDX (vld4, float64x1x4_t, float64_t*, f64);
> +TEST_LDX (vld4, poly64x1x4_t, poly64_t*, p64);
> +
> +TEST_LDX (vld4q, int8x16x4_t, int8_t*, s8);
> +TEST_LDX (vld4q, uint8x16x4_t, uint8_t*, u8);
> +TEST_LDX (vld4q, poly8x16x4_t, poly8_t*, p8);
> +TEST_LDX (vld4q, int16x8x4_t, int16_t*, s16);
> +TEST_LDX (vld4q, uint16x8x4_t, uint16_t*, u16);
> +TEST_LDX (vld4q, poly16x8x4_t, poly16_t*, p16);
> +TEST_LDX (vld4q, float16x8x4_t, float16_t*, f16);
> +TEST_LDX (vld4q, bfloat16x8x4_t, bfloat16_t*, bf16);
> +TEST_LDX (vld4q, int32x4x4_t, int32_t*, s32);
> +TEST_LDX (vld4q, uint32x4x4_t, uint32_t*, u32);
> +TEST_LDX (vld4q, float32x4x4_t, float32_t*, f32);
> +TEST_LDX (vld4q, int64x2x4_t, int64_t*, s64);
> +TEST_LDX (vld4q, uint64x2x4_t, uint64_t*, u64);
> +TEST_LDX (vld4q, float64x2x4_t, float64_t*, f64);
> +TEST_LDX (vld4q, poly64x2x4_t, poly64_t*, p64);
>  
>  #define TEST_STX_LANE(name, tbltype, ptrtype, ts) \
>    void test_ ## name ## _ ## ts (ptrtype a, tbltype b) \
>  	{ \
> -		name ## _ ## ts (a, b, 1); \
> +		name ## _ ## ts (a, b, 0); \
>  	}
>  
> -TEST_STX_LANE (vst4q_lane, int8x16x4_t, int8_t*, s8);
> -TEST_STX_LANE (vst4q_lane, uint8x16x4_t, uint8_t*, u8);
> -TEST_STX_LANE (vst4q_lane, poly8x16x4_t, poly8_t*, p8);
> -TEST_STX_LANE (vst4q_lane, int16x8x4_t, int16_t*, s16);
> -TEST_STX_LANE (vst4q_lane, uint16x8x4_t, uint16_t*, u16);
> -TEST_STX_LANE (vst4q_lane, poly16x8x4_t, poly16_t*, p16);
> -TEST_STX_LANE (vst4q_lane, float16x8x4_t, float16_t*, f16);
> -TEST_STX_LANE (vst4q_lane, bfloat16x8x4_t, bfloat16_t*, bf16);
> -TEST_STX_LANE (vst4q_lane, int32x4x4_t, int32_t*, s32);
> -TEST_STX_LANE (vst4q_lane, uint32x4x4_t, uint32_t*, u32);
> -TEST_STX_LANE (vst4q_lane, float32x4x4_t, float32_t*, f32);
> -TEST_STX_LANE (vst4q_lane, int64x2x4_t, int64_t*, s64);
> -TEST_STX_LANE (vst4q_lane, uint64x2x4_t, uint64_t*, u64);
> -TEST_STX_LANE (vst4q_lane, float64x2x4_t, float64_t*, f64);
> -TEST_STX_LANE (vst4q_lane, poly64x2x4_t, poly64_t*, p64);
> +TEST_STX_LANE (vst2_lane, int8x8x2_t, int8_t*, s8);
> +TEST_STX_LANE (vst2_lane, uint8x8x2_t, uint8_t*, u8);
> +TEST_STX_LANE (vst2_lane, poly8x8x2_t, poly8_t*, p8);
> +TEST_STX_LANE (vst2_lane, int16x4x2_t, int16_t*, s16);
> +TEST_STX_LANE (vst2_lane, uint16x4x2_t, uint16_t*, u16);
> +TEST_STX_LANE (vst2_lane, poly16x4x2_t, poly16_t*, p16);
> +TEST_STX_LANE (vst2_lane, float16x4x2_t, float16_t*, f16);
> +TEST_STX_LANE (vst2_lane, bfloat16x4x2_t, bfloat16_t*, bf16);
> +TEST_STX_LANE (vst2_lane, int32x2x2_t, int32_t*, s32);
> +TEST_STX_LANE (vst2_lane, uint32x2x2_t, uint32_t*, u32);
> +TEST_STX_LANE (vst2_lane, float32x2x2_t, float32_t*, f32);
> +TEST_STX_LANE (vst2_lane, int64x1x2_t, int64_t*, s64);
> +TEST_STX_LANE (vst2_lane, uint64x1x2_t, uint64_t*, u64);
> +TEST_STX_LANE (vst2_lane, float64x1x2_t, float64_t*, f64);
> +TEST_STX_LANE (vst2_lane, poly64x1x2_t, poly64_t*, p64);
>  
>  TEST_STX_LANE (vst2q_lane, int8x16x2_t, int8_t*, s8);
>  TEST_STX_LANE (vst2q_lane, uint8x16x2_t, uint8_t*, u8);
> @@ -180,27 +305,171 @@ TEST_STX_LANE (vst2q_lane, uint64x2x2_t, uint64_t*, u64);
>  TEST_STX_LANE (vst2q_lane, float64x2x2_t, float64_t*, f64);
>  TEST_STX_LANE (vst2q_lane, poly64x2x2_t, poly64_t*, p64);
>  
> -#define TEST_ST3_LANE(name, tbltype, ptrtype, ts) \
> -  void test_ ## name ## _ ## ts (ptrtype a, int8x8_t dummy, tbltype b) \
> +TEST_STX_LANE (vst3_lane, int8x8x3_t, int8_t*, s8);
> +TEST_STX_LANE (vst3_lane, uint8x8x3_t, uint8_t*, u8);
> +TEST_STX_LANE (vst3_lane, poly8x8x3_t, poly8_t*, p8);
> +TEST_STX_LANE (vst3_lane, int16x4x3_t, int16_t*, s16);
> +TEST_STX_LANE (vst3_lane, uint16x4x3_t, uint16_t*, u16);
> +TEST_STX_LANE (vst3_lane, poly16x4x3_t, poly16_t*, p16);
> +TEST_STX_LANE (vst3_lane, float16x4x3_t, float16_t*, f16);
> +TEST_STX_LANE (vst3_lane, bfloat16x4x3_t, bfloat16_t*, bf16);
> +TEST_STX_LANE (vst3_lane, int32x2x3_t, int32_t*, s32);
> +TEST_STX_LANE (vst3_lane, uint32x2x3_t, uint32_t*, u32);
> +TEST_STX_LANE (vst3_lane, float32x2x3_t, float32_t*, f32);
> +TEST_STX_LANE (vst3_lane, int64x1x3_t, int64_t*, s64);
> +TEST_STX_LANE (vst3_lane, uint64x1x3_t, uint64_t*, u64);
> +TEST_STX_LANE (vst3_lane, float64x1x3_t, float64_t*, f64);
> +TEST_STX_LANE (vst3_lane, poly64x1x3_t, poly64_t*, p64);
> +
> +TEST_STX_LANE (vst3q_lane, int8x16x3_t, int8_t*, s8);
> +TEST_STX_LANE (vst3q_lane, uint8x16x3_t, uint8_t*, u8);
> +TEST_STX_LANE (vst3q_lane, poly8x16x3_t, poly8_t*, p8);
> +TEST_STX_LANE (vst3q_lane, int16x8x3_t, int16_t*, s16);
> +TEST_STX_LANE (vst3q_lane, uint16x8x3_t, uint16_t*, u16);
> +TEST_STX_LANE (vst3q_lane, poly16x8x3_t, poly16_t*, p16);
> +TEST_STX_LANE (vst3q_lane, float16x8x3_t, float16_t*, f16);
> +TEST_STX_LANE (vst3q_lane, bfloat16x8x3_t, bfloat16_t*, bf16);
> +TEST_STX_LANE (vst3q_lane, int32x4x3_t, int32_t*, s32);
> +TEST_STX_LANE (vst3q_lane, uint32x4x3_t, uint32_t*, u32);
> +TEST_STX_LANE (vst3q_lane, float32x4x3_t, float32_t*, f32);
> +TEST_STX_LANE (vst3q_lane, int64x2x3_t, int64_t*, s64);
> +TEST_STX_LANE (vst3q_lane, uint64x2x3_t, uint64_t*, u64);
> +TEST_STX_LANE (vst3q_lane, float64x2x3_t, float64_t*, f64);
> +TEST_STX_LANE (vst3q_lane, poly64x2x3_t, poly64_t*, p64);
> +
> +TEST_STX_LANE (vst4_lane, int8x8x4_t, int8_t*, s8);
> +TEST_STX_LANE (vst4_lane, uint8x8x4_t, uint8_t*, u8);
> +TEST_STX_LANE (vst4_lane, poly8x8x4_t, poly8_t*, p8);
> +TEST_STX_LANE (vst4_lane, int16x4x4_t, int16_t*, s16);
> +TEST_STX_LANE (vst4_lane, uint16x4x4_t, uint16_t*, u16);
> +TEST_STX_LANE (vst4_lane, poly16x4x4_t, poly16_t*, p16);
> +TEST_STX_LANE (vst4_lane, float16x4x4_t, float16_t*, f16);
> +TEST_STX_LANE (vst4_lane, bfloat16x4x4_t, bfloat16_t*, bf16);
> +TEST_STX_LANE (vst4_lane, int32x2x4_t, int32_t*, s32);
> +TEST_STX_LANE (vst4_lane, uint32x2x4_t, uint32_t*, u32);
> +TEST_STX_LANE (vst4_lane, float32x2x4_t, float32_t*, f32);
> +TEST_STX_LANE (vst4_lane, int64x1x4_t, int64_t*, s64);
> +TEST_STX_LANE (vst4_lane, uint64x1x4_t, uint64_t*, u64);
> +TEST_STX_LANE (vst4_lane, float64x1x4_t, float64_t*, f64);
> +TEST_STX_LANE (vst4_lane, poly64x1x4_t, poly64_t*, p64);
> +
> +TEST_STX_LANE (vst4q_lane, int8x16x4_t, int8_t*, s8);
> +TEST_STX_LANE (vst4q_lane, uint8x16x4_t, uint8_t*, u8);
> +TEST_STX_LANE (vst4q_lane, poly8x16x4_t, poly8_t*, p8);
> +TEST_STX_LANE (vst4q_lane, int16x8x4_t, int16_t*, s16);
> +TEST_STX_LANE (vst4q_lane, uint16x8x4_t, uint16_t*, u16);
> +TEST_STX_LANE (vst4q_lane, poly16x8x4_t, poly16_t*, p16);
> +TEST_STX_LANE (vst4q_lane, float16x8x4_t, float16_t*, f16);
> +TEST_STX_LANE (vst4q_lane, bfloat16x8x4_t, bfloat16_t*, bf16);
> +TEST_STX_LANE (vst4q_lane, int32x4x4_t, int32_t*, s32);
> +TEST_STX_LANE (vst4q_lane, uint32x4x4_t, uint32_t*, u32);
> +TEST_STX_LANE (vst4q_lane, float32x4x4_t, float32_t*, f32);
> +TEST_STX_LANE (vst4q_lane, int64x2x4_t, int64_t*, s64);
> +TEST_STX_LANE (vst4q_lane, uint64x2x4_t, uint64_t*, u64);
> +TEST_STX_LANE (vst4q_lane, float64x2x4_t, float64_t*, f64);
> +TEST_STX_LANE (vst4q_lane, poly64x2x4_t, poly64_t*, p64);
> +
> +#define TEST_LDX_LANE(name, rettype, ptrtype, ts) \
> +  rettype test_ ## name ## _ ## ts (ptrtype a, rettype b) \
>  	{ \
> -		name ## _ ## ts (a, b, 1); \
> +		return name ## _ ## ts (a, b, 0); \
>  	}
>  
> -TEST_ST3_LANE (vst3q_lane, int8x16x3_t, int8_t*, s8);
> -TEST_ST3_LANE (vst3q_lane, uint8x16x3_t, uint8_t*, u8);
> -TEST_ST3_LANE (vst3q_lane, poly8x16x3_t, poly8_t*, p8);
> -TEST_ST3_LANE (vst3q_lane, int16x8x3_t, int16_t*, s16);
> -TEST_ST3_LANE (vst3q_lane, uint16x8x3_t, uint16_t*, u16);
> -TEST_ST3_LANE (vst3q_lane, poly16x8x3_t, poly16_t*, p16);
> -TEST_ST3_LANE (vst3q_lane, float16x8x3_t, float16_t*, f16);
> -TEST_ST3_LANE (vst3q_lane, bfloat16x8x3_t, bfloat16_t*, bf16);
> -TEST_ST3_LANE (vst3q_lane, int32x4x3_t, int32_t*, s32);
> -TEST_ST3_LANE (vst3q_lane, uint32x4x3_t, uint32_t*, u32);
> -TEST_ST3_LANE (vst3q_lane, float32x4x3_t, float32_t*, f32);
> -TEST_ST3_LANE (vst3q_lane, int64x2x3_t, int64_t*, s64);
> -TEST_ST3_LANE (vst3q_lane, uint64x2x3_t, uint64_t*, u64);
> -TEST_ST3_LANE (vst3q_lane, float64x2x3_t, float64_t*, f64);
> -TEST_ST3_LANE (vst3q_lane, poly64x2x3_t, poly64_t*, p64);
> +TEST_LDX_LANE (vld2_lane, int8x8x2_t, int8_t*, s8);
> +TEST_LDX_LANE (vld2_lane, uint8x8x2_t, uint8_t*, u8);
> +TEST_LDX_LANE (vld2_lane, poly8x8x2_t, poly8_t*, p8);
> +TEST_LDX_LANE (vld2_lane, int16x4x2_t, int16_t*, s16);
> +TEST_LDX_LANE (vld2_lane, uint16x4x2_t, uint16_t*, u16);
> +TEST_LDX_LANE (vld2_lane, poly16x4x2_t, poly16_t*, p16);
> +TEST_LDX_LANE (vld2_lane, float16x4x2_t, float16_t*, f16);
> +TEST_LDX_LANE (vld2_lane, bfloat16x4x2_t, bfloat16_t*, bf16);
> +TEST_LDX_LANE (vld2_lane, int32x2x2_t, int32_t*, s32);
> +TEST_LDX_LANE (vld2_lane, uint32x2x2_t, uint32_t*, u32);
> +TEST_LDX_LANE (vld2_lane, float32x2x2_t, float32_t*, f32);
> +TEST_LDX_LANE (vld2_lane, int64x1x2_t, int64_t*, s64);
> +TEST_LDX_LANE (vld2_lane, uint64x1x2_t, uint64_t*, u64);
> +TEST_LDX_LANE (vld2_lane, float64x1x2_t, float64_t*, f64);
> +TEST_LDX_LANE (vld2_lane, poly64x1x2_t, poly64_t*, p64);
> +
> +TEST_LDX_LANE (vld2q_lane, int8x16x2_t, int8_t*, s8);
> +TEST_LDX_LANE (vld2q_lane, uint8x16x2_t, uint8_t*, u8);
> +TEST_LDX_LANE (vld2q_lane, poly8x16x2_t, poly8_t*, p8);
> +TEST_LDX_LANE (vld2q_lane, int16x8x2_t, int16_t*, s16);
> +TEST_LDX_LANE (vld2q_lane, uint16x8x2_t, uint16_t*, u16);
> +TEST_LDX_LANE (vld2q_lane, poly16x8x2_t, poly16_t*, p16);
> +TEST_LDX_LANE (vld2q_lane, float16x8x2_t, float16_t*, f16);
> +TEST_LDX_LANE (vld2q_lane, bfloat16x8x2_t, bfloat16_t*, bf16);
> +TEST_LDX_LANE (vld2q_lane, int32x4x2_t, int32_t*, s32);
> +TEST_LDX_LANE (vld2q_lane, uint32x4x2_t, uint32_t*, u32);
> +TEST_LDX_LANE (vld2q_lane, float32x4x2_t, float32_t*, f32);
> +TEST_LDX_LANE (vld2q_lane, int64x2x2_t, int64_t*, s64);
> +TEST_LDX_LANE (vld2q_lane, uint64x2x2_t, uint64_t*, u64);
> +TEST_LDX_LANE (vld2q_lane, float64x2x2_t, float64_t*, f64);
> +TEST_LDX_LANE (vld2q_lane, poly64x2x2_t, poly64_t*, p64);
> +
> +TEST_LDX_LANE (vld3_lane, int8x8x3_t, int8_t*, s8);
> +TEST_LDX_LANE (vld3_lane, uint8x8x3_t, uint8_t*, u8);
> +TEST_LDX_LANE (vld3_lane, poly8x8x3_t, poly8_t*, p8);
> +TEST_LDX_LANE (vld3_lane, int16x4x3_t, int16_t*, s16);
> +TEST_LDX_LANE (vld3_lane, uint16x4x3_t, uint16_t*, u16);
> +TEST_LDX_LANE (vld3_lane, poly16x4x3_t, poly16_t*, p16);
> +TEST_LDX_LANE (vld3_lane, float16x4x3_t, float16_t*, f16);
> +TEST_LDX_LANE (vld3_lane, bfloat16x4x3_t, bfloat16_t*, bf16);
> +TEST_LDX_LANE (vld3_lane, int32x2x3_t, int32_t*, s32);
> +TEST_LDX_LANE (vld3_lane, uint32x2x3_t, uint32_t*, u32);
> +TEST_LDX_LANE (vld3_lane, float32x2x3_t, float32_t*, f32);
> +TEST_LDX_LANE (vld3_lane, int64x1x3_t, int64_t*, s64);
> +TEST_LDX_LANE (vld3_lane, uint64x1x3_t, uint64_t*, u64);
> +TEST_LDX_LANE (vld3_lane, float64x1x3_t, float64_t*, f64);
> +TEST_LDX_LANE (vld3_lane, poly64x1x3_t, poly64_t*, p64);
> +
> +TEST_LDX_LANE (vld3q_lane, int8x16x3_t, int8_t*, s8);
> +TEST_LDX_LANE (vld3q_lane, uint8x16x3_t, uint8_t*, u8);
> +TEST_LDX_LANE (vld3q_lane, poly8x16x3_t, poly8_t*, p8);
> +TEST_LDX_LANE (vld3q_lane, int16x8x3_t, int16_t*, s16);
> +TEST_LDX_LANE (vld3q_lane, uint16x8x3_t, uint16_t*, u16);
> +TEST_LDX_LANE (vld3q_lane, poly16x8x3_t, poly16_t*, p16);
> +TEST_LDX_LANE (vld3q_lane, float16x8x3_t, float16_t*, f16);
> +TEST_LDX_LANE (vld3q_lane, bfloat16x8x3_t, bfloat16_t*, bf16);
> +TEST_LDX_LANE (vld3q_lane, int32x4x3_t, int32_t*, s32);
> +TEST_LDX_LANE (vld3q_lane, uint32x4x3_t, uint32_t*, u32);
> +TEST_LDX_LANE (vld3q_lane, float32x4x3_t, float32_t*, f32);
> +TEST_LDX_LANE (vld3q_lane, int64x2x3_t, int64_t*, s64);
> +TEST_LDX_LANE (vld3q_lane, uint64x2x3_t, uint64_t*, u64);
> +TEST_LDX_LANE (vld3q_lane, float64x2x3_t, float64_t*, f64);
> +TEST_LDX_LANE (vld3q_lane, poly64x2x3_t, poly64_t*, p64);
> +
> +TEST_LDX_LANE (vld4_lane, int8x8x4_t, int8_t*, s8);
> +TEST_LDX_LANE (vld4_lane, uint8x8x4_t, uint8_t*, u8);
> +TEST_LDX_LANE (vld4_lane, poly8x8x4_t, poly8_t*, p8);
> +TEST_LDX_LANE (vld4_lane, int16x4x4_t, int16_t*, s16);
> +TEST_LDX_LANE (vld4_lane, uint16x4x4_t, uint16_t*, u16);
> +TEST_LDX_LANE (vld4_lane, poly16x4x4_t, poly16_t*, p16);
> +TEST_LDX_LANE (vld4_lane, float16x4x4_t, float16_t*, f16);
> +TEST_LDX_LANE (vld4_lane, bfloat16x4x4_t, bfloat16_t*, bf16);
> +TEST_LDX_LANE (vld4_lane, int32x2x4_t, int32_t*, s32);
> +TEST_LDX_LANE (vld4_lane, uint32x2x4_t, uint32_t*, u32);
> +TEST_LDX_LANE (vld4_lane, float32x2x4_t, float32_t*, f32);
> +TEST_LDX_LANE (vld4_lane, int64x1x4_t, int64_t*, s64);
> +TEST_LDX_LANE (vld4_lane, uint64x1x4_t, uint64_t*, u64);
> +TEST_LDX_LANE (vld4_lane, float64x1x4_t, float64_t*, f64);
> +TEST_LDX_LANE (vld4_lane, poly64x1x4_t, poly64_t*, p64);
> +
> +TEST_LDX_LANE (vld4q_lane, int8x16x4_t, int8_t*, s8);
> +TEST_LDX_LANE (vld4q_lane, uint8x16x4_t, uint8_t*, u8);
> +TEST_LDX_LANE (vld4q_lane, poly8x16x4_t, poly8_t*, p8);
> +TEST_LDX_LANE (vld4q_lane, int16x8x4_t, int16_t*, s16);
> +TEST_LDX_LANE (vld4q_lane, uint16x8x4_t, uint16_t*, u16);
> +TEST_LDX_LANE (vld4q_lane, poly16x8x4_t, poly16_t*, p16);
> +TEST_LDX_LANE (vld4q_lane, float16x8x4_t, float16_t*, f16);
> +TEST_LDX_LANE (vld4q_lane, bfloat16x8x4_t, bfloat16_t*, bf16);
> +TEST_LDX_LANE (vld4q_lane, int32x4x4_t, int32_t*, s32);
> +TEST_LDX_LANE (vld4q_lane, uint32x4x4_t, uint32_t*, u32);
> +TEST_LDX_LANE (vld4q_lane, float32x4x4_t, float32_t*, f32);
> +TEST_LDX_LANE (vld4q_lane, int64x2x4_t, int64_t*, s64);
> +TEST_LDX_LANE (vld4q_lane, uint64x2x4_t, uint64_t*, u64);
> +TEST_LDX_LANE (vld4q_lane, float64x2x4_t, float64_t*, f64);
> +TEST_LDX_LANE (vld4q_lane, poly64x2x4_t, poly64_t*, p64);
>  
>  #define TEST_ST1xN(name, tbltype, ptrtype, ts, xn) \
>    void test_ ## name ## _ ## ts ## _ ## xn (ptrtype a, tbltype b) \
> @@ -208,21 +477,21 @@ TEST_ST3_LANE (vst3q_lane, poly64x2x3_t, poly64_t*, p64);
>  		name ## _ ## ts ## _ ## xn (a, b); \
>  	}
>  
> -TEST_ST1xN (vst1q, int8x16x4_t, int8_t*, s8, x4);
> -TEST_ST1xN (vst1q, uint8x16x4_t, uint8_t*, u8, x4);
> -TEST_ST1xN (vst1q, poly8x16x4_t, poly8_t*, p8, x4);
> -TEST_ST1xN (vst1q, int16x8x4_t, int16_t*, s16, x4);
> -TEST_ST1xN (vst1q, uint16x8x4_t, uint16_t*, u16, x4);
> -TEST_ST1xN (vst1q, poly16x8x4_t, poly16_t*, p16, x4);
> -TEST_ST1xN (vst1q, float16x8x4_t, float16_t*, f16, x4);
> -TEST_ST1xN (vst1q, bfloat16x8x4_t, bfloat16_t*, bf16, x4);
> -TEST_ST1xN (vst1q, int32x4x4_t, int32_t*, s32, x4);
> -TEST_ST1xN (vst1q, uint32x4x4_t, uint32_t*, u32, x4);
> -TEST_ST1xN (vst1q, float32x4x4_t, float32_t*, f32, x4);
> -TEST_ST1xN (vst1q, int64x2x4_t, int64_t*, s64, x4);
> -TEST_ST1xN (vst1q, uint64x2x4_t, uint64_t*, u64, x4);
> -TEST_ST1xN (vst1q, poly64x2x4_t, poly64_t*, p64, x4);
> -TEST_ST1xN (vst1q, float64x2x4_t, float64_t*, f64, x4);
> +TEST_ST1xN (vst1, int8x8x2_t, int8_t*, s8, x2);
> +TEST_ST1xN (vst1, uint8x8x2_t, uint8_t*, u8, x2);
> +TEST_ST1xN (vst1, poly8x8x2_t, poly8_t*, p8, x2);
> +TEST_ST1xN (vst1, int16x4x2_t, int16_t*, s16, x2);
> +TEST_ST1xN (vst1, uint16x4x2_t, uint16_t*, u16, x2);
> +TEST_ST1xN (vst1, poly16x4x2_t, poly16_t*, p16, x2);
> +TEST_ST1xN (vst1, float16x4x2_t, float16_t*, f16, x2);
> +TEST_ST1xN (vst1, bfloat16x4x2_t, bfloat16_t*, bf16, x2);
> +TEST_ST1xN (vst1, int32x2x2_t, int32_t*, s32, x2);
> +TEST_ST1xN (vst1, uint32x2x2_t, uint32_t*, u32, x2);
> +TEST_ST1xN (vst1, float32x2x2_t, float32_t*, f32, x2);
> +TEST_ST1xN (vst1, int64x1x2_t, int64_t*, s64, x2);
> +TEST_ST1xN (vst1, uint64x1x2_t, uint64_t*, u64, x2);
> +TEST_ST1xN (vst1, poly64x1x2_t, poly64_t*, p64, x2);
> +TEST_ST1xN (vst1, float64x1x2_t, float64_t*, f64, x2);
>  
>  TEST_ST1xN (vst1q, int8x16x2_t, int8_t*, s8, x2);
>  TEST_ST1xN (vst1q, uint8x16x2_t, uint8_t*, u8, x2);
> @@ -240,34 +509,184 @@ TEST_ST1xN (vst1q, uint64x2x2_t, uint64_t*, u64, x2);
>  TEST_ST1xN (vst1q, poly64x2x2_t, poly64_t*, p64, x2);
>  TEST_ST1xN (vst1q, float64x2x2_t, float64_t*, f64, x2);
>  
> -#define TEST_ST1x3(name, tbltype, ptrtype, ts, xn) \
> -  void test_ ## name ## _ ## ts ## _ ## xn (ptrtype a, int8x8_t dummy, \
> -					    tbltype b) \
> +TEST_ST1xN (vst1, int8x8x3_t, int8_t*, s8, x3);
> +TEST_ST1xN (vst1, uint8x8x3_t, uint8_t*, u8, x3);
> +TEST_ST1xN (vst1, poly8x8x3_t, poly8_t*, p8, x3);
> +TEST_ST1xN (vst1, int16x4x3_t, int16_t*, s16, x3);
> +TEST_ST1xN (vst1, uint16x4x3_t, uint16_t*, u16, x3);
> +TEST_ST1xN (vst1, poly16x4x3_t, poly16_t*, p16, x3);
> +TEST_ST1xN (vst1, float16x4x3_t, float16_t*, f16, x3);
> +TEST_ST1xN (vst1, bfloat16x4x3_t, bfloat16_t*, bf16, x3);
> +TEST_ST1xN (vst1, int32x2x3_t, int32_t*, s32, x3);
> +TEST_ST1xN (vst1, uint32x2x3_t, uint32_t*, u32, x3);
> +TEST_ST1xN (vst1, float32x2x3_t, float32_t*, f32, x3);
> +TEST_ST1xN (vst1, int64x1x3_t, int64_t*, s64, x3);
> +TEST_ST1xN (vst1, uint64x1x3_t, uint64_t*, u64, x3);
> +TEST_ST1xN (vst1, poly64x1x3_t, poly64_t*, p64, x3);
> +TEST_ST1xN (vst1, float64x1x3_t, float64_t*, f64, x3);
> +
> +TEST_ST1xN (vst1q, int8x16x3_t, int8_t*, s8, x3);
> +TEST_ST1xN (vst1q, uint8x16x3_t, uint8_t*, u8, x3);
> +TEST_ST1xN (vst1q, poly8x16x3_t, poly8_t*, p8, x3);
> +TEST_ST1xN (vst1q, int16x8x3_t, int16_t*, s16, x3);
> +TEST_ST1xN (vst1q, uint16x8x3_t, uint16_t*, u16, x3);
> +TEST_ST1xN (vst1q, poly16x8x3_t, poly16_t*, p16, x3);
> +TEST_ST1xN (vst1q, float16x8x3_t, float16_t*, f16, x3);
> +TEST_ST1xN (vst1q, bfloat16x8x3_t, bfloat16_t*, bf16, x3);
> +TEST_ST1xN (vst1q, int32x4x3_t, int32_t*, s32, x3);
> +TEST_ST1xN (vst1q, uint32x4x3_t, uint32_t*, u32, x3);
> +TEST_ST1xN (vst1q, float32x4x3_t, float32_t*, f32, x3);
> +TEST_ST1xN (vst1q, int64x2x3_t, int64_t*, s64, x3);
> +TEST_ST1xN (vst1q, uint64x2x3_t, uint64_t*, u64, x3);
> +TEST_ST1xN (vst1q, poly64x2x3_t, poly64_t*, p64, x3);
> +TEST_ST1xN (vst1q, float64x2x3_t, float64_t*, f64, x3);
> +
> +TEST_ST1xN (vst1, int8x8x4_t, int8_t*, s8, x4);
> +TEST_ST1xN (vst1, uint8x8x4_t, uint8_t*, u8, x4);
> +TEST_ST1xN (vst1, poly8x8x4_t, poly8_t*, p8, x4);
> +TEST_ST1xN (vst1, int16x4x4_t, int16_t*, s16, x4);
> +TEST_ST1xN (vst1, uint16x4x4_t, uint16_t*, u16, x4);
> +TEST_ST1xN (vst1, poly16x4x4_t, poly16_t*, p16, x4);
> +TEST_ST1xN (vst1, float16x4x4_t, float16_t*, f16, x4);
> +TEST_ST1xN (vst1, bfloat16x4x4_t, bfloat16_t*, bf16, x4);
> +TEST_ST1xN (vst1, int32x2x4_t, int32_t*, s32, x4);
> +TEST_ST1xN (vst1, uint32x2x4_t, uint32_t*, u32, x4);
> +TEST_ST1xN (vst1, float32x2x4_t, float32_t*, f32, x4);
> +TEST_ST1xN (vst1, int64x1x4_t, int64_t*, s64, x4);
> +TEST_ST1xN (vst1, uint64x1x4_t, uint64_t*, u64, x4);
> +TEST_ST1xN (vst1, poly64x1x4_t, poly64_t*, p64, x4);
> +TEST_ST1xN (vst1, float64x1x4_t, float64_t*, f64, x4);
> +
> +TEST_ST1xN (vst1q, int8x16x4_t, int8_t*, s8, x4);
> +TEST_ST1xN (vst1q, uint8x16x4_t, uint8_t*, u8, x4);
> +TEST_ST1xN (vst1q, poly8x16x4_t, poly8_t*, p8, x4);
> +TEST_ST1xN (vst1q, int16x8x4_t, int16_t*, s16, x4);
> +TEST_ST1xN (vst1q, uint16x8x4_t, uint16_t*, u16, x4);
> +TEST_ST1xN (vst1q, poly16x8x4_t, poly16_t*, p16, x4);
> +TEST_ST1xN (vst1q, float16x8x4_t, float16_t*, f16, x4);
> +TEST_ST1xN (vst1q, bfloat16x8x4_t, bfloat16_t*, bf16, x4);
> +TEST_ST1xN (vst1q, int32x4x4_t, int32_t*, s32, x4);
> +TEST_ST1xN (vst1q, uint32x4x4_t, uint32_t*, u32, x4);
> +TEST_ST1xN (vst1q, float32x4x4_t, float32_t*, f32, x4);
> +TEST_ST1xN (vst1q, int64x2x4_t, int64_t*, s64, x4);
> +TEST_ST1xN (vst1q, uint64x2x4_t, uint64_t*, u64, x4);
> +TEST_ST1xN (vst1q, poly64x2x4_t, poly64_t*, p64, x4);
> +TEST_ST1xN (vst1q, float64x2x4_t, float64_t*, f64, x4);
> +
> +#define TEST_LD1xN(name, rettype, ptrtype, ts, xn) \
> +  rettype test_ ## name ## _ ## ts ## _ ## xn (ptrtype a) \
>  	{ \
> -		name ## _ ## ts ## _ ## xn (a, b); \
> +		return name ## _ ## ts ## _ ## xn (a); \
>  	}
>  
> -TEST_ST1x3 (vst1q, int8x16x3_t, int8_t*, s8, x3);
> -TEST_ST1x3 (vst1q, uint8x16x3_t, uint8_t*, u8, x3);
> -TEST_ST1x3 (vst1q, poly8x16x3_t, poly8_t*, p8, x3);
> -TEST_ST1x3 (vst1q, int16x8x3_t, int16_t*, s16, x3);
> -TEST_ST1x3 (vst1q, uint16x8x3_t, uint16_t*, u16, x3);
> -TEST_ST1x3 (vst1q, poly16x8x3_t, poly16_t*, p16, x3);
> -TEST_ST1x3 (vst1q, float16x8x3_t, float16_t*, f16, x3);
> -TEST_ST1x3 (vst1q, bfloat16x8x3_t, bfloat16_t*, bf16, x3);
> -TEST_ST1x3 (vst1q, int32x4x3_t, int32_t*, s32, x3);
> -TEST_ST1x3 (vst1q, uint32x4x3_t, uint32_t*, u32, x3);
> -TEST_ST1x3 (vst1q, float32x4x3_t, float32_t*, f32, x3);
> -TEST_ST1x3 (vst1q, int64x2x3_t, int64_t*, s64, x3);
> -TEST_ST1x3 (vst1q, uint64x2x3_t, uint64_t*, u64, x3);
> -TEST_ST1x3 (vst1q, poly64x2x3_t, poly64_t*, p64, x3);
> -TEST_ST1x3 (vst1q, float64x2x3_t, float64_t*, f64, x3);
> +TEST_LD1xN (vld1, int8x8x2_t, int8_t*, s8, x2);
> +TEST_LD1xN (vld1, uint8x8x2_t, uint8_t*, u8, x2);
> +TEST_LD1xN (vld1, poly8x8x2_t, poly8_t*, p8, x2);
> +TEST_LD1xN (vld1, int16x4x2_t, int16_t*, s16, x2);
> +TEST_LD1xN (vld1, uint16x4x2_t, uint16_t*, u16, x2);
> +TEST_LD1xN (vld1, poly16x4x2_t, poly16_t*, p16, x2);
> +TEST_LD1xN (vld1, float16x4x2_t, float16_t*, f16, x2);
> +TEST_LD1xN (vld1, bfloat16x4x2_t, bfloat16_t*, bf16, x2);
> +TEST_LD1xN (vld1, int32x2x2_t, int32_t*, s32, x2);
> +TEST_LD1xN (vld1, uint32x2x2_t, uint32_t*, u32, x2);
> +TEST_LD1xN (vld1, float32x2x2_t, float32_t*, f32, x2);
> +TEST_LD1xN (vld1, int64x1x2_t, int64_t*, s64, x2);
> +TEST_LD1xN (vld1, uint64x1x2_t, uint64_t*, u64, x2);
> +TEST_LD1xN (vld1, poly64x1x2_t, poly64_t*, p64, x2);
> +TEST_LD1xN (vld1, float64x1x2_t, float64_t*, f64, x2);
> +
> +TEST_LD1xN (vld1q, int8x16x2_t, int8_t*, s8, x2);
> +TEST_LD1xN (vld1q, uint8x16x2_t, uint8_t*, u8, x2);
> +TEST_LD1xN (vld1q, poly8x16x2_t, poly8_t*, p8, x2);
> +TEST_LD1xN (vld1q, int16x8x2_t, int16_t*, s16, x2);
> +TEST_LD1xN (vld1q, uint16x8x2_t, uint16_t*, u16, x2);
> +TEST_LD1xN (vld1q, poly16x8x2_t, poly16_t*, p16, x2);
> +TEST_LD1xN (vld1q, float16x8x2_t, float16_t*, f16, x2);
> +TEST_LD1xN (vld1q, bfloat16x8x2_t, bfloat16_t*, bf16, x2);
> +TEST_LD1xN (vld1q, int32x4x2_t, int32_t*, s32, x2);
> +TEST_LD1xN (vld1q, uint32x4x2_t, uint32_t*, u32, x2);
> +TEST_LD1xN (vld1q, float32x4x2_t, float32_t*, f32, x2);
> +TEST_LD1xN (vld1q, int64x2x2_t, int64_t*, s64, x2);
> +TEST_LD1xN (vld1q, uint64x2x2_t, uint64_t*, u64, x2);
> +TEST_LD1xN (vld1q, poly64x2x2_t, poly64_t*, p64, x2);
> +TEST_LD1xN (vld1q, float64x2x2_t, float64_t*, f64, x2);
> +
> +TEST_LD1xN (vld1, int8x8x3_t, int8_t*, s8, x3);
> +TEST_LD1xN (vld1, uint8x8x3_t, uint8_t*, u8, x3);
> +TEST_LD1xN (vld1, poly8x8x3_t, poly8_t*, p8, x3);
> +TEST_LD1xN (vld1, int16x4x3_t, int16_t*, s16, x3);
> +TEST_LD1xN (vld1, uint16x4x3_t, uint16_t*, u16, x3);
> +TEST_LD1xN (vld1, poly16x4x3_t, poly16_t*, p16, x3);
> +TEST_LD1xN (vld1, float16x4x3_t, float16_t*, f16, x3);
> +TEST_LD1xN (vld1, bfloat16x4x3_t, bfloat16_t*, bf16, x3);
> +TEST_LD1xN (vld1, int32x2x3_t, int32_t*, s32, x3);
> +TEST_LD1xN (vld1, uint32x2x3_t, uint32_t*, u32, x3);
> +TEST_LD1xN (vld1, float32x2x3_t, float32_t*, f32, x3);
> +TEST_LD1xN (vld1, int64x1x3_t, int64_t*, s64, x3);
> +TEST_LD1xN (vld1, uint64x1x3_t, uint64_t*, u64, x3);
> +TEST_LD1xN (vld1, poly64x1x3_t, poly64_t*, p64, x3);
> +TEST_LD1xN (vld1, float64x1x3_t, float64_t*, f64, x3);
> +
> +TEST_LD1xN (vld1q, int8x16x3_t, int8_t*, s8, x3);
> +TEST_LD1xN (vld1q, uint8x16x3_t, uint8_t*, u8, x3);
> +TEST_LD1xN (vld1q, poly8x16x3_t, poly8_t*, p8, x3);
> +TEST_LD1xN (vld1q, int16x8x3_t, int16_t*, s16, x3);
> +TEST_LD1xN (vld1q, uint16x8x3_t, uint16_t*, u16, x3);
> +TEST_LD1xN (vld1q, poly16x8x3_t, poly16_t*, p16, x3);
> +TEST_LD1xN (vld1q, float16x8x3_t, float16_t*, f16, x3);
> +TEST_LD1xN (vld1q, bfloat16x8x3_t, bfloat16_t*, bf16, x3);
> +TEST_LD1xN (vld1q, int32x4x3_t, int32_t*, s32, x3);
> +TEST_LD1xN (vld1q, uint32x4x3_t, uint32_t*, u32, x3);
> +TEST_LD1xN (vld1q, float32x4x3_t, float32_t*, f32, x3);
> +TEST_LD1xN (vld1q, int64x2x3_t, int64_t*, s64, x3);
> +TEST_LD1xN (vld1q, uint64x2x3_t, uint64_t*, u64, x3);
> +TEST_LD1xN (vld1q, poly64x2x3_t, poly64_t*, p64, x3);
> +TEST_LD1xN (vld1q, float64x2x3_t, float64_t*, f64, x3);
> +
> +TEST_LD1xN (vld1, int8x8x4_t, int8_t*, s8, x4);
> +TEST_LD1xN (vld1, uint8x8x4_t, uint8_t*, u8, x4);
> +TEST_LD1xN (vld1, poly8x8x4_t, poly8_t*, p8, x4);
> +TEST_LD1xN (vld1, int16x4x4_t, int16_t*, s16, x4);
> +TEST_LD1xN (vld1, uint16x4x4_t, uint16_t*, u16, x4);
> +TEST_LD1xN (vld1, poly16x4x4_t, poly16_t*, p16, x4);
> +TEST_LD1xN (vld1, float16x4x4_t, float16_t*, f16, x4);
> +TEST_LD1xN (vld1, bfloat16x4x4_t, bfloat16_t*, bf16, x4);
> +TEST_LD1xN (vld1, int32x2x4_t, int32_t*, s32, x4);
> +TEST_LD1xN (vld1, uint32x2x4_t, uint32_t*, u32, x4);
> +TEST_LD1xN (vld1, float32x2x4_t, float32_t*, f32, x4);
> +TEST_LD1xN (vld1, int64x1x4_t, int64_t*, s64, x4);
> +TEST_LD1xN (vld1, uint64x1x4_t, uint64_t*, u64, x4);
> +TEST_LD1xN (vld1, poly64x1x4_t, poly64_t*, p64, x4);
> +TEST_LD1xN (vld1, float64x1x4_t, float64_t*, f64, x4);
> +
> +TEST_LD1xN (vld1q, int8x16x4_t, int8_t*, s8, x4);
> +TEST_LD1xN (vld1q, uint8x16x4_t, uint8_t*, u8, x4);
> +TEST_LD1xN (vld1q, poly8x16x4_t, poly8_t*, p8, x4);
> +TEST_LD1xN (vld1q, int16x8x4_t, int16_t*, s16, x4);
> +TEST_LD1xN (vld1q, uint16x8x4_t, uint16_t*, u16, x4);
> +TEST_LD1xN (vld1q, poly16x8x4_t, poly16_t*, p16, x4);
> +TEST_LD1xN (vld1q, float16x8x4_t, float16_t*, f16, x4);
> +TEST_LD1xN (vld1q, bfloat16x8x4_t, bfloat16_t*, bf16, x4);
> +TEST_LD1xN (vld1q, int32x4x4_t, int32_t*, s32, x4);
> +TEST_LD1xN (vld1q, uint32x4x4_t, uint32_t*, u32, x4);
> +TEST_LD1xN (vld1q, float32x4x4_t, float32_t*, f32, x4);
> +TEST_LD1xN (vld1q, int64x2x4_t, int64_t*, s64, x4);
> +TEST_LD1xN (vld1q, uint64x2x4_t, uint64_t*, u64, x4);
> +TEST_LD1xN (vld1q, poly64x2x4_t, poly64_t*, p64, x4);
> +TEST_LD1xN (vld1q, float64x2x4_t, float64_t*, f64, x4);
>  
>  /* { dg-final { scan-assembler-not "mov\\t" { target aarch64_little_endian } } } */
> +/* { dg-final { scan-assembler-not "ldr\\t" { target aarch64_little_endian } } } */
> +/* { dg-final { scan-assembler-not "str\\t" { target aarch64_little_endian } } } */
> +/* { dg-final { scan-assembler-not "sp" { target aarch64_little_endian } } } */
>  
>  /* { dg-final { scan-assembler-times "tbl\\t" 18} }  */
>  /* { dg-final { scan-assembler-times "tbx\\t" 18} }  */
> -/* { dg-final { scan-assembler-times "st4\\t" 30} }  */
> -/* { dg-final { scan-assembler-times "st3\\t" 30} }  */
> -/* { dg-final { scan-assembler-times "st2\\t" 30} }  */
> -/* { dg-final { scan-assembler-times "st1\\t" 45} }  */
> +/* { dg-final { scan-assembler-times "st4\\t" 56} }  */
> +/* { dg-final { scan-assembler-times "st3\\t" 56} }  */
> +/* { dg-final { scan-assembler-times "st2\\t" 56} }  */
> +/* { dg-final { scan-assembler-times "st1\\t" 102} }  */
> +/* { dg-final { scan-assembler-times "ld4\\t" 56} }  */
> +/* { dg-final { scan-assembler-times "ld3\\t" 56} }  */
> +/* { dg-final { scan-assembler-times "ld2\\t" 56} }  */
> +/* { dg-final { scan-assembler-times "ld1\\t" 102} }  */


More information about the Gcc-patches mailing list