Index: gcc/config/aarch64/aarch64-simd.md
===================================================================
--- gcc/config/aarch64/aarch64-simd.md (revision 210319)
+++ gcc/config/aarch64/aarch64-simd.md (working copy)
@@ -19,8 +19,8 @@
;; .
(define_expand "mov"
- [(set (match_operand:VALL 0 "aarch64_simd_nonimmediate_operand" "")
- (match_operand:VALL 1 "aarch64_simd_general_operand" ""))]
+ [(set (match_operand:VALL 0 "nonimmediate_operand" "")
+ (match_operand:VALL 1 "general_operand" ""))]
"TARGET_SIMD"
"
if (GET_CODE (operands[0]) == MEM)
@@ -29,8 +29,8 @@
)
(define_expand "movmisalign"
- [(set (match_operand:VALL 0 "aarch64_simd_nonimmediate_operand" "")
- (match_operand:VALL 1 "aarch64_simd_general_operand" ""))]
+ [(set (match_operand:VALL 0 "nonimmediate_operand" "")
+ (match_operand:VALL 1 "general_operand" ""))]
"TARGET_SIMD"
{
/* This pattern is not permitted to fail during expansion: if both arguments
@@ -91,9 +91,9 @@
)
(define_insn "*aarch64_simd_mov"
- [(set (match_operand:VD 0 "aarch64_simd_nonimmediate_operand"
+ [(set (match_operand:VD 0 "nonimmediate_operand"
"=w, m, w, ?r, ?w, ?r, w")
- (match_operand:VD 1 "aarch64_simd_general_operand"
+ (match_operand:VD 1 "general_operand"
"m, w, w, w, r, r, Dn"))]
"TARGET_SIMD
&& (register_operand (operands[0], mode)
@@ -119,9 +119,9 @@
)
(define_insn "*aarch64_simd_mov"
- [(set (match_operand:VQ 0 "aarch64_simd_nonimmediate_operand"
+ [(set (match_operand:VQ 0 "nonimmediate_operand"
"=w, m, w, ?r, ?w, ?r, w")
- (match_operand:VQ 1 "aarch64_simd_general_operand"
+ (match_operand:VQ 1 "general_operand"
"m, w, w, w, r, r, Dn"))]
"TARGET_SIMD
&& (register_operand (operands[0], mode)
Index: gcc/config/aarch64/aarch64.c
===================================================================
--- gcc/config/aarch64/aarch64.c (revision 210319)
+++ gcc/config/aarch64/aarch64.c (working copy)
@@ -3075,11 +3075,11 @@ aarch64_classify_address (struct aarch64_address_i
enum rtx_code code = GET_CODE (x);
rtx op0, op1;
bool allow_reg_index_p =
- outer_code != PARALLEL && GET_MODE_SIZE(mode) != 16;
-
+ outer_code != PARALLEL && (GET_MODE_SIZE (mode) != 16
+ || aarch64_vector_mode_supported_p (mode));
/* Don't support anything other than POST_INC or REG addressing for
AdvSIMD. */
- if (aarch64_vector_mode_p (mode)
+ if (aarch64_vect_struct_mode_p (mode)
&& (code != POST_INC && code != REG))
return false;
@@ -4010,7 +4010,7 @@ aarch64_legitimize_reload_address (rtx *x_p,
rtx x = *x_p;
/* Do not allow mem (plus (reg, const)) if vector mode. */
- if (aarch64_vector_mode_p (mode)
+ if (aarch64_vect_struct_mode_p (mode)
&& GET_CODE (x) == PLUS
&& REG_P (XEXP (x, 0))
&& CONST_INT_P (XEXP (x, 1)))