[PATCH, rs6000] (2/3) Fix widening multiply high/low operations for little endian

Bill Schmidt wschmidt@linux.vnet.ibm.com
Mon Nov 4 18:05:00 GMT 2013


Per Richard S's suggestion, I'm reworking parts 1 and 3 of the patch
set, but this one will remain unchanged and is ready for review.

Thanks,
Bill

On Sun, 2013-11-03 at 23:34 -0600, Bill Schmidt wrote:
> Hi,
> 
> This patch fixes the widening multiply high/low operations to work
> correctly in the presence of the first patch of this series, which
> reverses the meanings of multiply even/odd instructions.  Here we
> reorder the input operands to the vector merge low/high instructions.
> 
> The general rule is that vmrghh(x,y) [BE] = vmrglh(y,x) [LE], and so on;
> that is, we need to reverse the usage of merge high and merge low, and
> also swap their inputs, to obtain the same semantics.  In this case we
> are only swapping the inputs, because the reversed usage of high and low
> has already been done for us in the generic handling code for
> VEC_WIDEN_MULT_LO_EXPR.
> 
> Bootstrapped and tested with the rest of the patch set on
> powerpc64{,le}-unknown-linux-gnu, with no regressions.  Is this ok for
> trunk?
> 
> Thanks,
> Bill
> 
> 
> 2013-11-03  Bill Schmidt  <wschmidt@linux.vnet.ibm.com>
> 
> 	* config/rs6000/altivec.md (vec_widen_umult_hi_v16qi): Swap
> 	arguments to merge instruction for little endian.
> 	(vec_widen_umult_lo_v16qi): Likewise.
> 	(vec_widen_smult_hi_v16qi): Likewise.
> 	(vec_widen_smult_lo_v16qi): Likewise.
> 	(vec_widen_umult_hi_v8hi): Likewise.
> 	(vec_widen_umult_lo_v8hi): Likewise.
> 	(vec_widen_smult_hi_v8hi): Likewise.
> 	(vec_widen_smult_lo_v8hi): Likewise.
> 
> 
> Index: gcc/config/rs6000/altivec.md
> ===================================================================
> --- gcc/config/rs6000/altivec.md	(revision 204192)
> +++ gcc/config/rs6000/altivec.md	(working copy)
> @@ -2185,7 +2235,10 @@
>    
>    emit_insn (gen_vec_widen_umult_even_v16qi (ve, operands[1], operands[2]));
>    emit_insn (gen_vec_widen_umult_odd_v16qi (vo, operands[1], operands[2]));
> -  emit_insn (gen_altivec_vmrghh (operands[0], ve, vo));
> +  if (BYTES_BIG_ENDIAN)
> +    emit_insn (gen_altivec_vmrghh (operands[0], ve, vo));
> +  else
> +    emit_insn (gen_altivec_vmrghh (operands[0], vo, ve));
>    DONE;
>  }")
> 
> @@ -2202,7 +2255,10 @@
>    
>    emit_insn (gen_vec_widen_umult_even_v16qi (ve, operands[1], operands[2]));
>    emit_insn (gen_vec_widen_umult_odd_v16qi (vo, operands[1], operands[2]));
> -  emit_insn (gen_altivec_vmrglh (operands[0], ve, vo));
> +  if (BYTES_BIG_ENDIAN)
> +    emit_insn (gen_altivec_vmrglh (operands[0], ve, vo));
> +  else
> +    emit_insn (gen_altivec_vmrglh (operands[0], vo, ve));
>    DONE;
>  }")
> 
> @@ -2219,7 +2275,10 @@
>    
>    emit_insn (gen_vec_widen_smult_even_v16qi (ve, operands[1], operands[2]));
>    emit_insn (gen_vec_widen_smult_odd_v16qi (vo, operands[1], operands[2]));
> -  emit_insn (gen_altivec_vmrghh (operands[0], ve, vo));
> +  if (BYTES_BIG_ENDIAN)
> +    emit_insn (gen_altivec_vmrghh (operands[0], ve, vo));
> +  else
> +    emit_insn (gen_altivec_vmrghh (operands[0], vo, ve));
>    DONE;
>  }")
> 
> @@ -2236,7 +2295,10 @@
>    
>    emit_insn (gen_vec_widen_smult_even_v16qi (ve, operands[1], operands[2]));
>    emit_insn (gen_vec_widen_smult_odd_v16qi (vo, operands[1], operands[2]));
> -  emit_insn (gen_altivec_vmrglh (operands[0], ve, vo));
> +  if (BYTES_BIG_ENDIAN)
> +    emit_insn (gen_altivec_vmrglh (operands[0], ve, vo));
> +  else
> +    emit_insn (gen_altivec_vmrglh (operands[0], vo, ve));
>    DONE;
>  }")
> 
> @@ -2253,7 +2315,10 @@
>    
>    emit_insn (gen_vec_widen_umult_even_v8hi (ve, operands[1], operands[2]));
>    emit_insn (gen_vec_widen_umult_odd_v8hi (vo, operands[1], operands[2]));
> -  emit_insn (gen_altivec_vmrghw (operands[0], ve, vo));
> +  if (BYTES_BIG_ENDIAN)
> +    emit_insn (gen_altivec_vmrghw (operands[0], ve, vo));
> +  else
> +    emit_insn (gen_altivec_vmrghw (operands[0], vo, ve));
>    DONE;
>  }")
> 
> @@ -2270,7 +2335,10 @@
>    
>    emit_insn (gen_vec_widen_umult_even_v8hi (ve, operands[1], operands[2]));
>    emit_insn (gen_vec_widen_umult_odd_v8hi (vo, operands[1], operands[2]));
> -  emit_insn (gen_altivec_vmrglw (operands[0], ve, vo));
> +  if (BYTES_BIG_ENDIAN)
> +    emit_insn (gen_altivec_vmrglw (operands[0], ve, vo));
> +  else
> +    emit_insn (gen_altivec_vmrglw (operands[0], vo, ve));
>    DONE;
>  }")
> 
> @@ -2287,7 +2355,10 @@
>    
>    emit_insn (gen_vec_widen_smult_even_v8hi (ve, operands[1], operands[2]));
>    emit_insn (gen_vec_widen_smult_odd_v8hi (vo, operands[1], operands[2]));
> -  emit_insn (gen_altivec_vmrghw (operands[0], ve, vo));
> +  if (BYTES_BIG_ENDIAN)
> +    emit_insn (gen_altivec_vmrghw (operands[0], ve, vo));
> +  else
> +    emit_insn (gen_altivec_vmrghw (operands[0], vo, ve));
>    DONE;
>  }")
> 
> @@ -2304,7 +2375,10 @@
>    
>    emit_insn (gen_vec_widen_smult_even_v8hi (ve, operands[1], operands[2]));
>    emit_insn (gen_vec_widen_smult_odd_v8hi (vo, operands[1], operands[2]));
> -  emit_insn (gen_altivec_vmrglw (operands[0], ve, vo));
> +  if (BYTES_BIG_ENDIAN)
> +    emit_insn (gen_altivec_vmrglw (operands[0], ve, vo));
> +  else
> +    emit_insn (gen_altivec_vmrglw (operands[0], vo, ve));
>    DONE;
>  }")
> 
> 
> 



More information about the Gcc-patches mailing list