[PATCH, rs6000] Don't use deprecated vector shift instructions
Bill Schmidt
wschmidt@linux.ibm.com
Thu Oct 25 07:28:00 GMT 2018
Hi,
The intrinsic compatibility headers make use of some deprecated functions for
vector shifts, which are not available in some compilers. For compatibility
reasons, this patch, replaces those with intrinsics guaranteed to be portable.
Bootstrapped and tested on powerpc64le-linux-gnu with no regressions.
Is this okay for trunk?
Thanks,
Bill
2018-10-24 Bill Schmidt <wschmidt@linux.ibm.com>
Jinsong Ji <jji@us.ibm.com>
* config/rs6000/emmintrin.h (_mm_slli_epi16): Replace deprecated
function with vec_sl.
(_mm_slli_epi32): Likewise.
(_mm_slli_epi64): Likewise.
(_mm_srai_epi16): Replace deprecated function with vec_sra.
(_mm_srai_epi32): Likewise.
(_mm_srli_epi16): Replace deprecated function with vec_sr.
(_mm_srli_epi32): Likewise.
(_mm_srli_epi64): Likewise.
(_mm_sll_epi16): Replace deprecated function with vec_sl.
(_mm_sll_epi32): Likewise.
(_mm_sll_epi64): Likewise.
(_mm_sra_epi16): Replace deprecated function with vec_sra.
(_mm_sra_epi32): Likewise.
(_mm_srl_epi16): Replace deprecated function with vec_sr.
(_mm_srl_epi32): Likewise.
(_mm_srl_epi64): Likewise.
Index: gcc/config/rs6000/emmintrin.h
===================================================================
--- gcc/config/rs6000/emmintrin.h (revision 265464)
+++ gcc/config/rs6000/emmintrin.h (working copy)
@@ -1504,7 +1504,7 @@ _mm_slli_epi16 (__m128i __A, int __B)
else
lshift = vec_splats ((unsigned short) __B);
- result = vec_vslh ((__v8hi) __A, lshift);
+ result = vec_sl ((__v8hi) __A, lshift);
}
return (__m128i) result;
@@ -1523,7 +1523,7 @@ _mm_slli_epi32 (__m128i __A, int __B)
else
lshift = vec_splats ((unsigned int) __B);
- result = vec_vslw ((__v4si) __A, lshift);
+ result = vec_sl ((__v4si) __A, lshift);
}
return (__m128i) result;
@@ -1543,7 +1543,7 @@ _mm_slli_epi64 (__m128i __A, int __B)
else
lshift = (__v2du) vec_splats ((unsigned int) __B);
- result = vec_vsld ((__v2di) __A, lshift);
+ result = vec_sl ((__v2di) __A, lshift);
}
return (__m128i) result;
@@ -1563,7 +1563,7 @@ _mm_srai_epi16 (__m128i __A, int __B)
else
rshift = vec_splats ((unsigned short) __B);
}
- result = vec_vsrah ((__v8hi) __A, rshift);
+ result = vec_sra ((__v8hi) __A, rshift);
return (__m128i) result;
}
@@ -1586,7 +1586,7 @@ _mm_srai_epi32 (__m128i __A, int __B)
else
rshift = vec_splats ((unsigned int) __B);
}
- result = vec_vsraw ((__v4si) __A, rshift);
+ result = vec_sra ((__v4si) __A, rshift);
return (__m128i) result;
}
@@ -1666,7 +1666,7 @@ _mm_srli_epi16 (__m128i __A, int __B)
else
rshift = vec_splats ((unsigned short) __B);
- result = vec_vsrh ((__v8hi) __A, rshift);
+ result = vec_sr ((__v8hi) __A, rshift);
}
return (__m128i) result;
@@ -1690,7 +1690,7 @@ _mm_srli_epi32 (__m128i __A, int __B)
else
rshift = vec_splats ((unsigned int) __B);
- result = vec_vsrw ((__v4si) __A, rshift);
+ result = vec_sr ((__v4si) __A, rshift);
}
return (__m128i) result;
@@ -1715,7 +1715,7 @@ _mm_srli_epi64 (__m128i __A, int __B)
else
rshift = (__v2du) vec_splats ((unsigned int) __B);
- result = vec_vsrd ((__v2di) __A, rshift);
+ result = vec_sr ((__v2di) __A, rshift);
}
return (__m128i) result;
@@ -1735,7 +1735,7 @@ _mm_sll_epi16 (__m128i __A, __m128i __B)
lshift = vec_splat ((__v8hu)__B, 3);
#endif
shmask = lshift <= shmax;
- result = vec_vslh ((__v8hu) __A, lshift);
+ result = vec_sl ((__v8hu) __A, lshift);
result = vec_sel (shmask, result, shmask);
return (__m128i) result;
@@ -1753,7 +1753,7 @@ _mm_sll_epi32 (__m128i __A, __m128i __B)
lshift = vec_splat ((__v4su)__B, 1);
#endif
shmask = lshift < shmax;
- result = vec_vslw ((__v4su) __A, lshift);
+ result = vec_sl ((__v4su) __A, lshift);
result = vec_sel (shmask, result, shmask);
return (__m128i) result;
@@ -1769,7 +1769,7 @@ _mm_sll_epi64 (__m128i __A, __m128i __B)
lshift = (__v2du) vec_splat ((__v2du)__B, 0);
shmask = lshift < shmax;
- result = vec_vsld ((__v2du) __A, lshift);
+ result = vec_sl ((__v2du) __A, lshift);
result = (__v2du) vec_sel ((__v2df) shmask, (__v2df) result,
(__v2df) shmask);
@@ -1790,7 +1790,7 @@ _mm_sra_epi16 (__m128i __A, __m128i __B)
rshift = vec_splat ((__v8hu)__B, 3);
#endif
rshift = vec_min (rshift, rshmax);
- result = vec_vsrah ((__v8hi) __A, rshift);
+ result = vec_sra ((__v8hi) __A, rshift);
return (__m128i) result;
}
@@ -1808,7 +1808,7 @@ _mm_sra_epi32 (__m128i __A, __m128i __B)
rshift = vec_splat ((__v4su)__B, 1);
#endif
rshift = vec_min (rshift, rshmax);
- result = vec_vsraw ((__v4si) __A, rshift);
+ result = vec_sra ((__v4si) __A, rshift);
return (__m128i) result;
}
@@ -1826,7 +1826,7 @@ _mm_srl_epi16 (__m128i __A, __m128i __B)
rshift = vec_splat ((__v8hu)__B, 3);
#endif
shmask = rshift <= shmax;
- result = vec_vsrh ((__v8hu) __A, rshift);
+ result = vec_sr ((__v8hu) __A, rshift);
result = vec_sel (shmask, result, shmask);
return (__m128i) result;
@@ -1845,7 +1845,7 @@ _mm_srl_epi32 (__m128i __A, __m128i __B)
rshift = vec_splat ((__v4su)__B, 1);
#endif
shmask = rshift < shmax;
- result = vec_vsrw ((__v4su) __A, rshift);
+ result = vec_sr ((__v4su) __A, rshift);
result = vec_sel (shmask, result, shmask);
return (__m128i) result;
@@ -1861,7 +1861,7 @@ _mm_srl_epi64 (__m128i __A, __m128i __B)
rshift = (__v2du) vec_splat ((__v2du)__B, 0);
shmask = rshift < shmax;
- result = vec_vsrd ((__v2du) __A, rshift);
+ result = vec_sr ((__v2du) __A, rshift);
result = (__v2du)vec_sel ((__v2du)shmask, (__v2du)result, (__v2du)shmask);
return (__m128i) result;
More information about the Gcc-patches
mailing list