VREPLACE_ELT_V2DF
[VEC_REPLACE_UN, vec_replace_unaligned, __builtin_vec_replace_un]
- vui __builtin_vec_replace_un (vui, unsigned int, const int);
+ vuc __builtin_vec_replace_un (vui, unsigned int, const int);
VREPLACE_UN_UV4SI
- vsi __builtin_vec_replace_un (vsi, signed int, const int);
+ vuc __builtin_vec_replace_un (vsi, signed int, const int);
VREPLACE_UN_V4SI
- vull __builtin_vec_replace_un (vull, unsigned long long, const int);
+ vuc __builtin_vec_replace_un (vull, unsigned long long, const int);
VREPLACE_UN_UV2DI
- vsll __builtin_vec_replace_un (vsll, signed long long, const int);
+ vuc __builtin_vec_replace_un (vsll, signed long long, const int);
VREPLACE_UN_V2DI
- vf __builtin_vec_replace_un (vf, float, const int);
+ vuc __builtin_vec_replace_un (vf, float, const int);
VREPLACE_UN_V4SF
- vd __builtin_vec_replace_un (vd, double, const int);
+ vuc __builtin_vec_replace_un (vd, double, const int);
VREPLACE_UN_V2DF
[VEC_REVB, vec_revb, __builtin_vec_revb]
signed int __builtin_vec_scalar_cmp_exp_eq (double, double);
VSCEDPEQ
signed int __builtin_vec_scalar_cmp_exp_eq (_Float128, _Float128);
- VSCEQPEQ
+ VSCEQPEQ_KF
+ signed int __builtin_vec_scalar_cmp_exp_eq (long double, long double);
+ VSCEQPEQ_TF
[VEC_VSCEGT, scalar_cmp_exp_gt, __builtin_vec_scalar_cmp_exp_gt]
signed int __builtin_vec_scalar_cmp_exp_gt (double, double);
VSCEDPGT
signed int __builtin_vec_scalar_cmp_exp_gt (_Float128, _Float128);
- VSCEQPGT
+ VSCEQPGT_KF
+ signed int __builtin_vec_scalar_cmp_exp_gt (long double, long double);
+ VSCEQPGT_TF
[VEC_VSCELT, scalar_cmp_exp_lt, __builtin_vec_scalar_cmp_exp_lt]
signed int __builtin_vec_scalar_cmp_exp_lt (double, double);
VSCEDPLT
signed int __builtin_vec_scalar_cmp_exp_lt (_Float128, _Float128);
- VSCEQPLT
+ VSCEQPLT_KF
+ signed int __builtin_vec_scalar_cmp_exp_lt (long double, long double);
+ VSCEQPLT_TF
[VEC_VSCEUO, scalar_cmp_exp_unordered, __builtin_vec_scalar_cmp_exp_unordered]
signed int __builtin_vec_scalar_cmp_exp_unordered (double, double);
VSCEDPUO
signed int __builtin_vec_scalar_cmp_exp_unordered (_Float128, _Float128);
- VSCEQPUO
+ VSCEQPUO_KF
+ signed int __builtin_vec_scalar_cmp_exp_unordered (long double, long double);
+ VSCEQPUO_TF
[VEC_VSEE, scalar_extract_exp, __builtin_vec_scalar_extract_exp]
unsigned int __builtin_vec_scalar_extract_exp (double);
VUPKLSW VUPKLSW_DEPR1
vbll __builtin_vec_vupklsw (vbi);
VUPKLSW VUPKLSW_DEPR2
+
+[ADDF128_ODD, SKIP, __builtin_addf128_round_to_odd]
+ long double __builtin_addf128_round_to_odd (long double, long double);
+ ADDF128_ODD_TF
+ _Float128 __builtin_addf128_round_to_odd (_Float128, _Float128);
+ ADDF128_ODD_KF
+
+[DIVF128_ODD, SKIP, __builtin_divf128_round_to_odd]
+ long double __builtin_divf128_round_to_odd (long double, long double);
+ DIVF128_ODD_TF
+ _Float128 __builtin_divf128_round_to_odd (_Float128, _Float128);
+ DIVF128_ODD_KF
+
+[FMAF128_ODD, SKIP, __builtin_fmaf128_round_to_odd]
+ long double __builtin_fmaf128_round_to_odd (long double, long double, \
+ long double);
+ FMAF128_ODD_TF
+ _Float128 __builtin_fmaf128_round_to_odd (_Float128, _Float128, \
+ _Float128);
+ FMAF128_ODD_KF
+
+[MULF128_ODD, SKIP, __builtin_mulf128_round_to_odd]
+ long double __builtin_mulf128_round_to_odd (long double, long double);
+ MULF128_ODD_TF
+ _Float128 __builtin_mulf128_round_to_odd (_Float128, _Float128);
+ MULF128_ODD_KF
+
+[SQRTF128_ODD, SKIP, __builtin_sqrtf128_round_to_odd]
+ long double __builtin_sqrtf128_round_to_odd (long double);
+ SQRTF128_ODD_TF
+ _Float128 __builtin_sqrtf128_round_to_odd (_Float128);
+ SQRTF128_ODD_KF
+
+[SUBF128_ODD, SKIP, __builtin_subf128_round_to_odd]
+ long double __builtin_subf128_round_to_odd (long double, long double);
+ SUBF128_ODD_TF
+ _Float128 __builtin_subf128_round_to_odd (_Float128, _Float128);
+ SUBF128_ODD_KF
+
+[TRUNCF128_ODD, SKIP, __builtin_truncf128_round_to_odd]
+ long double __builtin_truncf128_round_to_odd (long double);
+ TRUNCF128_ODD_TF
+ _Float128 __builtin_truncf128_round_to_odd (_Float128);
+ TRUNCF128_ODD_KF