]> gcc.gnu.org Git - gcc.git/blame - gcc/config/i386/avx512erintrin.h
Push evex512 target for 512 bit intrins
[gcc.git] / gcc / config / i386 / avx512erintrin.h
CommitLineData
83ffe9cd 1/* Copyright (C) 2013-2023 Free Software Foundation, Inc.
756c5857
AI
2
3 This file is part of GCC.
4
5 GCC is free software; you can redistribute it and/or modify
6 it under the terms of the GNU General Public License as published by
7 the Free Software Foundation; either version 3, or (at your option)
8 any later version.
9
10 GCC is distributed in the hope that it will be useful,
11 but WITHOUT ANY WARRANTY; without even the implied warranty of
12 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
13 GNU General Public License for more details.
14
15 Under Section 7 of GPL version 3, you are granted additional
16 permissions described in the GCC Runtime Library Exception, version
17 3.1, as published by the Free Software Foundation.
18
19 You should have received a copy of the GNU General Public License and
20 a copy of the GCC Runtime Library Exception along with this program;
21 see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
22 <http://www.gnu.org/licenses/>. */
23
24#ifndef _IMMINTRIN_H_INCLUDED
25#error "Never use <avx512erintrin.h> directly; include <immintrin.h> instead."
26#endif
27
28#ifndef _AVX512ERINTRIN_H_INCLUDED
29#define _AVX512ERINTRIN_H_INCLUDED
30
31#ifndef __AVX512ER__
32#pragma GCC push_options
2b77bcbb 33#pragma GCC target("avx512er,evex512")
756c5857
AI
34#define __DISABLE_AVX512ER__
35#endif /* __AVX512ER__ */
36
37/* Internal data types for implementing the intrinsics. */
38typedef double __v8df __attribute__ ((__vector_size__ (64)));
39typedef float __v16sf __attribute__ ((__vector_size__ (64)));
40
41/* The Intel API is flexible enough that we must allow aliasing with other
42 vector types, and their scalar components. */
43typedef float __m512 __attribute__ ((__vector_size__ (64), __may_alias__));
44typedef double __m512d __attribute__ ((__vector_size__ (64), __may_alias__));
45
46typedef unsigned char __mmask8;
47typedef unsigned short __mmask16;
48
49#ifdef __OPTIMIZE__
50extern __inline __m512d
51__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
52_mm512_exp2a23_round_pd (__m512d __A, int __R)
53{
756c5857 54 return (__m512d) __builtin_ia32_exp2pd_mask ((__v8df) __A,
41602390 55 (__v8df) _mm512_undefined_pd (),
756c5857
AI
56 (__mmask8) -1, __R);
57}
58
59extern __inline __m512d
60__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
61_mm512_mask_exp2a23_round_pd (__m512d __W, __mmask8 __U, __m512d __A, int __R)
62{
63 return (__m512d) __builtin_ia32_exp2pd_mask ((__v8df) __A,
64 (__v8df) __W,
65 (__mmask8) __U, __R);
66}
67
68extern __inline __m512d
69__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
70_mm512_maskz_exp2a23_round_pd (__mmask8 __U, __m512d __A, int __R)
71{
72 return (__m512d) __builtin_ia32_exp2pd_mask ((__v8df) __A,
73 (__v8df) _mm512_setzero_pd (),
74 (__mmask8) __U, __R);
75}
76
77extern __inline __m512
78__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
79_mm512_exp2a23_round_ps (__m512 __A, int __R)
80{
756c5857 81 return (__m512) __builtin_ia32_exp2ps_mask ((__v16sf) __A,
41602390 82 (__v16sf) _mm512_undefined_ps (),
756c5857
AI
83 (__mmask16) -1, __R);
84}
85
86extern __inline __m512
87__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
88_mm512_mask_exp2a23_round_ps (__m512 __W, __mmask16 __U, __m512 __A, int __R)
89{
90 return (__m512) __builtin_ia32_exp2ps_mask ((__v16sf) __A,
91 (__v16sf) __W,
92 (__mmask16) __U, __R);
93}
94
95extern __inline __m512
96__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
97_mm512_maskz_exp2a23_round_ps (__mmask16 __U, __m512 __A, int __R)
98{
99 return (__m512) __builtin_ia32_exp2ps_mask ((__v16sf) __A,
100 (__v16sf) _mm512_setzero_ps (),
101 (__mmask16) __U, __R);
102}
103
104extern __inline __m512d
105__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
106_mm512_rcp28_round_pd (__m512d __A, int __R)
107{
756c5857 108 return (__m512d) __builtin_ia32_rcp28pd_mask ((__v8df) __A,
41602390 109 (__v8df) _mm512_undefined_pd (),
756c5857
AI
110 (__mmask8) -1, __R);
111}
112
113extern __inline __m512d
114__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
115_mm512_mask_rcp28_round_pd (__m512d __W, __mmask8 __U, __m512d __A, int __R)
116{
117 return (__m512d) __builtin_ia32_rcp28pd_mask ((__v8df) __A,
118 (__v8df) __W,
119 (__mmask8) __U, __R);
120}
121
122extern __inline __m512d
123__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
124_mm512_maskz_rcp28_round_pd (__mmask8 __U, __m512d __A, int __R)
125{
126 return (__m512d) __builtin_ia32_rcp28pd_mask ((__v8df) __A,
127 (__v8df) _mm512_setzero_pd (),
128 (__mmask8) __U, __R);
129}
130
131extern __inline __m512
132__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
133_mm512_rcp28_round_ps (__m512 __A, int __R)
134{
756c5857 135 return (__m512) __builtin_ia32_rcp28ps_mask ((__v16sf) __A,
41602390 136 (__v16sf) _mm512_undefined_ps (),
756c5857
AI
137 (__mmask16) -1, __R);
138}
139
140extern __inline __m512
141__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
142_mm512_mask_rcp28_round_ps (__m512 __W, __mmask16 __U, __m512 __A, int __R)
143{
144 return (__m512) __builtin_ia32_rcp28ps_mask ((__v16sf) __A,
145 (__v16sf) __W,
146 (__mmask16) __U, __R);
147}
148
149extern __inline __m512
150__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
151_mm512_maskz_rcp28_round_ps (__mmask16 __U, __m512 __A, int __R)
152{
153 return (__m512) __builtin_ia32_rcp28ps_mask ((__v16sf) __A,
154 (__v16sf) _mm512_setzero_ps (),
155 (__mmask16) __U, __R);
156}
157
c56a42b9
KY
158extern __inline __m128d
159__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
160_mm_rcp28_round_sd (__m128d __A, __m128d __B, int __R)
161{
df62b4af
IT
162 return (__m128d) __builtin_ia32_rcp28sd_round ((__v2df) __B,
163 (__v2df) __A,
c56a42b9
KY
164 __R);
165}
166
93103603
SP
167extern __inline __m128d
168__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
169_mm_mask_rcp28_round_sd (__m128d __W, __mmask8 __U, __m128d __A,
170 __m128d __B, int __R)
171{
172 return (__m128d) __builtin_ia32_rcp28sd_mask_round ((__v2df) __B,
173 (__v2df) __A,
174 (__v2df) __W,
175 __U,
176 __R);
177}
178
179extern __inline __m128d
180__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
181_mm_maskz_rcp28_round_sd (__mmask8 __U, __m128d __A, __m128d __B, int __R)
182{
183 return (__m128d) __builtin_ia32_rcp28sd_mask_round ((__v2df) __B,
184 (__v2df) __A,
185 (__v2df)
186 _mm_setzero_pd (),
187 __U,
188 __R);
189}
190
c56a42b9
KY
191extern __inline __m128
192__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
193_mm_rcp28_round_ss (__m128 __A, __m128 __B, int __R)
194{
df62b4af
IT
195 return (__m128) __builtin_ia32_rcp28ss_round ((__v4sf) __B,
196 (__v4sf) __A,
c56a42b9
KY
197 __R);
198}
199
93103603
SP
200extern __inline __m128
201__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
202_mm_mask_rcp28_round_ss (__m128 __W, __mmask8 __U, __m128 __A,
203 __m128 __B, int __R)
204{
205 return (__m128) __builtin_ia32_rcp28ss_mask_round ((__v4sf) __B,
206 (__v4sf) __A,
207 (__v4sf) __W,
208 __U,
209 __R);
210}
211
212extern __inline __m128
213__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
214_mm_maskz_rcp28_round_ss (__mmask8 __U, __m128 __A, __m128 __B, int __R)
215{
216 return (__m128) __builtin_ia32_rcp28ss_mask_round ((__v4sf) __B,
217 (__v4sf) __A,
218 (__v4sf)
219 _mm_setzero_ps (),
220 __U,
221 __R);
222}
223
756c5857
AI
224extern __inline __m512d
225__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
226_mm512_rsqrt28_round_pd (__m512d __A, int __R)
227{
756c5857 228 return (__m512d) __builtin_ia32_rsqrt28pd_mask ((__v8df) __A,
41602390 229 (__v8df) _mm512_undefined_pd (),
756c5857
AI
230 (__mmask8) -1, __R);
231}
232
233extern __inline __m512d
234__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
235_mm512_mask_rsqrt28_round_pd (__m512d __W, __mmask8 __U, __m512d __A, int __R)
236{
237 return (__m512d) __builtin_ia32_rsqrt28pd_mask ((__v8df) __A,
238 (__v8df) __W,
239 (__mmask8) __U, __R);
240}
241
242extern __inline __m512d
243__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
244_mm512_maskz_rsqrt28_round_pd (__mmask8 __U, __m512d __A, int __R)
245{
246 return (__m512d) __builtin_ia32_rsqrt28pd_mask ((__v8df) __A,
247 (__v8df) _mm512_setzero_pd (),
248 (__mmask8) __U, __R);
249}
250
251extern __inline __m512
252__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
253_mm512_rsqrt28_round_ps (__m512 __A, int __R)
254{
756c5857 255 return (__m512) __builtin_ia32_rsqrt28ps_mask ((__v16sf) __A,
41602390 256 (__v16sf) _mm512_undefined_ps (),
756c5857
AI
257 (__mmask16) -1, __R);
258}
259
260extern __inline __m512
261__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
262_mm512_mask_rsqrt28_round_ps (__m512 __W, __mmask16 __U, __m512 __A, int __R)
263{
264 return (__m512) __builtin_ia32_rsqrt28ps_mask ((__v16sf) __A,
265 (__v16sf) __W,
266 (__mmask16) __U, __R);
267}
268
269extern __inline __m512
270__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
271_mm512_maskz_rsqrt28_round_ps (__mmask16 __U, __m512 __A, int __R)
272{
273 return (__m512) __builtin_ia32_rsqrt28ps_mask ((__v16sf) __A,
274 (__v16sf) _mm512_setzero_ps (),
275 (__mmask16) __U, __R);
276}
c56a42b9
KY
277
278extern __inline __m128d
279__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
280_mm_rsqrt28_round_sd (__m128d __A, __m128d __B, int __R)
281{
df62b4af
IT
282 return (__m128d) __builtin_ia32_rsqrt28sd_round ((__v2df) __B,
283 (__v2df) __A,
c56a42b9
KY
284 __R);
285}
286
93103603
SP
287extern __inline __m128d
288__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
289_mm_mask_rsqrt28_round_sd (__m128d __W, __mmask8 __U, __m128d __A,
290 __m128d __B, int __R)
291{
292 return (__m128d) __builtin_ia32_rsqrt28sd_mask_round ((__v2df) __B,
293 (__v2df) __A,
294 (__v2df) __W,
295 __U,
296 __R);
297}
298
299extern __inline __m128d
300__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
301_mm_maskz_rsqrt28_round_sd (__mmask8 __U, __m128d __A, __m128d __B, int __R)
302{
303 return (__m128d) __builtin_ia32_rsqrt28sd_mask_round ((__v2df) __B,
304 (__v2df) __A,
305 (__v2df)
306 _mm_setzero_pd (),
307 __U,
308 __R);
309}
310
c56a42b9
KY
311extern __inline __m128
312__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
313_mm_rsqrt28_round_ss (__m128 __A, __m128 __B, int __R)
314{
df62b4af
IT
315 return (__m128) __builtin_ia32_rsqrt28ss_round ((__v4sf) __B,
316 (__v4sf) __A,
c56a42b9
KY
317 __R);
318}
319
93103603
SP
320extern __inline __m128
321__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
322_mm_mask_rsqrt28_round_ss (__m128 __W, __mmask8 __U, __m128 __A,
323 __m128 __B, int __R)
324{
325 return (__m128) __builtin_ia32_rsqrt28ss_mask_round ((__v4sf) __B,
326 (__v4sf) __A,
327 (__v4sf) __W,
328 __U,
329 __R);
330}
331
332extern __inline __m128
333__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
334_mm_maskz_rsqrt28_round_ss (__mmask8 __U, __m128 __A, __m128 __B, int __R)
335{
336 return (__m128) __builtin_ia32_rsqrt28ss_mask_round ((__v4sf) __B,
337 (__v4sf) __A,
338 (__v4sf)
339 _mm_setzero_ps (),
340 __U,
341 __R);
342}
343
756c5857
AI
344#else
345#define _mm512_exp2a23_round_pd(A, C) \
346 __builtin_ia32_exp2pd_mask(A, (__v8df)_mm512_setzero_pd(), -1, C)
347
348#define _mm512_mask_exp2a23_round_pd(W, U, A, C) \
349 __builtin_ia32_exp2pd_mask(A, W, U, C)
350
351#define _mm512_maskz_exp2a23_round_pd(U, A, C) \
352 __builtin_ia32_exp2pd_mask(A, (__v8df)_mm512_setzero_pd(), U, C)
353
354#define _mm512_exp2a23_round_ps(A, C) \
355 __builtin_ia32_exp2ps_mask(A, (__v16sf)_mm512_setzero_ps(), -1, C)
356
357#define _mm512_mask_exp2a23_round_ps(W, U, A, C) \
358 __builtin_ia32_exp2ps_mask(A, W, U, C)
359
360#define _mm512_maskz_exp2a23_round_ps(U, A, C) \
361 __builtin_ia32_exp2ps_mask(A, (__v16sf)_mm512_setzero_ps(), U, C)
362
363#define _mm512_rcp28_round_pd(A, C) \
364 __builtin_ia32_rcp28pd_mask(A, (__v8df)_mm512_setzero_pd(), -1, C)
365
366#define _mm512_mask_rcp28_round_pd(W, U, A, C) \
367 __builtin_ia32_rcp28pd_mask(A, W, U, C)
368
369#define _mm512_maskz_rcp28_round_pd(U, A, C) \
370 __builtin_ia32_rcp28pd_mask(A, (__v8df)_mm512_setzero_pd(), U, C)
371
372#define _mm512_rcp28_round_ps(A, C) \
373 __builtin_ia32_rcp28ps_mask(A, (__v16sf)_mm512_setzero_ps(), -1, C)
374
375#define _mm512_mask_rcp28_round_ps(W, U, A, C) \
376 __builtin_ia32_rcp28ps_mask(A, W, U, C)
377
378#define _mm512_maskz_rcp28_round_ps(U, A, C) \
379 __builtin_ia32_rcp28ps_mask(A, (__v16sf)_mm512_setzero_ps(), U, C)
380
381#define _mm512_rsqrt28_round_pd(A, C) \
382 __builtin_ia32_rsqrt28pd_mask(A, (__v8df)_mm512_setzero_pd(), -1, C)
383
384#define _mm512_mask_rsqrt28_round_pd(W, U, A, C) \
385 __builtin_ia32_rsqrt28pd_mask(A, W, U, C)
386
387#define _mm512_maskz_rsqrt28_round_pd(U, A, C) \
388 __builtin_ia32_rsqrt28pd_mask(A, (__v8df)_mm512_setzero_pd(), U, C)
389
390#define _mm512_rsqrt28_round_ps(A, C) \
391 __builtin_ia32_rsqrt28ps_mask(A, (__v16sf)_mm512_setzero_ps(), -1, C)
392
393#define _mm512_mask_rsqrt28_round_ps(W, U, A, C) \
394 __builtin_ia32_rsqrt28ps_mask(A, W, U, C)
395
396#define _mm512_maskz_rsqrt28_round_ps(U, A, C) \
397 __builtin_ia32_rsqrt28ps_mask(A, (__v16sf)_mm512_setzero_ps(), U, C)
c56a42b9
KY
398
399#define _mm_rcp28_round_sd(A, B, R) \
400 __builtin_ia32_rcp28sd_round(A, B, R)
401
93103603
SP
402#define _mm_mask_rcp28_round_sd(W, U, A, B, R) \
403 __builtin_ia32_rcp28sd_mask_round ((A), (B), (W), (U), (R))
404
405#define _mm_maskz_rcp28_round_sd(U, A, B, R) \
406 __builtin_ia32_rcp28sd_mask_round ((A), (B), (__v2df) _mm_setzero_pd (), \
407 (U), (R))
408
c56a42b9
KY
409#define _mm_rcp28_round_ss(A, B, R) \
410 __builtin_ia32_rcp28ss_round(A, B, R)
411
93103603
SP
412#define _mm_mask_rcp28_round_ss(W, U, A, B, R) \
413 __builtin_ia32_rcp28ss_mask_round ((A), (B), (W), (U), (R))
414
415#define _mm_maskz_rcp28_round_ss(U, A, B, R) \
416 __builtin_ia32_rcp28ss_mask_round ((A), (B), (__v4sf) _mm_setzero_ps (), \
417 (U), (R))
418
c56a42b9
KY
419#define _mm_rsqrt28_round_sd(A, B, R) \
420 __builtin_ia32_rsqrt28sd_round(A, B, R)
421
93103603
SP
422#define _mm_mask_rsqrt28_round_sd(W, U, A, B, R) \
423 __builtin_ia32_rsqrt28sd_mask_round ((A), (B), (W), (U), (R))
424
425#define _mm_maskz_rsqrt28_round_sd(U, A, B, R) \
426 __builtin_ia32_rsqrt28sd_mask_round ((A), (B), (__v2df) _mm_setzero_pd (),\
427 (U), (R))
428
c56a42b9
KY
429#define _mm_rsqrt28_round_ss(A, B, R) \
430 __builtin_ia32_rsqrt28ss_round(A, B, R)
431
93103603
SP
432#define _mm_mask_rsqrt28_round_ss(W, U, A, B, R) \
433 __builtin_ia32_rsqrt28ss_mask_round ((A), (B), (W), (U), (R))
434
435#define _mm_maskz_rsqrt28_round_ss(U, A, B, R) \
436 __builtin_ia32_rsqrt28ss_mask_round ((A), (B), (__v4sf) _mm_setzero_ps (),\
437 (U), (R))
438
756c5857
AI
439#endif
440
93103603
SP
441#define _mm_mask_rcp28_sd(W, U, A, B)\
442 _mm_mask_rcp28_round_sd ((W), (U), (A), (B), _MM_FROUND_CUR_DIRECTION)
443
444#define _mm_maskz_rcp28_sd(U, A, B)\
445 _mm_maskz_rcp28_round_sd ((U), (A), (B), _MM_FROUND_CUR_DIRECTION)
446
447#define _mm_mask_rcp28_ss(W, U, A, B)\
448 _mm_mask_rcp28_round_ss ((W), (U), (A), (B), _MM_FROUND_CUR_DIRECTION)
449
450#define _mm_maskz_rcp28_ss(U, A, B)\
451 _mm_maskz_rcp28_round_ss ((U), (A), (B), _MM_FROUND_CUR_DIRECTION)
452
453#define _mm_mask_rsqrt28_sd(W, U, A, B)\
454 _mm_mask_rsqrt28_round_sd ((W), (U), (A), (B), _MM_FROUND_CUR_DIRECTION)
455
456#define _mm_maskz_rsqrt28_sd(U, A, B)\
457 _mm_maskz_rsqrt28_round_sd ((U), (A), (B), _MM_FROUND_CUR_DIRECTION)
458
459#define _mm_mask_rsqrt28_ss(W, U, A, B)\
460 _mm_mask_rsqrt28_round_ss ((W), (U), (A), (B), _MM_FROUND_CUR_DIRECTION)
461
462#define _mm_maskz_rsqrt28_ss(U, A, B)\
463 _mm_maskz_rsqrt28_round_ss ((U), (A), (B), _MM_FROUND_CUR_DIRECTION)
464
756c5857
AI
465#define _mm512_exp2a23_pd(A) \
466 _mm512_exp2a23_round_pd(A, _MM_FROUND_CUR_DIRECTION)
467
468#define _mm512_mask_exp2a23_pd(W, U, A) \
469 _mm512_mask_exp2a23_round_pd(W, U, A, _MM_FROUND_CUR_DIRECTION)
470
471#define _mm512_maskz_exp2a23_pd(U, A) \
472 _mm512_maskz_exp2a23_round_pd(U, A, _MM_FROUND_CUR_DIRECTION)
473
474#define _mm512_exp2a23_ps(A) \
475 _mm512_exp2a23_round_ps(A, _MM_FROUND_CUR_DIRECTION)
476
477#define _mm512_mask_exp2a23_ps(W, U, A) \
478 _mm512_mask_exp2a23_round_ps(W, U, A, _MM_FROUND_CUR_DIRECTION)
479
480#define _mm512_maskz_exp2a23_ps(U, A) \
481 _mm512_maskz_exp2a23_round_ps(U, A, _MM_FROUND_CUR_DIRECTION)
482
483#define _mm512_rcp28_pd(A) \
484 _mm512_rcp28_round_pd(A, _MM_FROUND_CUR_DIRECTION)
485
486#define _mm512_mask_rcp28_pd(W, U, A) \
487 _mm512_mask_rcp28_round_pd(W, U, A, _MM_FROUND_CUR_DIRECTION)
488
489#define _mm512_maskz_rcp28_pd(U, A) \
490 _mm512_maskz_rcp28_round_pd(U, A, _MM_FROUND_CUR_DIRECTION)
491
492#define _mm512_rcp28_ps(A) \
493 _mm512_rcp28_round_ps(A, _MM_FROUND_CUR_DIRECTION)
494
495#define _mm512_mask_rcp28_ps(W, U, A) \
496 _mm512_mask_rcp28_round_ps(W, U, A, _MM_FROUND_CUR_DIRECTION)
497
498#define _mm512_maskz_rcp28_ps(U, A) \
499 _mm512_maskz_rcp28_round_ps(U, A, _MM_FROUND_CUR_DIRECTION)
500
501#define _mm512_rsqrt28_pd(A) \
502 _mm512_rsqrt28_round_pd(A, _MM_FROUND_CUR_DIRECTION)
503
504#define _mm512_mask_rsqrt28_pd(W, U, A) \
505 _mm512_mask_rsqrt28_round_pd(W, U, A, _MM_FROUND_CUR_DIRECTION)
506
507#define _mm512_maskz_rsqrt28_pd(U, A) \
508 _mm512_maskz_rsqrt28_round_pd(U, A, _MM_FROUND_CUR_DIRECTION)
509
510#define _mm512_rsqrt28_ps(A) \
511 _mm512_rsqrt28_round_ps(A, _MM_FROUND_CUR_DIRECTION)
512
513#define _mm512_mask_rsqrt28_ps(W, U, A) \
514 _mm512_mask_rsqrt28_round_ps(W, U, A, _MM_FROUND_CUR_DIRECTION)
515
516#define _mm512_maskz_rsqrt28_ps(U, A) \
517 _mm512_maskz_rsqrt28_round_ps(U, A, _MM_FROUND_CUR_DIRECTION)
518
c56a42b9 519#define _mm_rcp28_sd(A, B) \
df62b4af 520 __builtin_ia32_rcp28sd_round(B, A, _MM_FROUND_CUR_DIRECTION)
c56a42b9
KY
521
522#define _mm_rcp28_ss(A, B) \
df62b4af 523 __builtin_ia32_rcp28ss_round(B, A, _MM_FROUND_CUR_DIRECTION)
c56a42b9
KY
524
525#define _mm_rsqrt28_sd(A, B) \
df62b4af 526 __builtin_ia32_rsqrt28sd_round(B, A, _MM_FROUND_CUR_DIRECTION)
c56a42b9
KY
527
528#define _mm_rsqrt28_ss(A, B) \
df62b4af 529 __builtin_ia32_rsqrt28ss_round(B, A, _MM_FROUND_CUR_DIRECTION)
c56a42b9 530
756c5857
AI
531#ifdef __DISABLE_AVX512ER__
532#undef __DISABLE_AVX512ER__
533#pragma GCC pop_options
534#endif /* __DISABLE_AVX512ER__ */
535
536#endif /* _AVX512ERINTRIN_H_INCLUDED */
This page took 3.027191 seconds and 5 git commands to generate.