diff options
| author | Igor Breger <igor.breger@intel.com> | 2016-02-08 12:36:48 +0000 |
|---|---|---|
| committer | Igor Breger <igor.breger@intel.com> | 2016-02-08 12:36:48 +0000 |
| commit | 9c2a0bfa13957e09ff376ba140611cb94855239e (patch) | |
| tree | fcf9eac6a20ffe540d9b889cae3787b8f14bf28b /clang/lib/Headers | |
| parent | 41b4973329490554594745b65913832ce5ae31ba (diff) | |
| download | bcm5719-llvm-9c2a0bfa13957e09ff376ba140611cb94855239e.tar.gz bcm5719-llvm-9c2a0bfa13957e09ff376ba140611cb94855239e.zip | |
AVX512: Change builtin function name for scalar intrinsics. Add "mask" to function name to reflect the function behavior.
Differential Revision: http://reviews.llvm.org/D16957
llvm-svn: 260088
Diffstat (limited to 'clang/lib/Headers')
| -rw-r--r-- | clang/lib/Headers/avx512erintrin.h | 24 | ||||
| -rw-r--r-- | clang/lib/Headers/avx512fintrin.h | 128 |
2 files changed, 76 insertions, 76 deletions
diff --git a/clang/lib/Headers/avx512erintrin.h b/clang/lib/Headers/avx512erintrin.h index 40a912189e5..f5b0abea1e7 100644 --- a/clang/lib/Headers/avx512erintrin.h +++ b/clang/lib/Headers/avx512erintrin.h @@ -126,19 +126,19 @@ _mm512_maskz_rsqrt28_round_ps((M), (A), _MM_FROUND_CUR_DIRECTION) #define _mm_rsqrt28_round_ss(A, B, R) __extension__ ({ \ - (__m128)__builtin_ia32_rsqrt28ss_round((__v4sf)(__m128)(A), \ + (__m128)__builtin_ia32_rsqrt28ss_round_mask((__v4sf)(__m128)(A), \ (__v4sf)(__m128)(B), \ (__v4sf)_mm_setzero_ps(), \ (__mmask8)-1, (R)); }) #define _mm_mask_rsqrt28_round_ss(S, M, A, B, R) __extension__ ({ \ - (__m128)__builtin_ia32_rsqrt28ss_round((__v4sf)(__m128)(A), \ + (__m128)__builtin_ia32_rsqrt28ss_round_mask((__v4sf)(__m128)(A), \ (__v4sf)(__m128)(B), \ (__v4sf)(__m128)(S), \ (__mmask8)(M), (R)); }) #define _mm_maskz_rsqrt28_round_ss(M, A, B, R) __extension__ ({ \ - (__m128)__builtin_ia32_rsqrt28ss_round((__v4sf)(__m128)(A), \ + (__m128)__builtin_ia32_rsqrt28ss_round_mask((__v4sf)(__m128)(A), \ (__v4sf)(__m128)(B), \ (__v4sf)_mm_setzero_ps(), \ (__mmask8)(M), (R)); }) @@ -153,19 +153,19 @@ _mm_maskz_rsqrt28_round_ss((M), (A), (B), _MM_FROUND_CUR_DIRECTION) #define _mm_rsqrt28_round_sd(A, B, R) __extension__ ({ \ - (__m128d)__builtin_ia32_rsqrt28sd_round((__v2df)(__m128d)(A), \ + (__m128d)__builtin_ia32_rsqrt28sd_round_mask((__v2df)(__m128d)(A), \ (__v2df)(__m128d)(B), \ (__v2df)_mm_setzero_pd(), \ (__mmask8)-1, (R)); }) #define _mm_mask_rsqrt28_round_sd(S, M, A, B, R) __extension__ ({ \ - (__m128d)__builtin_ia32_rsqrt28sd_round((__v2df)(__m128d)(A), \ + (__m128d)__builtin_ia32_rsqrt28sd_round_mask((__v2df)(__m128d)(A), \ (__v2df)(__m128d)(B), \ (__v2df)(__m128d)(S), \ (__mmask8)(M), (R)); }) #define _mm_maskz_rsqrt28_round_sd(M, A, B, R) __extension__ ({ \ - (__m128d)__builtin_ia32_rsqrt28sd_round((__v2df)(__m128d)(A), \ + (__m128d)__builtin_ia32_rsqrt28sd_round_mask((__v2df)(__m128d)(A), \ (__v2df)(__m128d)(B), \ (__v2df)_mm_setzero_pd(), \ (__mmask8)(M), (R)); }) @@ -229,19 +229,19 @@ _mm512_maskz_rcp28_round_ps((M), (A), _MM_FROUND_CUR_DIRECTION) #define _mm_rcp28_round_ss(A, B, R) __extension__ ({ \ - (__m128)__builtin_ia32_rcp28ss_round((__v4sf)(__m128)(A), \ + (__m128)__builtin_ia32_rcp28ss_round_mask((__v4sf)(__m128)(A), \ (__v4sf)(__m128)(B), \ (__v4sf)_mm_setzero_ps(), \ (__mmask8)-1, (R)); }) #define _mm_mask_rcp28_round_ss(S, M, A, B, R) __extension__ ({ \ - (__m128)__builtin_ia32_rcp28ss_round((__v4sf)(__m128)(A), \ + (__m128)__builtin_ia32_rcp28ss_round_mask((__v4sf)(__m128)(A), \ (__v4sf)(__m128)(B), \ (__v4sf)(__m128)(S), \ (__mmask8)(M), (R)); }) #define _mm_maskz_rcp28_round_ss(M, A, B, R) __extension__ ({ \ - (__m128)__builtin_ia32_rcp28ss_round((__v4sf)(__m128)(A), \ + (__m128)__builtin_ia32_rcp28ss_round_mask((__v4sf)(__m128)(A), \ (__v4sf)(__m128)(B), \ (__v4sf)_mm_setzero_ps(), \ (__mmask8)(M), (R)); }) @@ -256,19 +256,19 @@ _mm_maskz_rcp28_round_ss((M), (A), (B), _MM_FROUND_CUR_DIRECTION) #define _mm_rcp28_round_sd(A, B, R) __extension__ ({ \ - (__m128d)__builtin_ia32_rcp28sd_round((__v2df)(__m128d)(A), \ + (__m128d)__builtin_ia32_rcp28sd_round_mask((__v2df)(__m128d)(A), \ (__v2df)(__m128d)(B), \ (__v2df)_mm_setzero_pd(), \ (__mmask8)-1, (R)); }) #define _mm_mask_rcp28_round_sd(S, M, A, B, R) __extension__ ({ \ - (__m128d)__builtin_ia32_rcp28sd_round((__v2df)(__m128d)(A), \ + (__m128d)__builtin_ia32_rcp28sd_round_mask((__v2df)(__m128d)(A), \ (__v2df)(__m128d)(B), \ (__v2df)(__m128d)(S), \ (__mmask8)(M), (R)); }) #define _mm_maskz_rcp28_round_sd(M, A, B, R) __extension__ ({ \ - (__m128d)__builtin_ia32_rcp28sd_round((__v2df)(__m128d)(A), \ + (__m128d)__builtin_ia32_rcp28sd_round_mask((__v2df)(__m128d)(A), \ (__v2df)(__m128d)(B), \ (__v2df)_mm_setzero_pd(), \ (__mmask8)(M), (R)); }) diff --git a/clang/lib/Headers/avx512fintrin.h b/clang/lib/Headers/avx512fintrin.h index 5a976ccb8da..30dbd82c2c3 100644 --- a/clang/lib/Headers/avx512fintrin.h +++ b/clang/lib/Headers/avx512fintrin.h @@ -569,7 +569,7 @@ _mm512_max_ps(__m512 __A, __m512 __B) static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_mask_max_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { - return (__m128) __builtin_ia32_maxss_round ((__v4sf) __A, + return (__m128) __builtin_ia32_maxss_round_mask ((__v4sf) __A, (__v4sf) __B, (__v4sf) __W, (__mmask8) __U, @@ -578,7 +578,7 @@ _mm_mask_max_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_maskz_max_ss(__mmask8 __U,__m128 __A, __m128 __B) { - return (__m128) __builtin_ia32_maxss_round ((__v4sf) __A, + return (__m128) __builtin_ia32_maxss_round_mask ((__v4sf) __A, (__v4sf) __B, (__v4sf) _mm_setzero_ps (), (__mmask8) __U, @@ -586,20 +586,20 @@ _mm_maskz_max_ss(__mmask8 __U,__m128 __A, __m128 __B) { } #define _mm_max_round_ss(__A, __B, __R) __extension__ ({ \ - (__m128) __builtin_ia32_maxss_round ((__v4sf) __A, (__v4sf) __B, \ + (__m128) __builtin_ia32_maxss_round_mask ((__v4sf) __A, (__v4sf) __B, \ (__v4sf) _mm_setzero_ps(), (__mmask8) -1, __R); }) #define _mm_mask_max_round_ss(__W, __U, __A, __B, __R) __extension__ ({ \ - (__m128) __builtin_ia32_maxss_round ((__v4sf) __A, (__v4sf) __B, \ + (__m128) __builtin_ia32_maxss_round_mask ((__v4sf) __A, (__v4sf) __B, \ (__v4sf) __W, (__mmask8) __U,__R); }) #define _mm_maskz_max_round_ss(__U, __A, __B, __R) __extension__ ({ \ - (__m128) __builtin_ia32_maxss_round ((__v4sf) __A, (__v4sf) __B, \ + (__m128) __builtin_ia32_maxss_round_mask ((__v4sf) __A, (__v4sf) __B, \ (__v4sf) _mm_setzero_ps(), (__mmask8) __U,__R); }) static __inline__ __m128d __DEFAULT_FN_ATTRS _mm_mask_max_sd(__m128d __W, __mmask8 __U,__m128d __A, __m128d __B) { - return (__m128d) __builtin_ia32_maxsd_round ((__v2df) __A, + return (__m128d) __builtin_ia32_maxsd_round_mask ((__v2df) __A, (__v2df) __B, (__v2df) __W, (__mmask8) __U, @@ -608,7 +608,7 @@ _mm_mask_max_sd(__m128d __W, __mmask8 __U,__m128d __A, __m128d __B) { static __inline__ __m128d __DEFAULT_FN_ATTRS _mm_maskz_max_sd(__mmask8 __U,__m128d __A, __m128d __B) { - return (__m128d) __builtin_ia32_maxsd_round ((__v2df) __A, + return (__m128d) __builtin_ia32_maxsd_round_mask ((__v2df) __A, (__v2df) __B, (__v2df) _mm_setzero_pd (), (__mmask8) __U, @@ -616,15 +616,15 @@ _mm_maskz_max_sd(__mmask8 __U,__m128d __A, __m128d __B) { } #define _mm_max_round_sd(__A, __B, __R) __extension__ ({ \ - (__m128d) __builtin_ia32_maxsd_round ((__v2df) __A, (__v2df) __B, \ + (__m128d) __builtin_ia32_maxsd_round_mask ((__v2df) __A, (__v2df) __B, \ (__v2df) _mm_setzero_pd(), (__mmask8) -1, __R); }) #define _mm_mask_max_round_sd(__W, __U, __A, __B, __R) __extension__ ({ \ - (__m128d) __builtin_ia32_maxsd_round ((__v2df) __A, (__v2df) __B, \ + (__m128d) __builtin_ia32_maxsd_round_mask ((__v2df) __A, (__v2df) __B, \ (__v2df) __W, (__mmask8) __U,__R); }) #define _mm_maskz_max_round_sd(__U, __A, __B, __R) __extension__ ({ \ - (__m128d) __builtin_ia32_maxsd_round ((__v2df) __A, (__v2df) __B, \ + (__m128d) __builtin_ia32_maxsd_round_mask ((__v2df) __A, (__v2df) __B, \ (__v2df) _mm_setzero_pd(), (__mmask8) __U,__R); }) static __inline __m512i @@ -692,7 +692,7 @@ _mm512_min_ps(__m512 __A, __m512 __B) static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_mask_min_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { - return (__m128) __builtin_ia32_minss_round ((__v4sf) __A, + return (__m128) __builtin_ia32_minss_round_mask ((__v4sf) __A, (__v4sf) __B, (__v4sf) __W, (__mmask8) __U, @@ -701,7 +701,7 @@ _mm_mask_min_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_maskz_min_ss(__mmask8 __U,__m128 __A, __m128 __B) { - return (__m128) __builtin_ia32_minss_round ((__v4sf) __A, + return (__m128) __builtin_ia32_minss_round_mask ((__v4sf) __A, (__v4sf) __B, (__v4sf) _mm_setzero_ps (), (__mmask8) __U, @@ -709,20 +709,20 @@ _mm_maskz_min_ss(__mmask8 __U,__m128 __A, __m128 __B) { } #define _mm_min_round_ss(__A, __B, __R) __extension__ ({ \ - (__m128) __builtin_ia32_minss_round ((__v4sf) __A, (__v4sf) __B, \ + (__m128) __builtin_ia32_minss_round_mask ((__v4sf) __A, (__v4sf) __B, \ (__v4sf) _mm_setzero_ps(), (__mmask8) -1, __R); }) #define _mm_mask_min_round_ss(__W, __U, __A, __B, __R) __extension__ ({ \ - (__m128) __builtin_ia32_minss_round ((__v4sf) __A, (__v4sf) __B, \ + (__m128) __builtin_ia32_minss_round_mask ((__v4sf) __A, (__v4sf) __B, \ (__v4sf) __W, (__mmask8) __U,__R); }) #define _mm_maskz_min_round_ss(__U, __A, __B, __R) __extension__ ({ \ - (__m128) __builtin_ia32_minss_round ((__v4sf) __A, (__v4sf) __B, \ + (__m128) __builtin_ia32_minss_round_mask ((__v4sf) __A, (__v4sf) __B, \ (__v4sf) _mm_setzero_ps(), (__mmask8) __U,__R); }) static __inline__ __m128d __DEFAULT_FN_ATTRS _mm_mask_min_sd(__m128d __W, __mmask8 __U,__m128d __A, __m128d __B) { - return (__m128d) __builtin_ia32_minsd_round ((__v2df) __A, + return (__m128d) __builtin_ia32_minsd_round_mask ((__v2df) __A, (__v2df) __B, (__v2df) __W, (__mmask8) __U, @@ -731,7 +731,7 @@ _mm_mask_min_sd(__m128d __W, __mmask8 __U,__m128d __A, __m128d __B) { static __inline__ __m128d __DEFAULT_FN_ATTRS _mm_maskz_min_sd(__mmask8 __U,__m128d __A, __m128d __B) { - return (__m128d) __builtin_ia32_minsd_round ((__v2df) __A, + return (__m128d) __builtin_ia32_minsd_round_mask ((__v2df) __A, (__v2df) __B, (__v2df) _mm_setzero_pd (), (__mmask8) __U, @@ -739,15 +739,15 @@ _mm_maskz_min_sd(__mmask8 __U,__m128d __A, __m128d __B) { } #define _mm_min_round_sd(__A, __B, __R) __extension__ ({ \ - (__m128d) __builtin_ia32_minsd_round ((__v2df) __A, (__v2df) __B, \ + (__m128d) __builtin_ia32_minsd_round_mask ((__v2df) __A, (__v2df) __B, \ (__v2df) _mm_setzero_pd(), (__mmask8) -1, __R); }) #define _mm_mask_min_round_sd(__W, __U, __A, __B, __R) __extension__ ({ \ - (__m128d) __builtin_ia32_minsd_round ((__v2df) __A, (__v2df) __B, \ + (__m128d) __builtin_ia32_minsd_round_mask ((__v2df) __A, (__v2df) __B, \ (__v2df) __W, (__mmask8) __U,__R); }) #define _mm_maskz_min_round_sd(__U, __A, __B, __R) __extension__ ({ \ - (__m128d) __builtin_ia32_minsd_round ((__v2df) __A, (__v2df) __B, \ + (__m128d) __builtin_ia32_minsd_round_mask ((__v2df) __A, (__v2df) __B, \ (__v2df) _mm_setzero_pd(), (__mmask8) __U,__R); }) static __inline __m512i @@ -909,7 +909,7 @@ _mm512_rsqrt14_ps(__m512 __A) static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_rsqrt14_ss(__m128 __A, __m128 __B) { - return (__m128) __builtin_ia32_rsqrt14ss ((__v4sf) __A, + return (__m128) __builtin_ia32_rsqrt14ss_mask ((__v4sf) __A, (__v4sf) __B, (__v4sf) _mm_setzero_ps (), @@ -919,7 +919,7 @@ _mm_rsqrt14_ss(__m128 __A, __m128 __B) static __inline__ __m128d __DEFAULT_FN_ATTRS _mm_rsqrt14_sd(__m128d __A, __m128d __B) { - return (__m128d) __builtin_ia32_rsqrt14sd ((__v2df) __A, + return (__m128d) __builtin_ia32_rsqrt14sd_mask ((__v2df) __A, (__v2df) __B, (__v2df) _mm_setzero_pd (), @@ -946,7 +946,7 @@ _mm512_rcp14_ps(__m512 __A) static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_rcp14_ss(__m128 __A, __m128 __B) { - return (__m128) __builtin_ia32_rcp14ss ((__v4sf) __A, + return (__m128) __builtin_ia32_rcp14ss_mask ((__v4sf) __A, (__v4sf) __B, (__v4sf) _mm_setzero_ps (), @@ -956,7 +956,7 @@ _mm_rcp14_ss(__m128 __A, __m128 __B) static __inline__ __m128d __DEFAULT_FN_ATTRS _mm_rcp14_sd(__m128d __A, __m128d __B) { - return (__m128d) __builtin_ia32_rcp14sd ((__v2df) __A, + return (__m128d) __builtin_ia32_rcp14sd_mask ((__v2df) __A, (__v2df) __B, (__v2df) _mm_setzero_pd (), @@ -1019,7 +1019,7 @@ _mm512_abs_epi32(__m512i __A) static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_mask_add_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { - return (__m128) __builtin_ia32_addss_round ((__v4sf) __A, + return (__m128) __builtin_ia32_addss_round_mask ((__v4sf) __A, (__v4sf) __B, (__v4sf) __W, (__mmask8) __U, @@ -1028,7 +1028,7 @@ _mm_mask_add_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_maskz_add_ss(__mmask8 __U,__m128 __A, __m128 __B) { - return (__m128) __builtin_ia32_addss_round ((__v4sf) __A, + return (__m128) __builtin_ia32_addss_round_mask ((__v4sf) __A, (__v4sf) __B, (__v4sf) _mm_setzero_ps (), (__mmask8) __U, @@ -1036,20 +1036,20 @@ _mm_maskz_add_ss(__mmask8 __U,__m128 __A, __m128 __B) { } #define _mm_add_round_ss(__A, __B, __R) __extension__ ({ \ - (__m128) __builtin_ia32_addss_round ((__v4sf) __A, (__v4sf) __B, \ + (__m128) __builtin_ia32_addss_round_mask ((__v4sf) __A, (__v4sf) __B, \ (__v4sf) _mm_setzero_ps(), (__mmask8) -1, __R); }) #define _mm_mask_add_round_ss(__W, __U, __A, __B, __R) __extension__ ({ \ - (__m128) __builtin_ia32_addss_round ((__v4sf) __A, (__v4sf) __B, \ + (__m128) __builtin_ia32_addss_round_mask ((__v4sf) __A, (__v4sf) __B, \ (__v4sf) __W, (__mmask8) __U,__R); }) #define _mm_maskz_add_round_ss(__U, __A, __B, __R) __extension__ ({ \ - (__m128) __builtin_ia32_addss_round ((__v4sf) __A, (__v4sf) __B, \ + (__m128) __builtin_ia32_addss_round_mask ((__v4sf) __A, (__v4sf) __B, \ (__v4sf) _mm_setzero_ps(), (__mmask8) __U,__R); }) static __inline__ __m128d __DEFAULT_FN_ATTRS _mm_mask_add_sd(__m128d __W, __mmask8 __U,__m128d __A, __m128d __B) { - return (__m128d) __builtin_ia32_addsd_round ((__v2df) __A, + return (__m128d) __builtin_ia32_addsd_round_mask ((__v2df) __A, (__v2df) __B, (__v2df) __W, (__mmask8) __U, @@ -1058,22 +1058,22 @@ _mm_mask_add_sd(__m128d __W, __mmask8 __U,__m128d __A, __m128d __B) { static __inline__ __m128d __DEFAULT_FN_ATTRS _mm_maskz_add_sd(__mmask8 __U,__m128d __A, __m128d __B) { - return (__m128d) __builtin_ia32_addsd_round ((__v2df) __A, + return (__m128d) __builtin_ia32_addsd_round_mask ((__v2df) __A, (__v2df) __B, (__v2df) _mm_setzero_pd (), (__mmask8) __U, _MM_FROUND_CUR_DIRECTION); } #define _mm_add_round_sd(__A, __B, __R) __extension__ ({ \ - (__m128d) __builtin_ia32_addsd_round ((__v2df) __A, (__v2df) __B, \ + (__m128d) __builtin_ia32_addsd_round_mask ((__v2df) __A, (__v2df) __B, \ (__v2df) _mm_setzero_pd(), (__mmask8) -1, __R); }) #define _mm_mask_add_round_sd(__W, __U, __A, __B, __R) __extension__ ({ \ - (__m128d) __builtin_ia32_addsd_round ((__v2df) __A, (__v2df) __B, \ + (__m128d) __builtin_ia32_addsd_round_mask ((__v2df) __A, (__v2df) __B, \ (__v2df) __W, (__mmask8) __U,__R); }) #define _mm_maskz_add_round_sd(__U, __A, __B, __R) __extension__ ({ \ - (__m128d) __builtin_ia32_addsd_round ((__v2df) __A, (__v2df) __B, \ + (__m128d) __builtin_ia32_addsd_round_mask ((__v2df) __A, (__v2df) __B, \ (__v2df) _mm_setzero_pd(), (__mmask8) __U,__R); }) static __inline__ __m512d __DEFAULT_FN_ATTRS @@ -1138,7 +1138,7 @@ _mm512_maskz_add_ps(__mmask16 __U, __m512 __A, __m512 __B) { static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_mask_sub_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { - return (__m128) __builtin_ia32_subss_round ((__v4sf) __A, + return (__m128) __builtin_ia32_subss_round_mask ((__v4sf) __A, (__v4sf) __B, (__v4sf) __W, (__mmask8) __U, @@ -1147,27 +1147,27 @@ _mm_mask_sub_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_maskz_sub_ss(__mmask8 __U,__m128 __A, __m128 __B) { - return (__m128) __builtin_ia32_subss_round ((__v4sf) __A, + return (__m128) __builtin_ia32_subss_round_mask ((__v4sf) __A, (__v4sf) __B, (__v4sf) _mm_setzero_ps (), (__mmask8) __U, _MM_FROUND_CUR_DIRECTION); } #define _mm_sub_round_ss(__A, __B, __R) __extension__ ({ \ - (__m128) __builtin_ia32_subss_round ((__v4sf) __A, (__v4sf) __B, \ + (__m128) __builtin_ia32_subss_round_mask ((__v4sf) __A, (__v4sf) __B, \ (__v4sf) _mm_setzero_ps(), (__mmask8) -1, __R); }) #define _mm_mask_sub_round_ss(__W, __U, __A, __B, __R) __extension__ ({ \ - (__m128) __builtin_ia32_subss_round ((__v4sf) __A, (__v4sf) __B, \ + (__m128) __builtin_ia32_subss_round_mask ((__v4sf) __A, (__v4sf) __B, \ (__v4sf) __W, (__mmask8) __U,__R); }) #define _mm_maskz_sub_round_ss(__U, __A, __B, __R) __extension__ ({ \ - (__m128) __builtin_ia32_subss_round ((__v4sf) __A, (__v4sf) __B, \ + (__m128) __builtin_ia32_subss_round_mask ((__v4sf) __A, (__v4sf) __B, \ (__v4sf) _mm_setzero_ps(), (__mmask8) __U,__R); }) static __inline__ __m128d __DEFAULT_FN_ATTRS _mm_mask_sub_sd(__m128d __W, __mmask8 __U,__m128d __A, __m128d __B) { - return (__m128d) __builtin_ia32_subsd_round ((__v2df) __A, + return (__m128d) __builtin_ia32_subsd_round_mask ((__v2df) __A, (__v2df) __B, (__v2df) __W, (__mmask8) __U, @@ -1176,7 +1176,7 @@ _mm_mask_sub_sd(__m128d __W, __mmask8 __U,__m128d __A, __m128d __B) { static __inline__ __m128d __DEFAULT_FN_ATTRS _mm_maskz_sub_sd(__mmask8 __U,__m128d __A, __m128d __B) { - return (__m128d) __builtin_ia32_subsd_round ((__v2df) __A, + return (__m128d) __builtin_ia32_subsd_round_mask ((__v2df) __A, (__v2df) __B, (__v2df) _mm_setzero_pd (), (__mmask8) __U, @@ -1184,15 +1184,15 @@ _mm_maskz_sub_sd(__mmask8 __U,__m128d __A, __m128d __B) { } #define _mm_sub_round_sd(__A, __B, __R) __extension__ ({ \ - (__m128d) __builtin_ia32_subsd_round ((__v2df) __A, (__v2df) __B, \ + (__m128d) __builtin_ia32_subsd_round_mask ((__v2df) __A, (__v2df) __B, \ (__v2df) _mm_setzero_pd(), (__mmask8) -1, __R); }) #define _mm_mask_sub_round_sd(__W, __U, __A, __B, __R) __extension__ ({ \ - (__m128d) __builtin_ia32_subsd_round ((__v2df) __A, (__v2df) __B, \ + (__m128d) __builtin_ia32_subsd_round_mask ((__v2df) __A, (__v2df) __B, \ (__v2df) __W, (__mmask8) __U,__R); }) #define _mm_maskz_sub_round_sd(__U, __A, __B, __R) __extension__ ({ \ - (__m128d) __builtin_ia32_subsd_round ((__v2df) __A, (__v2df) __B, \ + (__m128d) __builtin_ia32_subsd_round_mask ((__v2df) __A, (__v2df) __B, \ (__v2df) _mm_setzero_pd(), (__mmask8) __U,__R); }) static __inline__ __m512d __DEFAULT_FN_ATTRS @@ -1259,7 +1259,7 @@ _mm512_maskz_sub_ps(__mmask16 __U, __m512 __A, __m512 __B) { static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_mask_mul_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { - return (__m128) __builtin_ia32_mulss_round ((__v4sf) __A, + return (__m128) __builtin_ia32_mulss_round_mask ((__v4sf) __A, (__v4sf) __B, (__v4sf) __W, (__mmask8) __U, @@ -1268,27 +1268,27 @@ _mm_mask_mul_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_maskz_mul_ss(__mmask8 __U,__m128 __A, __m128 __B) { - return (__m128) __builtin_ia32_mulss_round ((__v4sf) __A, + return (__m128) __builtin_ia32_mulss_round_mask ((__v4sf) __A, (__v4sf) __B, (__v4sf) _mm_setzero_ps (), (__mmask8) __U, _MM_FROUND_CUR_DIRECTION); } #define _mm_mul_round_ss(__A, __B, __R) __extension__ ({ \ - (__m128) __builtin_ia32_mulss_round ((__v4sf) __A, (__v4sf) __B, \ + (__m128) __builtin_ia32_mulss_round_mask ((__v4sf) __A, (__v4sf) __B, \ (__v4sf) _mm_setzero_ps(), (__mmask8) -1, __R); }) #define _mm_mask_mul_round_ss(__W, __U, __A, __B, __R) __extension__ ({ \ - (__m128) __builtin_ia32_mulss_round ((__v4sf) __A, (__v4sf) __B, \ + (__m128) __builtin_ia32_mulss_round_mask ((__v4sf) __A, (__v4sf) __B, \ (__v4sf) __W, (__mmask8) __U,__R); }) #define _mm_maskz_mul_round_ss(__U, __A, __B, __R) __extension__ ({ \ - (__m128) __builtin_ia32_mulss_round ((__v4sf) __A, (__v4sf) __B, \ + (__m128) __builtin_ia32_mulss_round_mask ((__v4sf) __A, (__v4sf) __B, \ (__v4sf) _mm_setzero_ps(), (__mmask8) __U,__R); }) static __inline__ __m128d __DEFAULT_FN_ATTRS _mm_mask_mul_sd(__m128d __W, __mmask8 __U,__m128d __A, __m128d __B) { - return (__m128d) __builtin_ia32_mulsd_round ((__v2df) __A, + return (__m128d) __builtin_ia32_mulsd_round_mask ((__v2df) __A, (__v2df) __B, (__v2df) __W, (__mmask8) __U, @@ -1297,7 +1297,7 @@ _mm_mask_mul_sd(__m128d __W, __mmask8 __U,__m128d __A, __m128d __B) { static __inline__ __m128d __DEFAULT_FN_ATTRS _mm_maskz_mul_sd(__mmask8 __U,__m128d __A, __m128d __B) { - return (__m128d) __builtin_ia32_mulsd_round ((__v2df) __A, + return (__m128d) __builtin_ia32_mulsd_round_mask ((__v2df) __A, (__v2df) __B, (__v2df) _mm_setzero_pd (), (__mmask8) __U, @@ -1305,15 +1305,15 @@ _mm_maskz_mul_sd(__mmask8 __U,__m128d __A, __m128d __B) { } #define _mm_mul_round_sd(__A, __B, __R) __extension__ ({ \ - (__m128d) __builtin_ia32_mulsd_round ((__v2df) __A, (__v2df) __B, \ + (__m128d) __builtin_ia32_mulsd_round_mask ((__v2df) __A, (__v2df) __B, \ (__v2df) _mm_setzero_pd(), (__mmask8) -1, __R); }) #define _mm_mask_mul_round_sd(__W, __U, __A, __B, __R) __extension__ ({ \ - (__m128d) __builtin_ia32_mulsd_round ((__v2df) __A, (__v2df) __B, \ + (__m128d) __builtin_ia32_mulsd_round_mask ((__v2df) __A, (__v2df) __B, \ (__v2df) __W, (__mmask8) __U,__R); }) #define _mm_maskz_mul_round_sd(__U, __A, __B, __R) __extension__ ({ \ - (__m128d) __builtin_ia32_mulsd_round ((__v2df) __A, (__v2df) __B, \ + (__m128d) __builtin_ia32_mulsd_round_mask ((__v2df) __A, (__v2df) __B, \ (__v2df) _mm_setzero_pd(), (__mmask8) __U,__R); }) static __inline__ __m512d __DEFAULT_FN_ATTRS @@ -1380,7 +1380,7 @@ _mm512_maskz_mul_ps(__mmask16 __U, __m512 __A, __m512 __B) { static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_mask_div_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { - return (__m128) __builtin_ia32_divss_round ((__v4sf) __A, + return (__m128) __builtin_ia32_divss_round_mask ((__v4sf) __A, (__v4sf) __B, (__v4sf) __W, (__mmask8) __U, @@ -1389,7 +1389,7 @@ _mm_mask_div_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_maskz_div_ss(__mmask8 __U,__m128 __A, __m128 __B) { - return (__m128) __builtin_ia32_divss_round ((__v4sf) __A, + return (__m128) __builtin_ia32_divss_round_mask ((__v4sf) __A, (__v4sf) __B, (__v4sf) _mm_setzero_ps (), (__mmask8) __U, @@ -1397,20 +1397,20 @@ _mm_maskz_div_ss(__mmask8 __U,__m128 __A, __m128 __B) { } #define _mm_div_round_ss(__A, __B, __R) __extension__ ({ \ - (__m128) __builtin_ia32_divss_round ((__v4sf) __A, (__v4sf) __B, \ + (__m128) __builtin_ia32_divss_round_mask ((__v4sf) __A, (__v4sf) __B, \ (__v4sf) _mm_setzero_ps(), (__mmask8) -1, __R); }) #define _mm_mask_div_round_ss(__W, __U, __A, __B, __R) __extension__ ({ \ - (__m128) __builtin_ia32_divss_round ((__v4sf) __A, (__v4sf) __B, \ + (__m128) __builtin_ia32_divss_round_mask ((__v4sf) __A, (__v4sf) __B, \ (__v4sf) __W, (__mmask8) __U,__R); }) #define _mm_maskz_div_round_ss(__U, __A, __B, __R) __extension__ ({ \ - (__m128) __builtin_ia32_divss_round ((__v4sf) __A, (__v4sf) __B, \ + (__m128) __builtin_ia32_divss_round_mask ((__v4sf) __A, (__v4sf) __B, \ (__v4sf) _mm_setzero_ps(), (__mmask8) __U,__R); }) static __inline__ __m128d __DEFAULT_FN_ATTRS _mm_mask_div_sd(__m128d __W, __mmask8 __U,__m128d __A, __m128d __B) { - return (__m128d) __builtin_ia32_divsd_round ((__v2df) __A, + return (__m128d) __builtin_ia32_divsd_round_mask ((__v2df) __A, (__v2df) __B, (__v2df) __W, (__mmask8) __U, @@ -1419,7 +1419,7 @@ _mm_mask_div_sd(__m128d __W, __mmask8 __U,__m128d __A, __m128d __B) { static __inline__ __m128d __DEFAULT_FN_ATTRS _mm_maskz_div_sd(__mmask8 __U,__m128d __A, __m128d __B) { - return (__m128d) __builtin_ia32_divsd_round ((__v2df) __A, + return (__m128d) __builtin_ia32_divsd_round_mask ((__v2df) __A, (__v2df) __B, (__v2df) _mm_setzero_pd (), (__mmask8) __U, @@ -1427,15 +1427,15 @@ _mm_maskz_div_sd(__mmask8 __U,__m128d __A, __m128d __B) { } #define _mm_div_round_sd(__A, __B, __R) __extension__ ({ \ - (__m128d) __builtin_ia32_divsd_round ((__v2df) __A, (__v2df) __B, \ + (__m128d) __builtin_ia32_divsd_round_mask ((__v2df) __A, (__v2df) __B, \ (__v2df) _mm_setzero_pd(), (__mmask8) -1, __R); }) #define _mm_mask_div_round_sd(__W, __U, __A, __B, __R) __extension__ ({ \ - (__m128d) __builtin_ia32_divsd_round ((__v2df) __A, (__v2df) __B, \ + (__m128d) __builtin_ia32_divsd_round_mask ((__v2df) __A, (__v2df) __B, \ (__v2df) __W, (__mmask8) __U,__R); }) #define _mm_maskz_div_round_sd(__U, __A, __B, __R) __extension__ ({ \ - (__m128d) __builtin_ia32_divsd_round ((__v2df) __A, (__v2df) __B, \ + (__m128d) __builtin_ia32_divsd_round_mask ((__v2df) __A, (__v2df) __B, \ (__v2df) _mm_setzero_pd(), (__mmask8) __U,__R); }) static __inline__ __m512d __DEFAULT_FN_ATTRS |

