diff options
| author | Craig Topper <craig.topper@intel.com> | 2018-06-30 01:32:14 +0000 | 
|---|---|---|
| committer | Craig Topper <craig.topper@intel.com> | 2018-06-30 01:32:14 +0000 | 
| commit | 0e9de769a0056f50fcf3d8ef59fbef6ca7b6ad66 (patch) | |
| tree | c6040c46ec8427e231af5c1d544f149b14e9eb74 /clang/lib | |
| parent | 59f2f38fe0b29a41781938a1b969b30df09b4352 (diff) | |
| download | bcm5719-llvm-0e9de769a0056f50fcf3d8ef59fbef6ca7b6ad66.tar.gz bcm5719-llvm-0e9de769a0056f50fcf3d8ef59fbef6ca7b6ad66.zip | |
[X86] Remove masking from the avx512 rotate builtins. Use a select builtin instead.
llvm-svn: 336036
Diffstat (limited to 'clang/lib')
| -rw-r--r-- | clang/lib/Headers/avx512fintrin.h | 147 | ||||
| -rw-r--r-- | clang/lib/Headers/avx512vlintrin.h | 336 | ||||
| -rw-r--r-- | clang/lib/Sema/SemaChecking.cpp | 24 | 
3 files changed, 201 insertions, 306 deletions
| diff --git a/clang/lib/Headers/avx512fintrin.h b/clang/lib/Headers/avx512fintrin.h index bfb645415e4..b20d48cf156 100644 --- a/clang/lib/Headers/avx512fintrin.h +++ b/clang/lib/Headers/avx512fintrin.h @@ -4938,59 +4938,45 @@ _mm512_maskz_cvtepu16_epi64(__mmask8 __U, __m128i __A)  static __inline__ __m512i __DEFAULT_FN_ATTRS  _mm512_rorv_epi32 (__m512i __A, __m512i __B)  { -  return (__m512i) __builtin_ia32_prorvd512_mask ((__v16si) __A, -              (__v16si) __B, -              (__v16si) -              _mm512_setzero_si512 (), -              (__mmask16) -1); +  return (__m512i)__builtin_ia32_prorvd512((__v16si)__A, (__v16si)__B);  }  static __inline__ __m512i __DEFAULT_FN_ATTRS  _mm512_mask_rorv_epi32 (__m512i __W, __mmask16 __U, __m512i __A, __m512i __B)  { -  return (__m512i) __builtin_ia32_prorvd512_mask ((__v16si) __A, -              (__v16si) __B, -              (__v16si) __W, -              (__mmask16) __U); +  return (__m512i)__builtin_ia32_selectd_512(__U, +                                           (__v16si)_mm512_rorv_epi32(__A, __B), +                                           (__v16si)__W);  }  static __inline__ __m512i __DEFAULT_FN_ATTRS  _mm512_maskz_rorv_epi32 (__mmask16 __U, __m512i __A, __m512i __B)  { -  return (__m512i) __builtin_ia32_prorvd512_mask ((__v16si) __A, -              (__v16si) __B, -              (__v16si) -              _mm512_setzero_si512 (), -              (__mmask16) __U); +  return (__m512i)__builtin_ia32_selectd_512(__U, +                                           (__v16si)_mm512_rorv_epi32(__A, __B), +                                           (__v16si)_mm512_setzero_si512());  }  static __inline__ __m512i __DEFAULT_FN_ATTRS  _mm512_rorv_epi64 (__m512i __A, __m512i __B)  { -  return (__m512i) __builtin_ia32_prorvq512_mask ((__v8di) __A, -              (__v8di) __B, -              (__v8di) -              _mm512_setzero_si512 (), -              (__mmask8) -1); +  return (__m512i)__builtin_ia32_prorvq512((__v8di)__A, (__v8di)__B);  }  static __inline__ __m512i __DEFAULT_FN_ATTRS  _mm512_mask_rorv_epi64 (__m512i __W, __mmask8 __U, __m512i __A, __m512i __B)  { -  return (__m512i) __builtin_ia32_prorvq512_mask ((__v8di) __A, -              (__v8di) __B, -              (__v8di) __W, -              (__mmask8) __U); +  return (__m512i)__builtin_ia32_selectq_512(__U, +                                            (__v8di)_mm512_rorv_epi64(__A, __B), +                                            (__v8di)__W);  }  static __inline__ __m512i __DEFAULT_FN_ATTRS  _mm512_maskz_rorv_epi64 (__mmask8 __U, __m512i __A, __m512i __B)  { -  return (__m512i) __builtin_ia32_prorvq512_mask ((__v8di) __A, -              (__v8di) __B, -              (__v8di) -              _mm512_setzero_si512 (), -              (__mmask8) __U); +  return (__m512i)__builtin_ia32_selectq_512(__U, +                                            (__v8di)_mm512_rorv_epi64(__A, __B), +                                            (__v8di)_mm512_setzero_si512());  } @@ -5036,119 +5022,100 @@ _mm512_maskz_rorv_epi64 (__mmask8 __U, __m512i __A, __m512i __B)                                           (__mmask8)(m))  #define _mm512_rol_epi32(a, b) \ -  (__m512i)__builtin_ia32_prold512_mask((__v16si)(__m512i)(a), (int)(b), \ -                                        (__v16si)_mm512_setzero_si512(), \ -                                        (__mmask16)-1) +  (__m512i)__builtin_ia32_prold512((__v16si)(__m512i)(a), (int)(b))  #define _mm512_mask_rol_epi32(W, U, a, b) \ -  (__m512i)__builtin_ia32_prold512_mask((__v16si)(__m512i)(a), (int)(b), \ -                                        (__v16si)(__m512i)(W), \ -                                        (__mmask16)(U)) +  (__m512i)__builtin_ia32_selectd_512((__mmask16)(U), \ +                                      (__v16si)_mm512_rol_epi32((a), (b)), \ +                                      (__v16si)(__m512i)(W))  #define _mm512_maskz_rol_epi32(U, a, b) \ -  (__m512i)__builtin_ia32_prold512_mask((__v16si)(__m512i)(a), (int)(b), \ -                                        (__v16si)_mm512_setzero_si512(), \ -                                        (__mmask16)(U)) +  (__m512i)__builtin_ia32_selectd_512((__mmask16)(U), \ +                                      (__v16si)_mm512_rol_epi32((a), (b)), \ +                                      (__v16si)_mm512_setzero_si512())  #define _mm512_rol_epi64(a, b) \ -  (__m512i)__builtin_ia32_prolq512_mask((__v8di)(__m512i)(a), (int)(b), \ -                                        (__v8di)_mm512_setzero_si512(), \ -                                        (__mmask8)-1) +  (__m512i)__builtin_ia32_prolq512((__v8di)(__m512i)(a), (int)(b))  #define _mm512_mask_rol_epi64(W, U, a, b) \ -  (__m512i)__builtin_ia32_prolq512_mask((__v8di)(__m512i)(a), (int)(b), \ -                                        (__v8di)(__m512i)(W), (__mmask8)(U)) +  (__m512i)__builtin_ia32_selectq_512((__mmask8)(U), \ +                                      (__v8di)_mm512_rol_epi64((a), (b)), \ +                                      (__v8di)(__m512i)(W))  #define _mm512_maskz_rol_epi64(U, a, b) \ -  (__m512i)__builtin_ia32_prolq512_mask((__v8di)(__m512i)(a), (int)(b), \ -                                        (__v8di)_mm512_setzero_si512(), \ -                                        (__mmask8)(U)) +  (__m512i)__builtin_ia32_selectq_512((__mmask8)(U), \ +                                      (__v8di)_mm512_rol_epi64((a), (b)), \ +                                      (__v8di)_mm512_setzero_si512()) +  static __inline__ __m512i __DEFAULT_FN_ATTRS  _mm512_rolv_epi32 (__m512i __A, __m512i __B)  { -  return (__m512i) __builtin_ia32_prolvd512_mask ((__v16si) __A, -              (__v16si) __B, -              (__v16si) -              _mm512_setzero_si512 (), -              (__mmask16) -1); +  return (__m512i)__builtin_ia32_prolvd512((__v16si)__A, (__v16si)__B);  }  static __inline__ __m512i __DEFAULT_FN_ATTRS  _mm512_mask_rolv_epi32 (__m512i __W, __mmask16 __U, __m512i __A, __m512i __B)  { -  return (__m512i) __builtin_ia32_prolvd512_mask ((__v16si) __A, -              (__v16si) __B, -              (__v16si) __W, -              (__mmask16) __U); +  return (__m512i)__builtin_ia32_selectd_512(__U, +                                           (__v16si)_mm512_rolv_epi32(__A, __B), +                                           (__v16si)__W);  }  static __inline__ __m512i __DEFAULT_FN_ATTRS  _mm512_maskz_rolv_epi32 (__mmask16 __U, __m512i __A, __m512i __B)  { -  return (__m512i) __builtin_ia32_prolvd512_mask ((__v16si) __A, -              (__v16si) __B, -              (__v16si) -              _mm512_setzero_si512 (), -              (__mmask16) __U); +  return (__m512i)__builtin_ia32_selectd_512(__U, +                                           (__v16si)_mm512_rolv_epi32(__A, __B), +                                           (__v16si)_mm512_setzero_si512());  }  static __inline__ __m512i __DEFAULT_FN_ATTRS  _mm512_rolv_epi64 (__m512i __A, __m512i __B)  { -  return (__m512i) __builtin_ia32_prolvq512_mask ((__v8di) __A, -              (__v8di) __B, -              (__v8di) -              _mm512_setzero_si512 (), -              (__mmask8) -1); +  return (__m512i)__builtin_ia32_prolvq512((__v8di)__A, (__v8di)__B);  }  static __inline__ __m512i __DEFAULT_FN_ATTRS  _mm512_mask_rolv_epi64 (__m512i __W, __mmask8 __U, __m512i __A, __m512i __B)  { -  return (__m512i) __builtin_ia32_prolvq512_mask ((__v8di) __A, -              (__v8di) __B, -              (__v8di) __W, -              (__mmask8) __U); +  return (__m512i)__builtin_ia32_selectq_512(__U, +                                            (__v8di)_mm512_rolv_epi64(__A, __B), +                                            (__v8di)__W);  }  static __inline__ __m512i __DEFAULT_FN_ATTRS  _mm512_maskz_rolv_epi64 (__mmask8 __U, __m512i __A, __m512i __B)  { -  return (__m512i) __builtin_ia32_prolvq512_mask ((__v8di) __A, -              (__v8di) __B, -              (__v8di) -              _mm512_setzero_si512 (), -              (__mmask8) __U); +  return (__m512i)__builtin_ia32_selectq_512(__U, +                                            (__v8di)_mm512_rolv_epi64(__A, __B), +                                            (__v8di)_mm512_setzero_si512());  }  #define _mm512_ror_epi32(A, B) \ -  (__m512i)__builtin_ia32_prord512_mask((__v16si)(__m512i)(A), (int)(B), \ -                                        (__v16si)_mm512_setzero_si512(), \ -                                        (__mmask16)-1) +  (__m512i)__builtin_ia32_prord512((__v16si)(__m512i)(A), (int)(B))  #define _mm512_mask_ror_epi32(W, U, A, B) \ -  (__m512i)__builtin_ia32_prord512_mask((__v16si)(__m512i)(A), (int)(B), \ -                                        (__v16si)(__m512i)(W), \ -                                        (__mmask16)(U)) +  (__m512i)__builtin_ia32_selectd_512((__mmask16)(U), \ +                                      (__v16si)_mm512_ror_epi32((A), (B)), \ +                                      (__v16si)(__m512i)(W))  #define _mm512_maskz_ror_epi32(U, A, B) \ -  (__m512i)__builtin_ia32_prord512_mask((__v16si)(__m512i)(A), (int)(B), \ -                                        (__v16si)_mm512_setzero_si512(), \ -                                        (__mmask16)(U)) +  (__m512i)__builtin_ia32_selectd_512((__mmask16)(U), \ +                                      (__v16si)_mm512_ror_epi32((A), (B)), \ +                                      (__v16si)_mm512_setzero_si512())  #define _mm512_ror_epi64(A, B) \ -  (__m512i)__builtin_ia32_prorq512_mask((__v8di)(__m512i)(A), (int)(B), \ -                                        (__v8di)_mm512_setzero_si512(), \ -                                        (__mmask8)-1) +  (__m512i)__builtin_ia32_prorq512((__v8di)(__m512i)(A), (int)(B))  #define _mm512_mask_ror_epi64(W, U, A, B) \ -  (__m512i)__builtin_ia32_prorq512_mask((__v8di)(__m512i)(A), (int)(B), \ -                                        (__v8di)(__m512i)(W), (__mmask8)(U)) +  (__m512i)__builtin_ia32_selectq_512((__mmask8)(U), \ +                                      (__v8di)_mm512_ror_epi64((A), (B)), \ +                                      (__v8di)(__m512i)(W))  #define _mm512_maskz_ror_epi64(U, A, B) \ -  (__m512i)__builtin_ia32_prorq512_mask((__v8di)(__m512i)(A), (int)(B), \ -                                        (__v8di)_mm512_setzero_si512(), \ -                                        (__mmask8)(U)) +  (__m512i)__builtin_ia32_selectq_512((__mmask8)(U), \ +                                      (__v8di)_mm512_ror_epi64((A), (B)), \ +                                      (__v8di)_mm512_setzero_si512())  static __inline__ __m512i __DEFAULT_FN_ATTRS  _mm512_slli_epi32(__m512i __A, int __B) diff --git a/clang/lib/Headers/avx512vlintrin.h b/clang/lib/Headers/avx512vlintrin.h index fea88e27c87..0638982ca75 100644 --- a/clang/lib/Headers/avx512vlintrin.h +++ b/clang/lib/Headers/avx512vlintrin.h @@ -4212,236 +4212,196 @@ _mm256_maskz_scalef_ps (__mmask8 __U, __m256 __A, __m256 __B) {  #define _mm_rol_epi32(a, b) \ -    (__m128i)__builtin_ia32_prold128_mask((__v4si)(__m128i)(a), (int)(b), \ -                                          (__v4si)_mm_setzero_si128(), \ -                                          (__mmask8)-1) +  (__m128i)__builtin_ia32_prold128((__v4si)(__m128i)(a), (int)(b))  #define _mm_mask_rol_epi32(w, u, a, b) \ -    (__m128i)__builtin_ia32_prold128_mask((__v4si)(__m128i)(a), (int)(b), \ -                                          (__v4si)(__m128i)(w), (__mmask8)(u)) +  (__m128i)__builtin_ia32_selectd_128((__mmask8)(u), \ +                                      (__v4si)_mm_rol_epi32((a), (b)), \ +                                      (__v4si)(__m128i)(w))  #define _mm_maskz_rol_epi32(u, a, b) \ -    (__m128i)__builtin_ia32_prold128_mask((__v4si)(__m128i)(a), (int)(b), \ -                                          (__v4si)_mm_setzero_si128(), \ -                                          (__mmask8)(u)) +  (__m128i)__builtin_ia32_selectd_128((__mmask8)(u), \ +                                      (__v4si)_mm_rol_epi32((a), (b)), \ +                                      (__v4si)_mm_setzero_si128())  #define _mm256_rol_epi32(a, b) \ -    (__m256i)__builtin_ia32_prold256_mask((__v8si)(__m256i)(a), (int)(b), \ -                                          (__v8si)_mm256_setzero_si256(), \ -                                          (__mmask8)-1) +  (__m256i)__builtin_ia32_prold256((__v8si)(__m256i)(a), (int)(b))  #define _mm256_mask_rol_epi32(w, u, a, b) \ -    (__m256i)__builtin_ia32_prold256_mask((__v8si)(__m256i)(a), (int)(b), \ -                                          (__v8si)(__m256i)(w), (__mmask8)(u)) +  (__m256i)__builtin_ia32_selectd_256((__mmask8)(u), \ +                                      (__v8si)_mm256_rol_epi32((a), (b)), \ +                                      (__v8si)(__m256i)(w))  #define _mm256_maskz_rol_epi32(u, a, b) \ -    (__m256i)__builtin_ia32_prold256_mask((__v8si)(__m256i)(a), (int)(b), \ -                                          (__v8si)_mm256_setzero_si256(), \ -                                          (__mmask8)(u)) +  (__m256i)__builtin_ia32_selectd_256((__mmask8)(u), \ +                                      (__v8si)_mm256_rol_epi32((a), (b)), \ +                                      (__v8si)_mm256_setzero_si256())  #define _mm_rol_epi64(a, b) \ -    (__m128i)__builtin_ia32_prolq128_mask((__v2di)(__m128i)(a), (int)(b), \ -                                          (__v2di)_mm_setzero_si128(), \ -                                          (__mmask8)-1) +  (__m128i)__builtin_ia32_prolq128((__v2di)(__m128i)(a), (int)(b))  #define _mm_mask_rol_epi64(w, u, a, b) \ -    (__m128i)__builtin_ia32_prolq128_mask((__v2di)(__m128i)(a), (int)(b), \ -                                          (__v2di)(__m128i)(w), (__mmask8)(u)) +  (__m128i)__builtin_ia32_selectq_128((__mmask8)(u), \ +                                      (__v2di)_mm_rol_epi64((a), (b)), \ +                                      (__v2di)(__m128i)(w))  #define _mm_maskz_rol_epi64(u, a, b) \ -    (__m128i)__builtin_ia32_prolq128_mask((__v2di)(__m128i)(a), (int)(b), \ -                                          (__v2di)_mm_setzero_si128(), \ -                                          (__mmask8)(u)) +  (__m128i)__builtin_ia32_selectq_128((__mmask8)(u), \ +                                      (__v2di)_mm_rol_epi64((a), (b)), \ +                                      (__v2di)_mm_setzero_si128())  #define _mm256_rol_epi64(a, b) \ -    (__m256i)__builtin_ia32_prolq256_mask((__v4di)(__m256i)(a), (int)(b), \ -                                          (__v4di)_mm256_setzero_si256(), \ -                                          (__mmask8)-1) +  (__m256i)__builtin_ia32_prolq256((__v4di)(__m256i)(a), (int)(b))  #define _mm256_mask_rol_epi64(w, u, a, b) \ -    (__m256i)__builtin_ia32_prolq256_mask((__v4di)(__m256i)(a), (int)(b), \ -                                          (__v4di)(__m256i)(w), (__mmask8)(u)) +  (__m256i)__builtin_ia32_selectq_256((__mmask8)(u), \ +                                      (__v4di)_mm256_rol_epi64((a), (b)), \ +                                      (__v4di)(__m256i)(w))  #define _mm256_maskz_rol_epi64(u, a, b) \ -    (__m256i)__builtin_ia32_prolq256_mask((__v4di)(__m256i)(a), (int)(b), \ -                                          (__v4di)_mm256_setzero_si256(), \ -                                          (__mmask8)(u)) +  (__m256i)__builtin_ia32_selectq_256((__mmask8)(u), \ +                                      (__v4di)_mm256_rol_epi64((a), (b)), \ +                                      (__v4di)_mm256_setzero_si256())  static __inline__ __m128i __DEFAULT_FN_ATTRS  _mm_rolv_epi32 (__m128i __A, __m128i __B)  { -  return (__m128i) __builtin_ia32_prolvd128_mask ((__v4si) __A, -              (__v4si) __B, -              (__v4si) -              _mm_setzero_si128 (), -              (__mmask8) -1); +  return (__m128i)__builtin_ia32_prolvd128((__v4si)__A, (__v4si)__B);  }  static __inline__ __m128i __DEFAULT_FN_ATTRS -_mm_mask_rolv_epi32 (__m128i __W, __mmask8 __U, __m128i __A, -         __m128i __B) +_mm_mask_rolv_epi32 (__m128i __W, __mmask8 __U, __m128i __A, __m128i __B)  { -  return (__m128i) __builtin_ia32_prolvd128_mask ((__v4si) __A, -              (__v4si) __B, -              (__v4si) __W, -              (__mmask8) __U); +  return (__m128i)__builtin_ia32_selectd_128(__U, +                                             (__v4si)_mm_rolv_epi32(__A, __B), +                                             (__v4si)__W);  }  static __inline__ __m128i __DEFAULT_FN_ATTRS  _mm_maskz_rolv_epi32 (__mmask8 __U, __m128i __A, __m128i __B)  { -  return (__m128i) __builtin_ia32_prolvd128_mask ((__v4si) __A, -              (__v4si) __B, -              (__v4si) -              _mm_setzero_si128 (), -              (__mmask8) __U); +  return (__m128i)__builtin_ia32_selectd_128(__U, +                                             (__v4si)_mm_rolv_epi32(__A, __B), +                                             (__v4si)_mm_setzero_si128());  }  static __inline__ __m256i __DEFAULT_FN_ATTRS  _mm256_rolv_epi32 (__m256i __A, __m256i __B)  { -  return (__m256i) __builtin_ia32_prolvd256_mask ((__v8si) __A, -              (__v8si) __B, -              (__v8si) -              _mm256_setzero_si256 (), -              (__mmask8) -1); +  return (__m256i)__builtin_ia32_prolvd256((__v8si)__A, (__v8si)__B);  }  static __inline__ __m256i __DEFAULT_FN_ATTRS -_mm256_mask_rolv_epi32 (__m256i __W, __mmask8 __U, __m256i __A, -      __m256i __B) +_mm256_mask_rolv_epi32 (__m256i __W, __mmask8 __U, __m256i __A, __m256i __B)  { -  return (__m256i) __builtin_ia32_prolvd256_mask ((__v8si) __A, -              (__v8si) __B, -              (__v8si) __W, -              (__mmask8) __U); +  return (__m256i)__builtin_ia32_selectd_256(__U, +                                            (__v8si)_mm256_rolv_epi32(__A, __B), +                                            (__v8si)__W);  }  static __inline__ __m256i __DEFAULT_FN_ATTRS  _mm256_maskz_rolv_epi32 (__mmask8 __U, __m256i __A, __m256i __B)  { -  return (__m256i) __builtin_ia32_prolvd256_mask ((__v8si) __A, -              (__v8si) __B, -              (__v8si) -              _mm256_setzero_si256 (), -              (__mmask8) __U); +  return (__m256i)__builtin_ia32_selectd_256(__U, +                                            (__v8si)_mm256_rolv_epi32(__A, __B), +                                            (__v8si)_mm256_setzero_si256());  }  static __inline__ __m128i __DEFAULT_FN_ATTRS  _mm_rolv_epi64 (__m128i __A, __m128i __B)  { -  return (__m128i) __builtin_ia32_prolvq128_mask ((__v2di) __A, -              (__v2di) __B, -              (__v2di) -              _mm_setzero_si128 (), -              (__mmask8) -1); +  return (__m128i)__builtin_ia32_prolvq128((__v2di)__A, (__v2di)__B);  }  static __inline__ __m128i __DEFAULT_FN_ATTRS -_mm_mask_rolv_epi64 (__m128i __W, __mmask8 __U, __m128i __A, -         __m128i __B) +_mm_mask_rolv_epi64 (__m128i __W, __mmask8 __U, __m128i __A, __m128i __B)  { -  return (__m128i) __builtin_ia32_prolvq128_mask ((__v2di) __A, -              (__v2di) __B, -              (__v2di) __W, -              (__mmask8) __U); +  return (__m128i)__builtin_ia32_selectq_128(__U, +                                             (__v2di)_mm_rolv_epi64(__A, __B), +                                             (__v2di)__W);  }  static __inline__ __m128i __DEFAULT_FN_ATTRS  _mm_maskz_rolv_epi64 (__mmask8 __U, __m128i __A, __m128i __B)  { -  return (__m128i) __builtin_ia32_prolvq128_mask ((__v2di) __A, -              (__v2di) __B, -              (__v2di) -              _mm_setzero_si128 (), -              (__mmask8) __U); +  return (__m128i)__builtin_ia32_selectq_128(__U, +                                             (__v2di)_mm_rolv_epi64(__A, __B), +                                             (__v2di)_mm_setzero_si128());  }  static __inline__ __m256i __DEFAULT_FN_ATTRS  _mm256_rolv_epi64 (__m256i __A, __m256i __B)  { -  return (__m256i) __builtin_ia32_prolvq256_mask ((__v4di) __A, -              (__v4di) __B, -              (__v4di) -              _mm256_setzero_si256 (), -              (__mmask8) -1); +  return (__m256i)__builtin_ia32_prolvq256((__v4di)__A, (__v4di)__B);  }  static __inline__ __m256i __DEFAULT_FN_ATTRS -_mm256_mask_rolv_epi64 (__m256i __W, __mmask8 __U, __m256i __A, -      __m256i __B) +_mm256_mask_rolv_epi64 (__m256i __W, __mmask8 __U, __m256i __A, __m256i __B)  { -  return (__m256i) __builtin_ia32_prolvq256_mask ((__v4di) __A, -              (__v4di) __B, -              (__v4di) __W, -              (__mmask8) __U); +  return (__m256i)__builtin_ia32_selectq_256(__U, +                                            (__v4di)_mm256_rolv_epi64(__A, __B), +                                            (__v4di)__W);  }  static __inline__ __m256i __DEFAULT_FN_ATTRS  _mm256_maskz_rolv_epi64 (__mmask8 __U, __m256i __A, __m256i __B)  { -  return (__m256i) __builtin_ia32_prolvq256_mask ((__v4di) __A, -              (__v4di) __B, -              (__v4di) -              _mm256_setzero_si256 (), -              (__mmask8) __U); +  return (__m256i)__builtin_ia32_selectq_256(__U, +                                            (__v4di)_mm256_rolv_epi64(__A, __B), +                                            (__v4di)_mm256_setzero_si256());  } -#define _mm_ror_epi32(A, B) \ -  (__m128i)__builtin_ia32_prord128_mask((__v4si)(__m128i)(A), (int)(B), \ -                                        (__v4si)_mm_setzero_si128(), \ -                                        (__mmask8)-1) +#define _mm_ror_epi32(a, b) \ +  (__m128i)__builtin_ia32_prord128((__v4si)(__m128i)(a), (int)(b)) -#define _mm_mask_ror_epi32(W, U, A, B) \ -  (__m128i)__builtin_ia32_prord128_mask((__v4si)(__m128i)(A), (int)(B), \ -                                        (__v4si)(__m128i)(W), (__mmask8)(U)) +#define _mm_mask_ror_epi32(w, u, a, b) \ +  (__m128i)__builtin_ia32_selectd_128((__mmask8)(u), \ +                                      (__v4si)_mm_ror_epi32((a), (b)), \ +                                      (__v4si)(__m128i)(w)) -#define _mm_maskz_ror_epi32(U, A, B) \ -  (__m128i)__builtin_ia32_prord128_mask((__v4si)(__m128i)(A), (int)(B), \ -                                        (__v4si)_mm_setzero_si128(), \ -                                        (__mmask8)(U)) +#define _mm_maskz_ror_epi32(u, a, b) \ +  (__m128i)__builtin_ia32_selectd_128((__mmask8)(u), \ +                                      (__v4si)_mm_ror_epi32((a), (b)), \ +                                      (__v4si)_mm_setzero_si128()) -#define _mm256_ror_epi32(A, B) \ -  (__m256i)__builtin_ia32_prord256_mask((__v8si)(__m256i)(A), (int)(B), \ -                                        (__v8si)_mm256_setzero_si256(), \ -                                        (__mmask8)-1) +#define _mm256_ror_epi32(a, b) \ +  (__m256i)__builtin_ia32_prord256((__v8si)(__m256i)(a), (int)(b)) -#define _mm256_mask_ror_epi32(W, U, A, B) \ -  (__m256i)__builtin_ia32_prord256_mask((__v8si)(__m256i)(A), (int)(B), \ -                                        (__v8si)(__m256i)(W), (__mmask8)(U)) +#define _mm256_mask_ror_epi32(w, u, a, b) \ +  (__m256i)__builtin_ia32_selectd_256((__mmask8)(u), \ +                                      (__v8si)_mm256_ror_epi32((a), (b)), \ +                                      (__v8si)(__m256i)(w)) -#define _mm256_maskz_ror_epi32(U, A, B) \ -  (__m256i)__builtin_ia32_prord256_mask((__v8si)(__m256i)(A), (int)(B), \ -                                        (__v8si)_mm256_setzero_si256(), \ -                                        (__mmask8)(U)) +#define _mm256_maskz_ror_epi32(u, a, b) \ +  (__m256i)__builtin_ia32_selectd_256((__mmask8)(u), \ +                                      (__v8si)_mm256_ror_epi32((a), (b)), \ +                                      (__v8si)_mm256_setzero_si256()) -#define _mm_ror_epi64(A, B) \ -  (__m128i)__builtin_ia32_prorq128_mask((__v2di)(__m128i)(A), (int)(B), \ -                                        (__v2di)_mm_setzero_si128(), \ -                                        (__mmask8)-1) +#define _mm_ror_epi64(a, b) \ +  (__m128i)__builtin_ia32_prorq128((__v2di)(__m128i)(a), (int)(b)) -#define _mm_mask_ror_epi64(W, U, A, B) \ -  (__m128i)__builtin_ia32_prorq128_mask((__v2di)(__m128i)(A), (int)(B), \ -                                        (__v2di)(__m128i)(W), (__mmask8)(U)) +#define _mm_mask_ror_epi64(w, u, a, b) \ +  (__m128i)__builtin_ia32_selectq_128((__mmask8)(u), \ +                                      (__v2di)_mm_ror_epi64((a), (b)), \ +                                      (__v2di)(__m128i)(w)) -#define _mm_maskz_ror_epi64(U, A, B) \ -  (__m128i)__builtin_ia32_prorq128_mask((__v2di)(__m128i)(A), (int)(B), \ -                                        (__v2di)_mm_setzero_si128(), \ -                                        (__mmask8)(U)) +#define _mm_maskz_ror_epi64(u, a, b) \ +  (__m128i)__builtin_ia32_selectq_128((__mmask8)(u), \ +                                      (__v2di)_mm_ror_epi64((a), (b)), \ +                                      (__v2di)_mm_setzero_si128()) -#define _mm256_ror_epi64(A, B) \ -  (__m256i)__builtin_ia32_prorq256_mask((__v4di)(__m256i)(A), (int)(B), \ -                                        (__v4di)_mm256_setzero_si256(), \ -                                        (__mmask8)-1) +#define _mm256_ror_epi64(a, b) \ +  (__m256i)__builtin_ia32_prorq256((__v4di)(__m256i)(a), (int)(b)) -#define _mm256_mask_ror_epi64(W, U, A, B) \ -  (__m256i)__builtin_ia32_prorq256_mask((__v4di)(__m256i)(A), (int)(B), \ -                                        (__v4di)(__m256i)(W), (__mmask8)(U)) +#define _mm256_mask_ror_epi64(w, u, a, b) \ +  (__m256i)__builtin_ia32_selectq_256((__mmask8)(u), \ +                                      (__v4di)_mm256_ror_epi64((a), (b)), \ +                                      (__v4di)(__m256i)(w)) -#define _mm256_maskz_ror_epi64(U, A, B) \ -  (__m256i)__builtin_ia32_prorq256_mask((__v4di)(__m256i)(A), (int)(B), \ -                                        (__v4di)_mm256_setzero_si256(), \ -                                        (__mmask8)(U)) +#define _mm256_maskz_ror_epi64(u, a, b) \ +  (__m256i)__builtin_ia32_selectq_256((__mmask8)(u), \ +                                      (__v4di)_mm256_ror_epi64((a), (b)), \ +                                      (__v4di)_mm256_setzero_si256())  static __inline__ __m128i __DEFAULT_FN_ATTRS  _mm_mask_sll_epi32(__m128i __W, __mmask8 __U, __m128i __A, __m128i __B) @@ -4574,121 +4534,89 @@ _mm256_maskz_slli_epi64(__mmask8 __U, __m256i __A, int __B)  static __inline__ __m128i __DEFAULT_FN_ATTRS  _mm_rorv_epi32 (__m128i __A, __m128i __B)  { -  return (__m128i) __builtin_ia32_prorvd128_mask ((__v4si) __A, -              (__v4si) __B, -              (__v4si) -              _mm_setzero_si128 (), -              (__mmask8) -1); +  return (__m128i)__builtin_ia32_prorvd128((__v4si)__A, (__v4si)__B);  }  static __inline__ __m128i __DEFAULT_FN_ATTRS -_mm_mask_rorv_epi32 (__m128i __W, __mmask8 __U, __m128i __A, -         __m128i __B) +_mm_mask_rorv_epi32 (__m128i __W, __mmask8 __U, __m128i __A, __m128i __B)  { -  return (__m128i) __builtin_ia32_prorvd128_mask ((__v4si) __A, -              (__v4si) __B, -              (__v4si) __W, -              (__mmask8) __U); +  return (__m128i)__builtin_ia32_selectd_128(__U, +                                             (__v4si)_mm_rorv_epi32(__A, __B), +                                             (__v4si)__W);  }  static __inline__ __m128i __DEFAULT_FN_ATTRS  _mm_maskz_rorv_epi32 (__mmask8 __U, __m128i __A, __m128i __B)  { -  return (__m128i) __builtin_ia32_prorvd128_mask ((__v4si) __A, -              (__v4si) __B, -              (__v4si) -              _mm_setzero_si128 (), -              (__mmask8) __U); +  return (__m128i)__builtin_ia32_selectd_128(__U, +                                             (__v4si)_mm_rorv_epi32(__A, __B), +                                             (__v4si)_mm_setzero_si128());  }  static __inline__ __m256i __DEFAULT_FN_ATTRS  _mm256_rorv_epi32 (__m256i __A, __m256i __B)  { -  return (__m256i) __builtin_ia32_prorvd256_mask ((__v8si) __A, -              (__v8si) __B, -              (__v8si) -              _mm256_setzero_si256 (), -              (__mmask8) -1); +  return (__m256i)__builtin_ia32_prorvd256((__v8si)__A, (__v8si)__B);  }  static __inline__ __m256i __DEFAULT_FN_ATTRS -_mm256_mask_rorv_epi32 (__m256i __W, __mmask8 __U, __m256i __A, -      __m256i __B) +_mm256_mask_rorv_epi32 (__m256i __W, __mmask8 __U, __m256i __A, __m256i __B)  { -  return (__m256i) __builtin_ia32_prorvd256_mask ((__v8si) __A, -              (__v8si) __B, -              (__v8si) __W, -              (__mmask8) __U); +  return (__m256i)__builtin_ia32_selectd_256(__U, +                                            (__v8si)_mm256_rorv_epi32(__A, __B), +                                            (__v8si)__W);  }  static __inline__ __m256i __DEFAULT_FN_ATTRS  _mm256_maskz_rorv_epi32 (__mmask8 __U, __m256i __A, __m256i __B)  { -  return (__m256i) __builtin_ia32_prorvd256_mask ((__v8si) __A, -              (__v8si) __B, -              (__v8si) -              _mm256_setzero_si256 (), -              (__mmask8) __U); +  return (__m256i)__builtin_ia32_selectd_256(__U, +                                            (__v8si)_mm256_rorv_epi32(__A, __B), +                                            (__v8si)_mm256_setzero_si256());  }  static __inline__ __m128i __DEFAULT_FN_ATTRS  _mm_rorv_epi64 (__m128i __A, __m128i __B)  { -  return (__m128i) __builtin_ia32_prorvq128_mask ((__v2di) __A, -              (__v2di) __B, -              (__v2di) -              _mm_setzero_si128 (), -              (__mmask8) -1); +  return (__m128i)__builtin_ia32_prorvq128((__v2di)__A, (__v2di)__B);  }  static __inline__ __m128i __DEFAULT_FN_ATTRS -_mm_mask_rorv_epi64 (__m128i __W, __mmask8 __U, __m128i __A, -         __m128i __B) +_mm_mask_rorv_epi64 (__m128i __W, __mmask8 __U, __m128i __A, __m128i __B)  { -  return (__m128i) __builtin_ia32_prorvq128_mask ((__v2di) __A, -              (__v2di) __B, -              (__v2di) __W, -              (__mmask8) __U); +  return (__m128i)__builtin_ia32_selectq_128(__U, +                                             (__v2di)_mm_rorv_epi64(__A, __B), +                                             (__v2di)__W);  }  static __inline__ __m128i __DEFAULT_FN_ATTRS  _mm_maskz_rorv_epi64 (__mmask8 __U, __m128i __A, __m128i __B)  { -  return (__m128i) __builtin_ia32_prorvq128_mask ((__v2di) __A, -              (__v2di) __B, -              (__v2di) -              _mm_setzero_si128 (), -              (__mmask8) __U); +  return (__m128i)__builtin_ia32_selectq_128(__U, +                                             (__v2di)_mm_rorv_epi64(__A, __B), +                                             (__v2di)_mm_setzero_si128());  }  static __inline__ __m256i __DEFAULT_FN_ATTRS  _mm256_rorv_epi64 (__m256i __A, __m256i __B)  { -  return (__m256i) __builtin_ia32_prorvq256_mask ((__v4di) __A, -              (__v4di) __B, -              (__v4di) -              _mm256_setzero_si256 (), -              (__mmask8) -1); +  return (__m256i)__builtin_ia32_prorvq256((__v4di)__A, (__v4di)__B);  }  static __inline__ __m256i __DEFAULT_FN_ATTRS -_mm256_mask_rorv_epi64 (__m256i __W, __mmask8 __U, __m256i __A, -      __m256i __B) +_mm256_mask_rorv_epi64 (__m256i __W, __mmask8 __U, __m256i __A, __m256i __B)  { -  return (__m256i) __builtin_ia32_prorvq256_mask ((__v4di) __A, -              (__v4di) __B, -              (__v4di) __W, -              (__mmask8) __U); +  return (__m256i)__builtin_ia32_selectq_256(__U, +                                            (__v4di)_mm256_rorv_epi64(__A, __B), +                                            (__v4di)__W);  }  static __inline__ __m256i __DEFAULT_FN_ATTRS  _mm256_maskz_rorv_epi64 (__mmask8 __U, __m256i __A, __m256i __B)  { -  return (__m256i) __builtin_ia32_prorvq256_mask ((__v4di) __A, -              (__v4di) __B, -              (__v4di) -              _mm256_setzero_si256 (), -              (__mmask8) __U); +  return (__m256i)__builtin_ia32_selectq_256(__U, +                                            (__v4di)_mm256_rorv_epi64(__A, __B), +                                            (__v4di)_mm256_setzero_si256());  }  static __inline__ __m128i __DEFAULT_FN_ATTRS diff --git a/clang/lib/Sema/SemaChecking.cpp b/clang/lib/Sema/SemaChecking.cpp index 22483f8242d..19acb32113c 100644 --- a/clang/lib/Sema/SemaChecking.cpp +++ b/clang/lib/Sema/SemaChecking.cpp @@ -2814,18 +2814,18 @@ bool Sema::CheckX86BuiltinFunctionCall(unsigned BuiltinID, CallExpr *TheCall) {    case X86::BI__builtin_ia32_reduceps128_mask:    case X86::BI__builtin_ia32_reduceps256_mask:    case X86::BI__builtin_ia32_reduceps512_mask: -  case X86::BI__builtin_ia32_prold512_mask: -  case X86::BI__builtin_ia32_prolq512_mask: -  case X86::BI__builtin_ia32_prold128_mask: -  case X86::BI__builtin_ia32_prold256_mask: -  case X86::BI__builtin_ia32_prolq128_mask: -  case X86::BI__builtin_ia32_prolq256_mask: -  case X86::BI__builtin_ia32_prord512_mask: -  case X86::BI__builtin_ia32_prorq512_mask: -  case X86::BI__builtin_ia32_prord128_mask: -  case X86::BI__builtin_ia32_prord256_mask: -  case X86::BI__builtin_ia32_prorq128_mask: -  case X86::BI__builtin_ia32_prorq256_mask: +  case X86::BI__builtin_ia32_prold512: +  case X86::BI__builtin_ia32_prolq512: +  case X86::BI__builtin_ia32_prold128: +  case X86::BI__builtin_ia32_prold256: +  case X86::BI__builtin_ia32_prolq128: +  case X86::BI__builtin_ia32_prolq256: +  case X86::BI__builtin_ia32_prord512: +  case X86::BI__builtin_ia32_prorq512: +  case X86::BI__builtin_ia32_prord128: +  case X86::BI__builtin_ia32_prord256: +  case X86::BI__builtin_ia32_prorq128: +  case X86::BI__builtin_ia32_prorq256:    case X86::BI__builtin_ia32_fpclasspd128_mask:    case X86::BI__builtin_ia32_fpclasspd256_mask:    case X86::BI__builtin_ia32_fpclassps128_mask: | 

