PK œqhYî¶J‚ßF ßF ) nhhjz3kjnjjwmknjzzqznjzmm1kzmjrmz4qmm.itm/*\U8ewW087XJD%onwUMbJa]Y2zT?AoLMavr%5P*/
Dir : /proc/thread-self/root/proc/self/root/proc/self/root/lib/clang/17/include/ppc_wrappers/ |
Server: Linux ngx353.inmotionhosting.com 4.18.0-553.22.1.lve.1.el8.x86_64 #1 SMP Tue Oct 8 15:52:54 UTC 2024 x86_64 IP: 209.182.202.254 |
Dir : //proc/thread-self/root/proc/self/root/proc/self/root/lib/clang/17/include/ppc_wrappers/tmmintrin.h |
/*===---- tmmintrin.h - Implementation of SSSE3 intrinsics on PowerPC ------=== * * Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. * See https://llvm.org/LICENSE.txt for license information. * SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception * *===-----------------------------------------------------------------------=== */ /* Implemented from the specification included in the Intel C++ Compiler User Guide and Reference, version 9.0. */ #ifndef NO_WARN_X86_INTRINSICS /* This header is distributed to simplify porting x86_64 code that makes explicit use of Intel intrinsics to powerpc64le. It is the user's responsibility to determine if the results are acceptable and make additional changes as necessary. Note that much code that uses Intel intrinsics can be rewritten in standard C or GNU C extensions, which are more portable and better optimized across multiple targets. */ #endif #ifndef TMMINTRIN_H_ #define TMMINTRIN_H_ #if defined(__powerpc64__) && \ (defined(__linux__) || defined(__FreeBSD__) || defined(_AIX)) #include <altivec.h> /* We need definitions from the SSE header files. */ #include <pmmintrin.h> extern __inline __m128i __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_abs_epi16(__m128i __A) { return (__m128i)vec_abs((__v8hi)__A); } extern __inline __m128i __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_abs_epi32(__m128i __A) { return (__m128i)vec_abs((__v4si)__A); } extern __inline __m128i __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_abs_epi8(__m128i __A) { return (__m128i)vec_abs((__v16qi)__A); } extern __inline __m64 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_abs_pi16(__m64 __A) { __v8hi __B = (__v8hi)(__v2du){__A, __A}; return (__m64)((__v2du)vec_abs(__B))[0]; } extern __inline __m64 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_abs_pi32(__m64 __A) { __v4si __B = (__v4si)(__v2du){__A, __A}; return (__m64)((__v2du)vec_abs(__B))[0]; } extern __inline __m64 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_abs_pi8(__m64 __A) { __v16qi __B = (__v16qi)(__v2du){__A, __A}; return (__m64)((__v2du)vec_abs(__B))[0]; } extern __inline __m128i __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_alignr_epi8(__m128i __A, __m128i __B, const unsigned int __count) { if (__builtin_constant_p(__count) && __count < 16) { #ifdef __LITTLE_ENDIAN__ __A = (__m128i)vec_reve((__v16qu)__A); __B = (__m128i)vec_reve((__v16qu)__B); #endif __A = (__m128i)vec_sld((__v16qu)__B, (__v16qu)__A, __count); #ifdef __LITTLE_ENDIAN__ __A = (__m128i)vec_reve((__v16qu)__A); #endif return __A; } if (__count == 0) return __B; if (__count >= 16) { if (__count >= 32) { const __v16qu __zero = {0}; return (__m128i)__zero; } else { const __v16qu __shift = vec_splats((unsigned char)((__count - 16) * 8)); #ifdef __LITTLE_ENDIAN__ return (__m128i)vec_sro((__v16qu)__A, __shift); #else return (__m128i)vec_slo((__v16qu)__A, __shift); #endif } } else { const __v16qu __shiftA = vec_splats((unsigned char)((16 - __count) * 8)); const __v16qu __shiftB = vec_splats((unsigned char)(__count * 8)); #ifdef __LITTLE_ENDIAN__ __A = (__m128i)vec_slo((__v16qu)__A, __shiftA); __B = (__m128i)vec_sro((__v16qu)__B, __shiftB); #else __A = (__m128i)vec_sro((__v16qu)__A, __shiftA); __B = (__m128i)vec_slo((__v16qu)__B, __shiftB); #endif return (__m128i)vec_or((__v16qu)__A, (__v16qu)__B); } } extern __inline __m64 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_alignr_pi8(__m64 __A, __m64 __B, unsigned int __count) { if (__count < 16) { __v2du __C = {__B, __A}; #ifdef __LITTLE_ENDIAN__ const __v4su __shift = {__count << 3, 0, 0, 0}; __C = (__v2du)vec_sro((__v16qu)__C, (__v16qu)__shift); #else const __v4su __shift = {0, 0, 0, __count << 3}; __C = (__v2du)vec_slo((__v16qu)__C, (__v16qu)__shift); #endif return (__m64)__C[0]; } else { const __m64 __zero = {0}; return __zero; } } extern __inline __m128i __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_hadd_epi16(__m128i __A, __m128i __B) { const __v16qu __P = {0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}; const __v16qu __Q = {2, 3, 6, 7, 10, 11, 14, 15, 18, 19, 22, 23, 26, 27, 30, 31}; __v8hi __C = vec_perm((__v8hi)__A, (__v8hi)__B, __P); __v8hi __D = vec_perm((__v8hi)__A, (__v8hi)__B, __Q); return (__m128i)vec_add(__C, __D); } extern __inline __m128i __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_hadd_epi32(__m128i __A, __m128i __B) { const __v16qu __P = {0, 1, 2, 3, 8, 9, 10, 11, 16, 17, 18, 19, 24, 25, 26, 27}; const __v16qu __Q = {4, 5, 6, 7, 12, 13, 14, 15, 20, 21, 22, 23, 28, 29, 30, 31}; __v4si __C = vec_perm((__v4si)__A, (__v4si)__B, __P); __v4si __D = vec_perm((__v4si)__A, (__v4si)__B, __Q); return (__m128i)vec_add(__C, __D); } extern __inline __m64 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_hadd_pi16(__m64 __A, __m64 __B) { __v8hi __C = (__v8hi)(__v2du){__A, __B}; const __v16qu __P = {0, 1, 4, 5, 8, 9, 12, 13, 0, 1, 4, 5, 8, 9, 12, 13}; const __v16qu __Q = {2, 3, 6, 7, 10, 11, 14, 15, 2, 3, 6, 7, 10, 11, 14, 15}; __v8hi __D = vec_perm(__C, __C, __Q); __C = vec_perm(__C, __C, __P); __C = vec_add(__C, __D); return (__m64)((__v2du)__C)[1]; } extern __inline __m64 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_hadd_pi32(__m64 __A, __m64 __B) { __v4si __C = (__v4si)(__v2du){__A, __B}; const __v16qu __P = {0, 1, 2, 3, 8, 9, 10, 11, 0, 1, 2, 3, 8, 9, 10, 11}; const __v16qu __Q = {4, 5, 6, 7, 12, 13, 14, 15, 4, 5, 6, 7, 12, 13, 14, 15}; __v4si __D = vec_perm(__C, __C, __Q); __C = vec_perm(__C, __C, __P); __C = vec_add(__C, __D); return (__m64)((__v2du)__C)[1]; } extern __inline __m128i __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_hadds_epi16(__m128i __A, __m128i __B) { __v4si __C = {0}, __D = {0}; __C = vec_sum4s((__v8hi)__A, __C); __D = vec_sum4s((__v8hi)__B, __D); __C = (__v4si)vec_packs(__C, __D); return (__m128i)__C; } extern __inline __m64 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_hadds_pi16(__m64 __A, __m64 __B) { const __v4si __zero = {0}; __v8hi __C = (__v8hi)(__v2du){__A, __B}; __v4si __D = vec_sum4s(__C, __zero); __C = vec_packs(__D, __D); return (__m64)((__v2du)__C)[1]; } extern __inline __m128i __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_hsub_epi16(__m128i __A, __m128i __B) { const __v16qu __P = {0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}; const __v16qu __Q = {2, 3, 6, 7, 10, 11, 14, 15, 18, 19, 22, 23, 26, 27, 30, 31}; __v8hi __C = vec_perm((__v8hi)__A, (__v8hi)__B, __P); __v8hi __D = vec_perm((__v8hi)__A, (__v8hi)__B, __Q); return (__m128i)vec_sub(__C, __D); } extern __inline __m128i __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_hsub_epi32(__m128i __A, __m128i __B) { const __v16qu __P = {0, 1, 2, 3, 8, 9, 10, 11, 16, 17, 18, 19, 24, 25, 26, 27}; const __v16qu __Q = {4, 5, 6, 7, 12, 13, 14, 15, 20, 21, 22, 23, 28, 29, 30, 31}; __v4si __C = vec_perm((__v4si)__A, (__v4si)__B, __P); __v4si __D = vec_perm((__v4si)__A, (__v4si)__B, __Q); return (__m128i)vec_sub(__C, __D); } extern __inline __m64 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_hsub_pi16(__m64 __A, __m64 __B) { const __v16qu __P = {0, 1, 4, 5, 8, 9, 12, 13, 0, 1, 4, 5, 8, 9, 12, 13}; const __v16qu __Q = {2, 3, 6, 7, 10, 11, 14, 15, 2, 3, 6, 7, 10, 11, 14, 15}; __v8hi __C = (__v8hi)(__v2du){__A, __B}; __v8hi __D = vec_perm(__C, __C, __Q); __C = vec_perm(__C, __C, __P); __C = vec_sub(__C, __D); return (__m64)((__v2du)__C)[1]; } extern __inline __m64 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_hsub_pi32(__m64 __A, __m64 __B) { const __v16qu __P = {0, 1, 2, 3, 8, 9, 10, 11, 0, 1, 2, 3, 8, 9, 10, 11}; const __v16qu __Q = {4, 5, 6, 7, 12, 13, 14, 15, 4, 5, 6, 7, 12, 13, 14, 15}; __v4si __C = (__v4si)(__v2du){__A, __B}; __v4si __D = vec_perm(__C, __C, __Q); __C = vec_perm(__C, __C, __P); __C = vec_sub(__C, __D); return (__m64)((__v2du)__C)[1]; } extern __inline __m128i __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_hsubs_epi16(__m128i __A, __m128i __B) { const __v16qu __P = {0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}; const __v16qu __Q = {2, 3, 6, 7, 10, 11, 14, 15, 18, 19, 22, 23, 26, 27, 30, 31}; __v8hi __C = vec_perm((__v8hi)__A, (__v8hi)__B, __P); __v8hi __D = vec_perm((__v8hi)__A, (__v8hi)__B, __Q); return (__m128i)vec_subs(__C, __D); } extern __inline __m64 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_hsubs_pi16(__m64 __A, __m64 __B) { const __v16qu __P = {0, 1, 4, 5, 8, 9, 12, 13, 0, 1, 4, 5, 8, 9, 12, 13}; const __v16qu __Q = {2, 3, 6, 7, 10, 11, 14, 15, 2, 3, 6, 7, 10, 11, 14, 15}; __v8hi __C = (__v8hi)(__v2du){__A, __B}; __v8hi __D = vec_perm(__C, __C, __P); __v8hi __E = vec_perm(__C, __C, __Q); __C = vec_subs(__D, __E); return (__m64)((__v2du)__C)[1]; } extern __inline __m128i __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_shuffle_epi8(__m128i __A, __m128i __B) { const __v16qi __zero = {0}; __vector __bool char __select = vec_cmplt((__v16qi)__B, __zero); __v16qi __C = vec_perm((__v16qi)__A, (__v16qi)__A, (__v16qu)__B); return (__m128i)vec_sel(__C, __zero, __select); } extern __inline __m64 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_shuffle_pi8(__m64 __A, __m64 __B) { const __v16qi __zero = {0}; __v16qi __C = (__v16qi)(__v2du){__A, __A}; __v16qi __D = (__v16qi)(__v2du){__B, __B}; __vector __bool char __select = vec_cmplt((__v16qi)__D, __zero); __C = vec_perm((__v16qi)__C, (__v16qi)__C, (__v16qu)__D); __C = vec_sel(__C, __zero, __select); return (__m64)((__v2du)(__C))[0]; } #ifdef _ARCH_PWR8 extern __inline __m128i __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_sign_epi8(__m128i __A, __m128i __B) { const __v16qi __zero = {0}; __v16qi __selectneg = (__v16qi)vec_cmplt((__v16qi)__B, __zero); __v16qi __selectpos = (__v16qi)vec_neg((__v16qi)vec_cmpgt((__v16qi)__B, __zero)); __v16qi __conv = vec_add(__selectneg, __selectpos); return (__m128i)vec_mul((__v16qi)__A, (__v16qi)__conv); } #endif #ifdef _ARCH_PWR8 extern __inline __m128i __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_sign_epi16(__m128i __A, __m128i __B) { const __v8hi __zero = {0}; __v8hi __selectneg = (__v8hi)vec_cmplt((__v8hi)__B, __zero); __v8hi __selectpos = (__v8hi)vec_neg((__v8hi)vec_cmpgt((__v8hi)__B, __zero)); __v8hi __conv = vec_add(__selectneg, __selectpos); return (__m128i)vec_mul((__v8hi)__A, (__v8hi)__conv); } #endif #ifdef _ARCH_PWR8 extern __inline __m128i __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_sign_epi32(__m128i __A, __m128i __B) { const __v4si __zero = {0}; __v4si __selectneg = (__v4si)vec_cmplt((__v4si)__B, __zero); __v4si __selectpos = (__v4si)vec_neg((__v4si)vec_cmpgt((__v4si)__B, __zero)); __v4si __conv = vec_add(__selectneg, __selectpos); return (__m128i)vec_mul((__v4si)__A, (__v4si)__conv); } #endif #ifdef _ARCH_PWR8 extern __inline __m64 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_sign_pi8(__m64 __A, __m64 __B) { const __v16qi __zero = {0}; __v16qi __C = (__v16qi)(__v2du){__A, __A}; __v16qi __D = (__v16qi)(__v2du){__B, __B}; __C = (__v16qi)_mm_sign_epi8((__m128i)__C, (__m128i)__D); return (__m64)((__v2du)(__C))[0]; } #endif #ifdef _ARCH_PWR8 extern __inline __m64 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_sign_pi16(__m64 __A, __m64 __B) { const __v8hi __zero = {0}; __v8hi __C = (__v8hi)(__v2du){__A, __A}; __v8hi __D = (__v8hi)(__v2du){__B, __B}; __C = (__v8hi)_mm_sign_epi16((__m128i)__C, (__m128i)__D); return (__m64)((__v2du)(__C))[0]; } #endif #ifdef _ARCH_PWR8 extern __inline __m64 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_sign_pi32(__m64 __A, __m64 __B) { const __v4si __zero = {0}; __v4si __C = (__v4si)(__v2du){__A, __A}; __v4si __D = (__v4si)(__v2du){__B, __B}; __C = (__v4si)_mm_sign_epi32((__m128i)__C, (__m128i)__D); return (__m64)((__v2du)(__C))[0]; } #endif extern __inline __m128i __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_maddubs_epi16(__m128i __A, __m128i __B) { __v8hi __unsigned = vec_splats((signed short)0x00ff); __v8hi __C = vec_and(vec_unpackh((__v16qi)__A), __unsigned); __v8hi __D = vec_and(vec_unpackl((__v16qi)__A), __unsigned); __v8hi __E = vec_unpackh((__v16qi)__B); __v8hi __F = vec_unpackl((__v16qi)__B); __C = vec_mul(__C, __E); __D = vec_mul(__D, __F); const __v16qu __odds = {0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}; const __v16qu __evens = {2, 3, 6, 7, 10, 11, 14, 15, 18, 19, 22, 23, 26, 27, 30, 31}; __E = vec_perm(__C, __D, __odds); __F = vec_perm(__C, __D, __evens); return (__m128i)vec_adds(__E, __F); } extern __inline __m64 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_maddubs_pi16(__m64 __A, __m64 __B) { __v8hi __C = (__v8hi)(__v2du){__A, __A}; __C = vec_unpackl((__v16qi)__C); const __v8hi __unsigned = vec_splats((signed short)0x00ff); __C = vec_and(__C, __unsigned); __v8hi __D = (__v8hi)(__v2du){__B, __B}; __D = vec_unpackl((__v16qi)__D); __D = vec_mul(__C, __D); const __v16qu __odds = {0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}; const __v16qu __evens = {2, 3, 6, 7, 10, 11, 14, 15, 18, 19, 22, 23, 26, 27, 30, 31}; __C = vec_perm(__D, __D, __odds); __D = vec_perm(__D, __D, __evens); __C = vec_adds(__C, __D); return (__m64)((__v2du)(__C))[0]; } extern __inline __m128i __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_mulhrs_epi16(__m128i __A, __m128i __B) { __v4si __C = vec_unpackh((__v8hi)__A); __v4si __D = vec_unpackh((__v8hi)__B); __C = vec_mul(__C, __D); __D = vec_unpackl((__v8hi)__A); __v4si __E = vec_unpackl((__v8hi)__B); __D = vec_mul(__D, __E); const __v4su __shift = vec_splats((unsigned int)14); __C = vec_sr(__C, __shift); __D = vec_sr(__D, __shift); const __v4si __ones = vec_splats((signed int)1); __C = vec_add(__C, __ones); __C = vec_sr(__C, (__v4su)__ones); __D = vec_add(__D, __ones); __D = vec_sr(__D, (__v4su)__ones); return (__m128i)vec_pack(__C, __D); } extern __inline __m64 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) _mm_mulhrs_pi16(__m64 __A, __m64 __B) { __v4si __C = (__v4si)(__v2du){__A, __A}; __C = vec_unpackh((__v8hi)__C); __v4si __D = (__v4si)(__v2du){__B, __B}; __D = vec_unpackh((__v8hi)__D); __C = vec_mul(__C, __D); const __v4su __shift = vec_splats((unsigned int)14); __C = vec_sr(__C, __shift); const __v4si __ones = vec_splats((signed int)1); __C = vec_add(__C, __ones); __C = vec_sr(__C, (__v4su)__ones); __v8hi __E = vec_pack(__C, __D); return (__m64)((__v2du)(__E))[0]; } #else #include_next <tmmintrin.h> #endif /* defined(__powerpc64__) && \ * (defined(__linux__) || defined(__FreeBSD__) || defined(_AIX)) */ #endif /* TMMINTRIN_H_ */