//===----------------------------------------------------------------------===// // // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. // See https://llvm.org/LICENSE.txt for license information. // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception // //===----------------------------------------------------------------------===// #ifndef _LIBCPP___ALGORITHM_SIMD_UTILS_H #define _LIBCPP___ALGORITHM_SIMD_UTILS_H #include <__algorithm/min.h> #include <__bit/bit_cast.h> #include <__bit/countl.h> #include <__bit/countr.h> #include <__config> #include <__type_traits/is_arithmetic.h> #include <__type_traits/is_same.h> #include <__utility/integer_sequence.h> #include #include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) # pragma GCC system_header #endif _LIBCPP_PUSH_MACROS #include <__undef_macros> // TODO: Find out how altivec changes things and allow vectorizations there too. #if _LIBCPP_STD_VER >= 14 && (defined(_LIBCPP_CLANG_VER) && _LIBCPP_CLANG_VER >= 1500) && !defined(__ALTIVEC__) # define _LIBCPP_HAS_ALGORITHM_VECTOR_UTILS 1 #else # define _LIBCPP_HAS_ALGORITHM_VECTOR_UTILS 0 #endif #if _LIBCPP_HAS_ALGORITHM_VECTOR_UTILS && !defined(__OPTIMIZE_SIZE__) # define _LIBCPP_VECTORIZE_ALGORITHMS 1 #else # define _LIBCPP_VECTORIZE_ALGORITHMS 0 #endif #if _LIBCPP_HAS_ALGORITHM_VECTOR_UTILS _LIBCPP_BEGIN_NAMESPACE_STD template inline constexpr bool __can_map_to_integer_v = sizeof(_Tp) == alignof(_Tp) && (sizeof(_Tp) == 1 || sizeof(_Tp) == 2 || sizeof(_Tp) == 4 || sizeof(_Tp) == 8); template struct __get_as_integer_type_impl; template <> struct __get_as_integer_type_impl<1> { using type = uint8_t; }; template <> struct __get_as_integer_type_impl<2> { using type = uint16_t; }; template <> struct __get_as_integer_type_impl<4> { using type = uint32_t; }; template <> struct __get_as_integer_type_impl<8> { using type = uint64_t; }; template using __get_as_integer_type_t = typename __get_as_integer_type_impl::type; // This isn't specialized for 64 byte vectors on purpose. They have the potential to significantly reduce performance // in mixed simd/non-simd workloads and don't provide any performance improvement for currently vectorized algorithms // as far as benchmarks are concerned. # if defined(__AVX__) || defined(__MVS__) template inline constexpr size_t __native_vector_size = 32 / sizeof(_Tp); # elif defined(__SSE__) || defined(__ARM_NEON__) template inline constexpr size_t __native_vector_size = 16 / sizeof(_Tp); # elif defined(__MMX__) template inline constexpr size_t __native_vector_size = 8 / sizeof(_Tp); # else template inline constexpr size_t __native_vector_size = 1; # endif template using __simd_vector __attribute__((__ext_vector_type__(_Np))) = _ArithmeticT; template inline constexpr size_t __simd_vector_size_v = []() -> size_t { static_assert(_False, "Not a vector!"); }(); template inline constexpr size_t __simd_vector_size_v<__simd_vector<_Tp, _Np>> = _Np; template _LIBCPP_HIDE_FROM_ABI _Tp __simd_vector_underlying_type_impl(__simd_vector<_Tp, _Np>) { return _Tp{}; } template using __simd_vector_underlying_type_t = decltype(std::__simd_vector_underlying_type_impl(_VecT{})); // This isn't inlined without always_inline when loading chars. template _LIBCPP_NODISCARD _LIBCPP_ALWAYS_INLINE _LIBCPP_HIDE_FROM_ABI _VecT __load_vector(_Iter __iter) noexcept { return [=](index_sequence<_Indices...>) _LIBCPP_ALWAYS_INLINE noexcept { return _VecT{__iter[_Indices]...}; }(make_index_sequence<__simd_vector_size_v<_VecT>>{}); } template _LIBCPP_NODISCARD _LIBCPP_HIDE_FROM_ABI bool __all_of(__simd_vector<_Tp, _Np> __vec) noexcept { return __builtin_reduce_and(__builtin_convertvector(__vec, __simd_vector)); } template _LIBCPP_NODISCARD _LIBCPP_HIDE_FROM_ABI size_t __find_first_set(__simd_vector<_Tp, _Np> __vec) noexcept { using __mask_vec = __simd_vector; // This has MSan disabled du to https://github.com/llvm/llvm-project/issues/85876 auto __impl = [&](_MaskT) _LIBCPP_NO_SANITIZE("memory") noexcept { # if defined(_LIBCPP_BIG_ENDIAN) return std::min( _Np, std::__countl_zero(__builtin_bit_cast(_MaskT, __builtin_convertvector(__vec, __mask_vec)))); # else return std::min( _Np, std::__countr_zero(__builtin_bit_cast(_MaskT, __builtin_convertvector(__vec, __mask_vec)))); # endif }; if constexpr (sizeof(__mask_vec) == sizeof(uint8_t)) { return __impl(uint8_t{}); } else if constexpr (sizeof(__mask_vec) == sizeof(uint16_t)) { return __impl(uint16_t{}); } else if constexpr (sizeof(__mask_vec) == sizeof(uint32_t)) { return __impl(uint32_t{}); } else if constexpr (sizeof(__mask_vec) == sizeof(uint64_t)) { return __impl(uint64_t{}); } else { static_assert(sizeof(__mask_vec) == 0, "unexpected required size for mask integer type"); return 0; } } template _LIBCPP_NODISCARD _LIBCPP_HIDE_FROM_ABI size_t __find_first_not_set(__simd_vector<_Tp, _Np> __vec) noexcept { return std::__find_first_set(~__vec); } _LIBCPP_END_NAMESPACE_STD #endif // _LIBCPP_HAS_ALGORITHM_VECTOR_UTILS _LIBCPP_POP_MACROS #endif // _LIBCPP___ALGORITHM_SIMD_UTILS_H