|
#define | SSE2NEON_PRECISE_MINMAX |
|
#define | SSE2NEON_PRECISE_DIV |
|
#define | SSE2NEON_PRECISE_SQRT |
|
#define | SSE2NEON_PRECISE_DP |
|
#define | SSE2NEON_INCLUDE_WINDOWS_H |
|
#define | _sse2neon_const |
|
#define | _sse2neon_define0(type, a, body) |
|
#define | _sse2neon_define1(type, a, body) |
|
#define | _sse2neon_define2(type, a, b, body) |
|
#define | _sse2neon_return(ret) |
|
#define | _sse2neon_init(...) |
|
#define | SSE2NEON_BARRIER() |
|
#define | SSE2NEON_CACHELINE_SIZE |
|
#define | __has_builtin(x) |
|
#define | _MM_SHUFFLE(fp3, fp2, fp1, fp0) |
|
#define | _MM_SHUFFLE2(fp1, fp0) |
|
#define | _MM_FROUND_TO_NEAREST_INT |
|
#define | _MM_FROUND_TO_NEG_INF |
|
#define | _MM_FROUND_TO_POS_INF |
|
#define | _MM_FROUND_TO_ZERO |
|
#define | _MM_FROUND_CUR_DIRECTION |
|
#define | _MM_FROUND_NO_EXC |
|
#define | _MM_FROUND_RAISE_EXC |
|
#define | _MM_FROUND_NINT |
|
#define | _MM_FROUND_FLOOR |
|
#define | _MM_FROUND_CEIL |
|
#define | _MM_FROUND_TRUNC |
|
#define | _MM_FROUND_RINT |
|
#define | _MM_FROUND_NEARBYINT |
|
#define | _MM_ROUND_NEAREST |
|
#define | _MM_ROUND_DOWN |
|
#define | _MM_ROUND_UP |
|
#define | _MM_ROUND_TOWARD_ZERO |
|
#define | _MM_FLUSH_ZERO_MASK |
|
#define | _MM_FLUSH_ZERO_ON |
|
#define | _MM_FLUSH_ZERO_OFF |
|
#define | _MM_DENORMALS_ZERO_MASK |
|
#define | _MM_DENORMALS_ZERO_ON |
|
#define | _MM_DENORMALS_ZERO_OFF |
|
#define | __constrange(a, b) |
|
#define | __int64 |
|
#define | vreinterpretq_m128_f16(x) |
|
#define | vreinterpretq_m128_f32(x) |
|
#define | vreinterpretq_m128_f64(x) |
|
#define | vreinterpretq_m128_u8(x) |
|
#define | vreinterpretq_m128_u16(x) |
|
#define | vreinterpretq_m128_u32(x) |
|
#define | vreinterpretq_m128_u64(x) |
|
#define | vreinterpretq_m128_s8(x) |
|
#define | vreinterpretq_m128_s16(x) |
|
#define | vreinterpretq_m128_s32(x) |
|
#define | vreinterpretq_m128_s64(x) |
|
#define | vreinterpretq_f16_m128(x) |
|
#define | vreinterpretq_f32_m128(x) |
|
#define | vreinterpretq_f64_m128(x) |
|
#define | vreinterpretq_u8_m128(x) |
|
#define | vreinterpretq_u16_m128(x) |
|
#define | vreinterpretq_u32_m128(x) |
|
#define | vreinterpretq_u64_m128(x) |
|
#define | vreinterpretq_s8_m128(x) |
|
#define | vreinterpretq_s16_m128(x) |
|
#define | vreinterpretq_s32_m128(x) |
|
#define | vreinterpretq_s64_m128(x) |
|
#define | vreinterpretq_m128i_s8(x) |
|
#define | vreinterpretq_m128i_s16(x) |
|
#define | vreinterpretq_m128i_s32(x) |
|
#define | vreinterpretq_m128i_s64(x) |
|
#define | vreinterpretq_m128i_u8(x) |
|
#define | vreinterpretq_m128i_u16(x) |
|
#define | vreinterpretq_m128i_u32(x) |
|
#define | vreinterpretq_m128i_u64(x) |
|
#define | vreinterpretq_f32_m128i(x) |
|
#define | vreinterpretq_f64_m128i(x) |
|
#define | vreinterpretq_s8_m128i(x) |
|
#define | vreinterpretq_s16_m128i(x) |
|
#define | vreinterpretq_s32_m128i(x) |
|
#define | vreinterpretq_s64_m128i(x) |
|
#define | vreinterpretq_u8_m128i(x) |
|
#define | vreinterpretq_u16_m128i(x) |
|
#define | vreinterpretq_u32_m128i(x) |
|
#define | vreinterpretq_u64_m128i(x) |
|
#define | vreinterpret_m64_s8(x) |
|
#define | vreinterpret_m64_s16(x) |
|
#define | vreinterpret_m64_s32(x) |
|
#define | vreinterpret_m64_s64(x) |
|
#define | vreinterpret_m64_u8(x) |
|
#define | vreinterpret_m64_u16(x) |
|
#define | vreinterpret_m64_u32(x) |
|
#define | vreinterpret_m64_u64(x) |
|
#define | vreinterpret_m64_f16(x) |
|
#define | vreinterpret_m64_f32(x) |
|
#define | vreinterpret_m64_f64(x) |
|
#define | vreinterpret_u8_m64(x) |
|
#define | vreinterpret_u16_m64(x) |
|
#define | vreinterpret_u32_m64(x) |
|
#define | vreinterpret_u64_m64(x) |
|
#define | vreinterpret_s8_m64(x) |
|
#define | vreinterpret_s16_m64(x) |
|
#define | vreinterpret_s32_m64(x) |
|
#define | vreinterpret_s64_m64(x) |
|
#define | vreinterpret_f32_m64(x) |
|
#define | vreinterpretq_m128d_s32(x) |
|
#define | vreinterpretq_m128d_s64(x) |
|
#define | vreinterpretq_m128d_u32(x) |
|
#define | vreinterpretq_m128d_u64(x) |
|
#define | vreinterpretq_m128d_f32(x) |
|
#define | vreinterpretq_s64_m128d(x) |
|
#define | vreinterpretq_u32_m128d(x) |
|
#define | vreinterpretq_u64_m128d(x) |
|
#define | vreinterpretq_f32_m128d(x) |
|
#define | vreinterpretq_nth_u64_m128i(x, n) |
|
#define | vreinterpretq_nth_u32_m128i(x, n) |
|
#define | vreinterpretq_nth_u8_m128i(x, n) |
|
#define | _MM_GET_FLUSH_ZERO_MODE |
|
#define | _MM_SET_FLUSH_ZERO_MODE |
|
#define | _MM_GET_DENORMALS_ZERO_MODE |
|
#define | _MM_SET_DENORMALS_ZERO_MODE |
|
#define | _mm_shuffle_epi32_default(a, imm) |
|
#define | _mm_shuffle_epi32_splat(a, imm) |
|
#define | _mm_shuffle_ps_default(a, b, imm) |
|
#define | _mm_shufflelo_epi16_function(a, imm) |
|
#define | _mm_shufflehi_epi16_function(a, imm) |
|
#define | _mm_cvtps_pi32(a) |
|
#define | _mm_cvtsi32_ss(a, b) |
|
#define | _mm_cvtss_si32(a) |
|
#define | _mm_cvttps_pi32(a) |
|
#define | _mm_cvttss_si32(a) |
|
#define | _mm_extract_pi16(a, imm) |
|
#define | _mm_insert_pi16(a, b, imm) |
|
#define | _mm_load_ps1 |
|
#define | _m_maskmovq(a, mask, mem_addr) |
|
#define | _m_pavgb(a, b) |
|
#define | _m_pavgw(a, b) |
|
#define | _m_pextrw(a, imm) |
|
#define | _m_pinsrw(a, i, imm) |
|
#define | _m_pmaxsw(a, b) |
|
#define | _m_pmaxub(a, b) |
|
#define | _m_pminsw(a, b) |
|
#define | _m_pminub(a, b) |
|
#define | _m_pmovmskb(a) |
|
#define | _m_pmulhuw(a, b) |
|
#define | _m_psadbw(a, b) |
|
#define | _m_pshufw(a, imm) |
|
#define | _mm_shuffle_pi16(a, imm) |
|
#define | _mm_shuffle_ps(a, b, imm) |
|
#define | _mm_store1_ps |
|
#define | _MM_TRANSPOSE4_PS(row0, row1, row2, row3) |
|
#define | _mm_ucomieq_ss |
|
#define | _mm_ucomige_ss |
|
#define | _mm_ucomigt_ss |
|
#define | _mm_ucomile_ss |
|
#define | _mm_ucomilt_ss |
|
#define | _mm_ucomineq_ss |
|
#define | _mm_bslli_si128(a, imm) |
|
#define | _mm_bsrli_si128(a, imm) |
|
#define | _mm_cvtsd_si64x |
|
#define | _mm_cvtsi128_si64x(a) |
|
#define | _mm_cvtsi128_si64x(a) |
|
#define | _mm_cvtsi64x_si128(a) |
|
#define | _mm_cvtsi64x_sd(a, b) |
|
#define | _mm_cvttsd_si64x(a) |
|
#define | _mm_extract_epi16(a, imm) |
|
#define | _mm_insert_epi16(a, b, imm) |
|
#define | _mm_load_pd1 |
|
#define | _mm_set_pd1 |
|
#define | _mm_shuffle_epi32(a, imm) |
|
#define | _mm_shuffle_pd(a, b, imm8) |
|
#define | _mm_shufflehi_epi16(a, imm) |
|
#define | _mm_shufflelo_epi16(a, imm) |
|
#define | _mm_slli_si128(a, imm) |
|
#define | _mm_srai_epi32(a, imm) |
|
#define | _mm_srli_epi16(a, imm) |
|
#define | _mm_srli_epi32(a, imm) |
|
#define | _mm_srli_epi64(a, imm) |
|
#define | _mm_srli_si128(a, imm) |
|
#define | _mm_store1_pd |
|
#define | _mm_ucomieq_sd |
|
#define | _mm_ucomige_sd |
|
#define | _mm_ucomigt_sd |
|
#define | _mm_ucomile_sd |
|
#define | _mm_ucomilt_sd |
|
#define | _mm_ucomineq_sd |
|
#define | _mm_lddqu_si128 |
|
#define | _mm_loaddup_pd |
|
#define | _mm_alignr_epi8(a, b, imm) |
|
#define | _mm_alignr_pi8(a, b, imm) |
|
#define | _mm_blend_epi16(a, b, imm) |
|
#define | _mm_blend_pd(a, b, imm) |
|
#define | _mm_extract_epi32(a, imm) |
|
#define | _mm_extract_epi64(a, imm) |
|
#define | _mm_extract_epi8(a, imm) |
|
#define | _mm_extract_ps(a, imm) |
|
#define | _mm_insert_epi32(a, b, imm) |
|
#define | _mm_insert_epi64(a, b, imm) |
|
#define | _mm_insert_epi8(a, b, imm) |
|
#define | _mm_insert_ps(a, b, imm8) |
|
#define | _mm_testnzc_si128(a, b) |
|
#define | _SIDD_UBYTE_OPS |
|
#define | _SIDD_UWORD_OPS |
|
#define | _SIDD_SBYTE_OPS |
|
#define | _SIDD_SWORD_OPS |
|
#define | _SIDD_CMP_EQUAL_ANY |
|
#define | _SIDD_CMP_RANGES |
|
#define | _SIDD_CMP_EQUAL_EACH |
|
#define | _SIDD_CMP_EQUAL_ORDERED |
|
#define | _SIDD_POSITIVE_POLARITY |
|
#define | _SIDD_MASKED_POSITIVE_POLARITY |
|
#define | _SIDD_NEGATIVE_POLARITY |
|
#define | _SIDD_MASKED_NEGATIVE_POLARITY |
|
#define | _SIDD_LEAST_SIGNIFICANT |
|
#define | _SIDD_MOST_SIGNIFICANT |
|
#define | _SIDD_BIT_MASK |
|
#define | _SIDD_UNIT_MASK |
|
#define | SSE2NEON_PRIMITIVE_CAT(a, ...) |
|
#define | SSE2NEON_CAT(a, b) |
|
#define | SSE2NEON_IIF(c) |
|
#define | SSE2NEON_IIF_0(t, ...) |
|
#define | SSE2NEON_IIF_1(t, ...) |
|
#define | SSE2NEON_COMPL(b) |
|
#define | SSE2NEON_COMPL_0 |
|
#define | SSE2NEON_COMPL_1 |
|
#define | SSE2NEON_DEC(x) |
|
#define | SSE2NEON_DEC_1 |
|
#define | SSE2NEON_DEC_2 |
|
#define | SSE2NEON_DEC_3 |
|
#define | SSE2NEON_DEC_4 |
|
#define | SSE2NEON_DEC_5 |
|
#define | SSE2NEON_DEC_6 |
|
#define | SSE2NEON_DEC_7 |
|
#define | SSE2NEON_DEC_8 |
|
#define | SSE2NEON_DEC_9 |
|
#define | SSE2NEON_DEC_10 |
|
#define | SSE2NEON_DEC_11 |
|
#define | SSE2NEON_DEC_12 |
|
#define | SSE2NEON_DEC_13 |
|
#define | SSE2NEON_DEC_14 |
|
#define | SSE2NEON_DEC_15 |
|
#define | SSE2NEON_DEC_16 |
|
#define | SSE2NEON_CHECK_N(x, n, ...) |
|
#define | SSE2NEON_CHECK(...) |
|
#define | SSE2NEON_PROBE(x) |
|
#define | SSE2NEON_NOT(x) |
|
#define | SSE2NEON_NOT_0 |
|
#define | SSE2NEON_BOOL(x) |
|
#define | SSE2NEON_IF(c) |
|
#define | SSE2NEON_EAT(...) |
|
#define | SSE2NEON_EXPAND(...) |
|
#define | SSE2NEON_WHEN(c) |
|
#define | SSE2NEON_EMPTY() |
|
#define | SSE2NEON_DEFER(id) |
|
#define | SSE2NEON_OBSTRUCT(...) |
|
#define | SSE2NEON_EXPAND(...) |
|
#define | SSE2NEON_EVAL(...) |
|
#define | SSE2NEON_EVAL1(...) |
|
#define | SSE2NEON_EVAL2(...) |
|
#define | SSE2NEON_EVAL3(...) |
|
#define | SSE2NEON_REPEAT(count, macro, ...) |
|
#define | SSE2NEON_REPEAT_INDIRECT() |
|
#define | SSE2NEON_SIZE_OF_byte |
|
#define | SSE2NEON_NUMBER_OF_LANES_byte |
|
#define | SSE2NEON_SIZE_OF_word |
|
#define | SSE2NEON_NUMBER_OF_LANES_word |
|
#define | SSE2NEON_COMPARE_EQUAL_THEN_FILL_LANE(i, type) |
|
#define | SSE2NEON_FILL_LANE(i, type) |
|
#define | PCMPSTR_RANGES(a, b, mtx, data_type_prefix, type_prefix, size, number_of_lanes, byte_or_word) |
|
#define | PCMPSTR_EQ(a, b, mtx, size, number_of_lanes) |
|
#define | SSE2NEON_CMP_EQUAL_ANY_IMPL(type) |
|
#define | SSE2NEON_CMP_RANGES_IMPL(type, data_type, us, byte_or_word) |
|
#define | SSE2NEON_CMP_EQUAL_ORDERED_IMPL(type) |
|
#define | SSE2NEON_GENERATE_CMP_EQUAL_ANY(prefix) |
|
#define | SSE2NEON_CMP_RANGES_IS_BYTE |
|
#define | SSE2NEON_CMP_RANGES_IS_WORD |
|
#define | SSE2NEON_GENERATE_CMP_RANGES(prefix) |
|
#define | SSE2NEON_AGGREGATE_EQUAL_ORDER_IS_UBYTE |
|
#define | SSE2NEON_AGGREGATE_EQUAL_ORDER_IS_UWORD |
|
#define | SSE2NEON_AGGREGATE_EQUAL_ORDER_IMPL(size, number_of_lanes, data_type) |
|
#define | SSE2NEON_GENERATE_AGGREGATE_EQUAL_ORDER(prefix) |
|
#define | SSE2NEON_GENERATE_CMP_EQUAL_ORDERED(prefix) |
|
#define | _(name, func_suffix) |
|
#define | _(name, func_suffix) |
|
#define | SSE2NEON_MIN(x, y) |
|
#define | SSE2NEON_CMPSTR_SET_UPPER(var, imm) |
|
#define | SSE2NEON_CMPESTRX_LEN_PAIR(a, b, la, lb) |
|
#define | SSE2NEON_COMP_AGG(a, b, la, lb, imm8, IE) |
|
#define | SSE2NEON_CMPSTR_GENERATE_INDEX(r2, bound, imm8) |
|
#define | SSE2NEON_CMPSTR_GENERATE_MASK(dst) |
|
#define | SSE2NEON_CMPISTRX_LENGTH(str, len, imm8) |
|
#define | SSE2NEON_CMPISTRX_LEN_PAIR(a, b, la, lb) |
|
#define | SSE2NEON_AES_SBOX(w) |
|
#define | SSE2NEON_AES_RSBOX(w) |
|
#define | SSE2NEON_AES_H0(x) |
|
#define | SSE2NEON_XT(x) |
|
#define | SSE2NEON_MULTIPLY(x, y) |
|
#define | SSE2NEON_AES_B2W(b0, b1, b2, b3) |
|
#define | SSE2NEON_AES_F2(x) |
|
#define | SSE2NEON_AES_F3(x) |
|
#define | SSE2NEON_AES_U0(p) |
|
#define | SSE2NEON_AES_U1(p) |
|
#define | SSE2NEON_AES_U2(p) |
|
#define | SSE2NEON_AES_U3(p) |
|
|
FORCE_INLINE double | sse2neon_recast_u64_f64 (uint64_t val) |
|
FORCE_INLINE int64_t | sse2neon_recast_f64_s64 (double val) |
|
FORCE_INLINE void | _sse2neon_smp_mb (void) |
|
union | ALIGN_STRUCT (16) SIMDVec |
|
FORCE_INLINE unsigned int | _MM_GET_ROUNDING_MODE (void) |
|
FORCE_INLINE __m128 | _mm_move_ss (__m128, __m128) |
|
FORCE_INLINE __m128 | _mm_or_ps (__m128, __m128) |
|
FORCE_INLINE __m128 | _mm_set_ps1 (float) |
|
FORCE_INLINE __m128 | _mm_setzero_ps (void) |
|
FORCE_INLINE __m128i | _mm_and_si128 (__m128i, __m128i) |
|
FORCE_INLINE __m128i | _mm_castps_si128 (__m128) |
|
FORCE_INLINE __m128i | _mm_cmpeq_epi32 (__m128i, __m128i) |
|
FORCE_INLINE __m128i | _mm_cvtps_epi32 (__m128) |
|
FORCE_INLINE __m128d | _mm_move_sd (__m128d, __m128d) |
|
FORCE_INLINE __m128i | _mm_or_si128 (__m128i, __m128i) |
|
FORCE_INLINE __m128i | _mm_set_epi32 (int, int, int, int) |
|
FORCE_INLINE __m128i | _mm_set_epi64x (int64_t, int64_t) |
|
FORCE_INLINE __m128d | _mm_set_pd (double, double) |
|
FORCE_INLINE __m128i | _mm_set1_epi32 (int) |
|
FORCE_INLINE __m128i | _mm_setzero_si128 (void) |
|
FORCE_INLINE __m128d | _mm_ceil_pd (__m128d) |
|
FORCE_INLINE __m128 | _mm_ceil_ps (__m128) |
|
FORCE_INLINE __m128d | _mm_floor_pd (__m128d) |
|
FORCE_INLINE __m128 | _mm_floor_ps (__m128) |
|
FORCE_INLINE __m128d | _mm_round_pd (__m128d, int) |
|
FORCE_INLINE __m128 | _mm_round_ps (__m128, int) |
|
FORCE_INLINE uint32_t | _mm_crc32_u8 (uint32_t, uint8_t) |
|
FORCE_INLINE uint8x16x4_t | _sse2neon_vld1q_u8_x4 (const uint8_t *p) |
|
FORCE_INLINE uint8_t | _sse2neon_vaddv_u8 (uint8x8_t v8) |
|
FORCE_INLINE uint8_t | _sse2neon_vaddvq_u8 (uint8x16_t a) |
|
FORCE_INLINE uint16_t | _sse2neon_vaddvq_u16 (uint16x8_t a) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_1032 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_2301 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_0321 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_2103 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_1010 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_1001 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_0101 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_3210 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_0011 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_0022 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_2200 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_3202 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_1133 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_2010 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_2001 (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_shuffle_ps_2032 (__m128 a, __m128 b) |
|
static uint64x2_t | _sse2neon_vmull_p64 (uint64x1_t _a, uint64x1_t _b) |
|
FORCE_INLINE __m128i | _mm_shuffle_epi_1032 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_shuffle_epi_2301 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_shuffle_epi_0321 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_shuffle_epi_2103 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_shuffle_epi_1010 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_shuffle_epi_1001 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_shuffle_epi_0101 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_shuffle_epi_2211 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_shuffle_epi_0122 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_shuffle_epi_3332 (__m128i a) |
|
FORCE_INLINE void | _mm_empty (void) |
|
FORCE_INLINE __m128 | _mm_add_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_add_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_and_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_andnot_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m64 | _mm_avg_pu16 (__m64 a, __m64 b) |
|
FORCE_INLINE __m64 | _mm_avg_pu8 (__m64 a, __m64 b) |
|
FORCE_INLINE __m128 | _mm_cmpeq_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpeq_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpge_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpge_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpgt_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpgt_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmple_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmple_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmplt_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmplt_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpneq_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpneq_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpnge_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpnge_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpngt_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpngt_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpnle_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpnle_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpnlt_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpnlt_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpord_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpord_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpunord_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cmpunord_ss (__m128 a, __m128 b) |
|
FORCE_INLINE int | _mm_comieq_ss (__m128 a, __m128 b) |
|
FORCE_INLINE int | _mm_comige_ss (__m128 a, __m128 b) |
|
FORCE_INLINE int | _mm_comigt_ss (__m128 a, __m128 b) |
|
FORCE_INLINE int | _mm_comile_ss (__m128 a, __m128 b) |
|
FORCE_INLINE int | _mm_comilt_ss (__m128 a, __m128 b) |
|
FORCE_INLINE int | _mm_comineq_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_cvt_pi2ps (__m128 a, __m64 b) |
|
FORCE_INLINE __m64 | _mm_cvt_ps2pi (__m128 a) |
|
FORCE_INLINE __m128 | _mm_cvt_si2ss (__m128 a, int b) |
|
FORCE_INLINE int | _mm_cvt_ss2si (__m128 a) |
|
FORCE_INLINE __m128 | _mm_cvtpi16_ps (__m64 a) |
|
FORCE_INLINE __m128 | _mm_cvtpi32_ps (__m128 a, __m64 b) |
|
FORCE_INLINE __m128 | _mm_cvtpi32x2_ps (__m64 a, __m64 b) |
|
FORCE_INLINE __m128 | _mm_cvtpi8_ps (__m64 a) |
|
FORCE_INLINE __m64 | _mm_cvtps_pi16 (__m128 a) |
|
FORCE_INLINE __m64 | _mm_cvtps_pi8 (__m128 a) |
|
FORCE_INLINE __m128 | _mm_cvtpu16_ps (__m64 a) |
|
FORCE_INLINE __m128 | _mm_cvtpu8_ps (__m64 a) |
|
FORCE_INLINE __m128 | _mm_cvtsi64_ss (__m128 a, int64_t b) |
|
FORCE_INLINE float | _mm_cvtss_f32 (__m128 a) |
|
FORCE_INLINE int64_t | _mm_cvtss_si64 (__m128 a) |
|
FORCE_INLINE __m64 | _mm_cvtt_ps2pi (__m128 a) |
|
FORCE_INLINE int | _mm_cvtt_ss2si (__m128 a) |
|
FORCE_INLINE int64_t | _mm_cvttss_si64 (__m128 a) |
|
FORCE_INLINE __m128 | _mm_div_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_div_ss (__m128 a, __m128 b) |
|
FORCE_INLINE void | _mm_free (void *addr) |
|
FORCE_INLINE uint64_t | _sse2neon_get_fpcr (void) |
|
FORCE_INLINE void | _sse2neon_set_fpcr (uint64_t value) |
|
FORCE_INLINE unsigned int | _sse2neon_mm_get_flush_zero_mode (void) |
|
FORCE_INLINE __m128 | _mm_load_ps (const float *p) |
|
FORCE_INLINE __m128 | _mm_load_ss (const float *p) |
|
FORCE_INLINE __m128 | _mm_load1_ps (const float *p) |
|
FORCE_INLINE __m128 | _mm_loadh_pi (__m128 a, __m64 const *p) |
|
FORCE_INLINE __m128 | _mm_loadl_pi (__m128 a, __m64 const *p) |
|
FORCE_INLINE __m128 | _mm_loadr_ps (const float *p) |
|
FORCE_INLINE __m128 | _mm_loadu_ps (const float *p) |
|
FORCE_INLINE __m128i | _mm_loadu_si16 (const void *p) |
|
FORCE_INLINE __m128i | _mm_loadu_si64 (const void *p) |
|
FORCE_INLINE void * | _mm_malloc (size_t size, size_t align) |
|
FORCE_INLINE void | _mm_maskmove_si64 (__m64 a, __m64 mask, char *mem_addr) |
|
FORCE_INLINE __m64 | _mm_max_pi16 (__m64 a, __m64 b) |
|
FORCE_INLINE __m128 | _mm_max_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m64 | _mm_max_pu8 (__m64 a, __m64 b) |
|
FORCE_INLINE __m128 | _mm_max_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m64 | _mm_min_pi16 (__m64 a, __m64 b) |
|
FORCE_INLINE __m128 | _mm_min_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m64 | _mm_min_pu8 (__m64 a, __m64 b) |
|
FORCE_INLINE __m128 | _mm_min_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_movehl_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_movelh_ps (__m128 __A, __m128 __B) |
|
FORCE_INLINE int | _mm_movemask_pi8 (__m64 a) |
|
FORCE_INLINE int | _mm_movemask_ps (__m128 a) |
|
FORCE_INLINE __m128 | _mm_mul_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_mul_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m64 | _mm_mulhi_pu16 (__m64 a, __m64 b) |
|
FORCE_INLINE void | _mm_prefetch (char const *p, int i) |
|
FORCE_INLINE __m128 | _mm_rcp_ps (__m128 in) |
|
FORCE_INLINE __m128 | _mm_rcp_ss (__m128 a) |
|
FORCE_INLINE __m128 | _mm_rsqrt_ps (__m128 in) |
|
FORCE_INLINE __m128 | _mm_rsqrt_ss (__m128 in) |
|
FORCE_INLINE __m64 | _mm_sad_pu8 (__m64 a, __m64 b) |
|
FORCE_INLINE void | _sse2neon_mm_set_flush_zero_mode (unsigned int flag) |
|
FORCE_INLINE __m128 | _mm_set_ps (float w, float z, float y, float x) |
|
FORCE_INLINE void | _MM_SET_ROUNDING_MODE (int rounding) |
|
FORCE_INLINE __m128 | _mm_set_ss (float a) |
|
FORCE_INLINE __m128 | _mm_set1_ps (float _w) |
|
FORCE_INLINE void | _mm_setcsr (unsigned int a) |
|
FORCE_INLINE unsigned int | _mm_getcsr (void) |
|
FORCE_INLINE __m128 | _mm_setr_ps (float w, float z, float y, float x) |
|
FORCE_INLINE void | _mm_sfence (void) |
|
FORCE_INLINE void | _mm_mfence (void) |
|
FORCE_INLINE void | _mm_lfence (void) |
|
FORCE_INLINE __m128 | _mm_sqrt_ps (__m128 in) |
|
FORCE_INLINE __m128 | _mm_sqrt_ss (__m128 in) |
|
FORCE_INLINE void | _mm_store_ps (float *p, __m128 a) |
|
FORCE_INLINE void | _mm_store_ps1 (float *p, __m128 a) |
|
FORCE_INLINE void | _mm_store_ss (float *p, __m128 a) |
|
FORCE_INLINE void | _mm_storeh_pi (__m64 *p, __m128 a) |
|
FORCE_INLINE void | _mm_storel_pi (__m64 *p, __m128 a) |
|
FORCE_INLINE void | _mm_storer_ps (float *p, __m128 a) |
|
FORCE_INLINE void | _mm_storeu_ps (float *p, __m128 a) |
|
FORCE_INLINE void | _mm_storeu_si16 (void *p, __m128i a) |
|
FORCE_INLINE void | _mm_storeu_si64 (void *p, __m128i a) |
|
FORCE_INLINE void | _mm_stream_pi (__m64 *p, __m64 a) |
|
FORCE_INLINE void | _mm_stream_ps (float *p, __m128 a) |
|
FORCE_INLINE __m128 | _mm_sub_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_sub_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128i | _mm_undefined_si128 (void) |
|
FORCE_INLINE __m128 | _mm_undefined_ps (void) |
|
FORCE_INLINE __m128 | _mm_unpackhi_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_unpacklo_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128 | _mm_xor_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128i | _mm_add_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_add_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_add_epi64 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_add_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_add_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_add_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m64 | _mm_add_si64 (__m64 a, __m64 b) |
|
FORCE_INLINE __m128i | _mm_adds_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_adds_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_adds_epu16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_adds_epu8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_and_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_andnot_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128i | _mm_andnot_si128 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_avg_epu16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_avg_epu8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128 | _mm_castpd_ps (__m128d a) |
|
FORCE_INLINE __m128i | _mm_castpd_si128 (__m128d a) |
|
FORCE_INLINE __m128d | _mm_castps_pd (__m128 a) |
|
FORCE_INLINE __m128d | _mm_castsi128_pd (__m128i a) |
|
FORCE_INLINE __m128 | _mm_castsi128_ps (__m128i a) |
|
FORCE_INLINE void | _mm_clflush (void const *p) |
|
FORCE_INLINE __m128i | _mm_cmpeq_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_cmpeq_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_cmpeq_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpeq_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpge_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpge_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128i | _mm_cmpgt_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_cmpgt_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_cmpgt_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_cmpgt_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpgt_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmple_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmple_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128i | _mm_cmplt_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_cmplt_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_cmplt_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_cmplt_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmplt_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpneq_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpneq_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpnge_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpnge_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpngt_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpngt_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpnle_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpnle_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpnlt_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpnlt_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpord_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpord_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpunord_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cmpunord_sd (__m128d a, __m128d b) |
|
FORCE_INLINE int | _mm_comige_sd (__m128d a, __m128d b) |
|
FORCE_INLINE int | _mm_comigt_sd (__m128d a, __m128d b) |
|
FORCE_INLINE int | _mm_comile_sd (__m128d a, __m128d b) |
|
FORCE_INLINE int | _mm_comilt_sd (__m128d a, __m128d b) |
|
FORCE_INLINE int | _mm_comieq_sd (__m128d a, __m128d b) |
|
FORCE_INLINE int | _mm_comineq_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_cvtepi32_pd (__m128i a) |
|
FORCE_INLINE __m128 | _mm_cvtepi32_ps (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtpd_epi32 (__m128d a) |
|
FORCE_INLINE __m64 | _mm_cvtpd_pi32 (__m128d a) |
|
FORCE_INLINE __m128 | _mm_cvtpd_ps (__m128d a) |
|
FORCE_INLINE __m128d | _mm_cvtpi32_pd (__m64 a) |
|
FORCE_INLINE __m128d | _mm_cvtps_pd (__m128 a) |
|
FORCE_INLINE double | _mm_cvtsd_f64 (__m128d a) |
|
FORCE_INLINE int32_t | _mm_cvtsd_si32 (__m128d a) |
|
FORCE_INLINE int64_t | _mm_cvtsd_si64 (__m128d a) |
|
FORCE_INLINE __m128 | _mm_cvtsd_ss (__m128 a, __m128d b) |
|
FORCE_INLINE int | _mm_cvtsi128_si32 (__m128i a) |
|
FORCE_INLINE int64_t | _mm_cvtsi128_si64 (__m128i a) |
|
FORCE_INLINE __m128d | _mm_cvtsi32_sd (__m128d a, int32_t b) |
|
FORCE_INLINE __m128i | _mm_cvtsi32_si128 (int a) |
|
FORCE_INLINE __m128d | _mm_cvtsi64_sd (__m128d a, int64_t b) |
|
FORCE_INLINE __m128i | _mm_cvtsi64_si128 (int64_t a) |
|
FORCE_INLINE __m128d | _mm_cvtss_sd (__m128d a, __m128 b) |
|
FORCE_INLINE __m128i | _mm_cvttpd_epi32 (__m128d a) |
|
FORCE_INLINE __m64 | _mm_cvttpd_pi32 (__m128d a) |
|
FORCE_INLINE __m128i | _mm_cvttps_epi32 (__m128 a) |
|
FORCE_INLINE int32_t | _mm_cvttsd_si32 (__m128d a) |
|
FORCE_INLINE int64_t | _mm_cvttsd_si64 (__m128d a) |
|
FORCE_INLINE __m128d | _mm_div_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_div_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_load_pd (const double *p) |
|
FORCE_INLINE __m128d | _mm_load_sd (const double *p) |
|
FORCE_INLINE __m128i | _mm_load_si128 (const __m128i *p) |
|
FORCE_INLINE __m128d | _mm_load1_pd (const double *p) |
|
FORCE_INLINE __m128d | _mm_loadh_pd (__m128d a, const double *p) |
|
FORCE_INLINE __m128i | _mm_loadl_epi64 (__m128i const *p) |
|
FORCE_INLINE __m128d | _mm_loadl_pd (__m128d a, const double *p) |
|
FORCE_INLINE __m128d | _mm_loadr_pd (const double *p) |
|
FORCE_INLINE __m128d | _mm_loadu_pd (const double *p) |
|
FORCE_INLINE __m128i | _mm_loadu_si128 (const __m128i *p) |
|
FORCE_INLINE __m128i | _mm_loadu_si32 (const void *p) |
|
FORCE_INLINE __m128i | _mm_madd_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE void | _mm_maskmoveu_si128 (__m128i a, __m128i mask, char *mem_addr) |
|
FORCE_INLINE __m128i | _mm_max_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_max_epu8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_max_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_max_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128i | _mm_min_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_min_epu8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_min_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_min_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128i | _mm_move_epi64 (__m128i a) |
|
FORCE_INLINE int | _mm_movemask_epi8 (__m128i a) |
|
FORCE_INLINE int | _mm_movemask_pd (__m128d a) |
|
FORCE_INLINE __m64 | _mm_movepi64_pi64 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_movpi64_epi64 (__m64 a) |
|
FORCE_INLINE __m128i | _mm_mul_epu32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_mul_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_mul_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m64 | _mm_mul_su32 (__m64 a, __m64 b) |
|
FORCE_INLINE __m128i | _mm_mulhi_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_mulhi_epu16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_mullo_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_or_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128i | _mm_packs_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_packs_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_packus_epi16 (const __m128i a, const __m128i b) |
|
FORCE_INLINE void | _mm_pause (void) |
|
FORCE_INLINE __m128i | _mm_sad_epu8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_set_epi16 (short i7, short i6, short i5, short i4, short i3, short i2, short i1, short i0) |
|
FORCE_INLINE __m128i | _mm_set_epi64 (__m64 i1, __m64 i2) |
|
FORCE_INLINE __m128i | _mm_set_epi8 (signed char b15, signed char b14, signed char b13, signed char b12, signed char b11, signed char b10, signed char b9, signed char b8, signed char b7, signed char b6, signed char b5, signed char b4, signed char b3, signed char b2, signed char b1, signed char b0) |
|
FORCE_INLINE __m128d | _mm_set_sd (double a) |
|
FORCE_INLINE __m128i | _mm_set1_epi16 (short w) |
|
FORCE_INLINE __m128i | _mm_set1_epi64 (__m64 _i) |
|
FORCE_INLINE __m128i | _mm_set1_epi64x (int64_t _i) |
|
FORCE_INLINE __m128i | _mm_set1_epi8 (signed char w) |
|
FORCE_INLINE __m128d | _mm_set1_pd (double d) |
|
FORCE_INLINE __m128i | _mm_setr_epi16 (short w0, short w1, short w2, short w3, short w4, short w5, short w6, short w7) |
|
FORCE_INLINE __m128i | _mm_setr_epi32 (int i3, int i2, int i1, int i0) |
|
FORCE_INLINE __m128i | _mm_setr_epi64 (__m64 e1, __m64 e0) |
|
FORCE_INLINE __m128i | _mm_setr_epi8 (signed char b0, signed char b1, signed char b2, signed char b3, signed char b4, signed char b5, signed char b6, signed char b7, signed char b8, signed char b9, signed char b10, signed char b11, signed char b12, signed char b13, signed char b14, signed char b15) |
|
FORCE_INLINE __m128d | _mm_setr_pd (double e1, double e0) |
|
FORCE_INLINE __m128d | _mm_setzero_pd (void) |
|
FORCE_INLINE __m128i | _mm_sll_epi16 (__m128i a, __m128i count) |
|
FORCE_INLINE __m128i | _mm_sll_epi32 (__m128i a, __m128i count) |
|
FORCE_INLINE __m128i | _mm_sll_epi64 (__m128i a, __m128i count) |
|
FORCE_INLINE __m128i | _mm_slli_epi16 (__m128i a, int imm) |
|
FORCE_INLINE __m128i | _mm_slli_epi32 (__m128i a, int imm) |
|
FORCE_INLINE __m128i | _mm_slli_epi64 (__m128i a, int imm) |
|
FORCE_INLINE __m128d | _mm_sqrt_pd (__m128d a) |
|
FORCE_INLINE __m128d | _mm_sqrt_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128i | _mm_sra_epi16 (__m128i a, __m128i count) |
|
FORCE_INLINE __m128i | _mm_sra_epi32 (__m128i a, __m128i count) |
|
FORCE_INLINE __m128i | _mm_srai_epi16 (__m128i a, int imm) |
|
FORCE_INLINE __m128i | _mm_srl_epi16 (__m128i a, __m128i count) |
|
FORCE_INLINE __m128i | _mm_srl_epi32 (__m128i a, __m128i count) |
|
FORCE_INLINE __m128i | _mm_srl_epi64 (__m128i a, __m128i count) |
|
FORCE_INLINE void | _mm_store_pd (double *mem_addr, __m128d a) |
|
FORCE_INLINE void | _mm_store_pd1 (double *mem_addr, __m128d a) |
|
FORCE_INLINE void | _mm_store_sd (double *mem_addr, __m128d a) |
|
FORCE_INLINE void | _mm_store_si128 (__m128i *p, __m128i a) |
|
FORCE_INLINE void | _mm_storeh_pd (double *mem_addr, __m128d a) |
|
FORCE_INLINE void | _mm_storel_epi64 (__m128i *a, __m128i b) |
|
FORCE_INLINE void | _mm_storel_pd (double *mem_addr, __m128d a) |
|
FORCE_INLINE void | _mm_storer_pd (double *mem_addr, __m128d a) |
|
FORCE_INLINE void | _mm_storeu_pd (double *mem_addr, __m128d a) |
|
FORCE_INLINE void | _mm_storeu_si128 (__m128i *p, __m128i a) |
|
FORCE_INLINE void | _mm_storeu_si32 (void *p, __m128i a) |
|
FORCE_INLINE void | _mm_stream_pd (double *p, __m128d a) |
|
FORCE_INLINE void | _mm_stream_si128 (__m128i *p, __m128i a) |
|
FORCE_INLINE void | _mm_stream_si32 (int *p, int a) |
|
FORCE_INLINE void | _mm_stream_si64 (int64_t *p, int64_t a) |
|
FORCE_INLINE __m128i | _mm_sub_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_sub_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_sub_epi64 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_sub_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_sub_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_sub_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m64 | _mm_sub_si64 (__m64 a, __m64 b) |
|
FORCE_INLINE __m128i | _mm_subs_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_subs_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_subs_epu16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_subs_epu8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_undefined_pd (void) |
|
FORCE_INLINE __m128i | _mm_unpackhi_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_unpackhi_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_unpackhi_epi64 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_unpackhi_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_unpackhi_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128i | _mm_unpacklo_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_unpacklo_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_unpacklo_epi64 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_unpacklo_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_unpacklo_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128d | _mm_xor_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128i | _mm_xor_si128 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_addsub_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128 | _mm_addsub_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128d | _mm_hadd_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128 | _mm_hadd_ps (__m128 a, __m128 b) |
|
FORCE_INLINE __m128d | _mm_hsub_pd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128 | _mm_hsub_ps (__m128 _a, __m128 _b) |
|
FORCE_INLINE __m128d | _mm_movedup_pd (__m128d a) |
|
FORCE_INLINE __m128 | _mm_movehdup_ps (__m128 a) |
|
FORCE_INLINE __m128 | _mm_moveldup_ps (__m128 a) |
|
FORCE_INLINE __m128i | _mm_abs_epi16 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_abs_epi32 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_abs_epi8 (__m128i a) |
|
FORCE_INLINE __m64 | _mm_abs_pi16 (__m64 a) |
|
FORCE_INLINE __m64 | _mm_abs_pi32 (__m64 a) |
|
FORCE_INLINE __m64 | _mm_abs_pi8 (__m64 a) |
|
FORCE_INLINE __m128i | _mm_hadd_epi16 (__m128i _a, __m128i _b) |
|
FORCE_INLINE __m128i | _mm_hadd_epi32 (__m128i _a, __m128i _b) |
|
FORCE_INLINE __m64 | _mm_hadd_pi16 (__m64 a, __m64 b) |
|
FORCE_INLINE __m64 | _mm_hadd_pi32 (__m64 a, __m64 b) |
|
FORCE_INLINE __m128i | _mm_hadds_epi16 (__m128i _a, __m128i _b) |
|
FORCE_INLINE __m64 | _mm_hadds_pi16 (__m64 _a, __m64 _b) |
|
FORCE_INLINE __m128i | _mm_hsub_epi16 (__m128i _a, __m128i _b) |
|
FORCE_INLINE __m128i | _mm_hsub_epi32 (__m128i _a, __m128i _b) |
|
FORCE_INLINE __m64 | _mm_hsub_pi16 (__m64 _a, __m64 _b) |
|
FORCE_INLINE __m64 | _mm_hsub_pi32 (__m64 _a, __m64 _b) |
|
FORCE_INLINE __m128i | _mm_hsubs_epi16 (__m128i _a, __m128i _b) |
|
FORCE_INLINE __m64 | _mm_hsubs_pi16 (__m64 _a, __m64 _b) |
|
FORCE_INLINE __m128i | _mm_maddubs_epi16 (__m128i _a, __m128i _b) |
|
FORCE_INLINE __m64 | _mm_maddubs_pi16 (__m64 _a, __m64 _b) |
|
FORCE_INLINE __m128i | _mm_mulhrs_epi16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m64 | _mm_mulhrs_pi16 (__m64 a, __m64 b) |
|
FORCE_INLINE __m128i | _mm_shuffle_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m64 | _mm_shuffle_pi8 (__m64 a, __m64 b) |
|
FORCE_INLINE __m128i | _mm_sign_epi16 (__m128i _a, __m128i _b) |
|
FORCE_INLINE __m128i | _mm_sign_epi32 (__m128i _a, __m128i _b) |
|
FORCE_INLINE __m128i | _mm_sign_epi8 (__m128i _a, __m128i _b) |
|
FORCE_INLINE __m64 | _mm_sign_pi16 (__m64 _a, __m64 _b) |
|
FORCE_INLINE __m64 | _mm_sign_pi32 (__m64 _a, __m64 _b) |
|
FORCE_INLINE __m64 | _mm_sign_pi8 (__m64 _a, __m64 _b) |
|
FORCE_INLINE __m128 | _mm_blend_ps (__m128 _a, __m128 _b, const char imm8) |
|
FORCE_INLINE __m128i | _mm_blendv_epi8 (__m128i _a, __m128i _b, __m128i _mask) |
|
FORCE_INLINE __m128d | _mm_blendv_pd (__m128d _a, __m128d _b, __m128d _mask) |
|
FORCE_INLINE __m128 | _mm_blendv_ps (__m128 _a, __m128 _b, __m128 _mask) |
|
FORCE_INLINE __m128d | _mm_ceil_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128 | _mm_ceil_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128i | _mm_cmpeq_epi64 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_cvtepi16_epi32 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtepi16_epi64 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtepi32_epi64 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtepi8_epi16 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtepi8_epi32 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtepi8_epi64 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtepu16_epi32 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtepu16_epi64 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtepu32_epi64 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtepu8_epi16 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtepu8_epi32 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_cvtepu8_epi64 (__m128i a) |
|
FORCE_INLINE __m128d | _mm_dp_pd (__m128d a, __m128d b, const int imm) |
|
FORCE_INLINE __m128 | _mm_dp_ps (__m128 a, __m128 b, const int imm) |
|
FORCE_INLINE __m128d | _mm_floor_sd (__m128d a, __m128d b) |
|
FORCE_INLINE __m128 | _mm_floor_ss (__m128 a, __m128 b) |
|
FORCE_INLINE __m128i | _mm_max_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_max_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_max_epu16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_max_epu32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_min_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_min_epi8 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_min_epu16 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_min_epu32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_minpos_epu16 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_mpsadbw_epu8 (__m128i a, __m128i b, const int imm) |
|
FORCE_INLINE __m128i | _mm_mul_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_mullo_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128i | _mm_packus_epi32 (__m128i a, __m128i b) |
|
FORCE_INLINE __m128d | _mm_round_sd (__m128d a, __m128d b, int rounding) |
|
FORCE_INLINE __m128 | _mm_round_ss (__m128 a, __m128 b, int rounding) |
|
FORCE_INLINE __m128i | _mm_stream_load_si128 (__m128i *p) |
|
FORCE_INLINE int | _mm_test_all_ones (__m128i a) |
|
FORCE_INLINE int | _mm_test_all_zeros (__m128i a, __m128i mask) |
|
FORCE_INLINE int | _mm_test_mix_ones_zeros (__m128i a, __m128i mask) |
|
FORCE_INLINE int | _mm_testc_si128 (__m128i a, __m128i b) |
|
FORCE_INLINE int | _mm_testz_si128 (__m128i a, __m128i b) |
|
static uint16_t | _sse2neon_aggregate_equal_any_8x16 (int la, int lb, __m128i mtx[16]) |
|
static uint16_t | _sse2neon_aggregate_equal_any_16x8 (int la, int lb, __m128i mtx[16]) |
|
static uint16_t | _sse2neon_cmp_byte_equal_any (__m128i a, int la, __m128i b, int lb) |
|
static uint16_t | _sse2neon_cmp_word_equal_any (__m128i a, int la, __m128i b, int lb) |
|
ssssssssssssstatic uint16_t | _sse2neon_aggregate_ranges_16x8 (int la, int lb, __m128i mtx[16]) |
|
static uint16_t | _sse2neon_aggregate_ranges_8x16 (int la, int lb, __m128i mtx[16]) |
|
static uint16_t | _sse2neon_cmp_ubyte_ranges (__m128i a, int la, __m128i b, int lb) |
|
static uint16_t | _sse2neon_cmp_sbyte_ranges (__m128i a, int la, __m128i b, int lb) |
|
static uint16_t | _sse2neon_cmp_uword_ranges (__m128i a, int la, __m128i b, int lb) |
|
static uint16_t | _sse2neon_cmp_sword_ranges (__m128i a, int la, __m128i b, int lb) |
|
static uint16_t | _sse2neon_cmp_byte_equal_each (__m128i a, int la, __m128i b, int lb) |
|
static uint16_t | _sse2neon_cmp_word_equal_each (__m128i a, int la, __m128i b, int lb) |
|
static uint16_t | _sse2neon_aggregate_equal_ordered_8x16 (int bound, int la, int lb, __m128i mtx[16]) |
|
static uint16_t | _sse2neon_aggregate_equal_ordered_16x8 (int bound, int la, int lb, __m128i mtx[16]) |
|
static uint16_t | _sse2neon_cmp_byte_equal_ordered (__m128i a, int la, __m128i b, int lb) |
|
static uint16_t | _sse2neon_cmp_word_equal_ordered (__m128i a, int la, __m128i b, int lb) |
|
| _ (CMP_UBYTE_EQUAL_ANY, cmp_byte_equal_any) _(CMP_UWORD_EQUAL_ANY |
|
cmp_word_equal_any | _ (CMP_SBYTE_EQUAL_ANY, cmp_byte_equal_any) _(CMP_SWORD_EQUAL_ANY |
|
cmp_word_equal_any cmp_word_equal_any | _ (CMP_UBYTE_RANGES, cmp_ubyte_ranges) _(CMP_UWORD_RANGES |
|
cmp_word_equal_any cmp_word_equal_any cmp_uword_ranges | _ (CMP_SBYTE_RANGES, cmp_sbyte_ranges) _(CMP_SWORD_RANGES |
|
cmp_word_equal_any cmp_word_equal_any cmp_uword_ranges cmp_sword_ranges | _ (CMP_UBYTE_EQUAL_EACH, cmp_byte_equal_each) _(CMP_UWORD_EQUAL_EACH |
|
cmp_word_equal_any cmp_word_equal_any cmp_uword_ranges cmp_sword_ranges cmp_word_equal_each | _ (CMP_SBYTE_EQUAL_EACH, cmp_byte_equal_each) _(CMP_SWORD_EQUAL_EACH |
|
cmp_word_equal_any cmp_word_equal_any cmp_uword_ranges cmp_sword_ranges cmp_word_equal_each cmp_word_equal_each | _ (CMP_UBYTE_EQUAL_ORDERED, cmp_byte_equal_ordered) _(CMP_UWORD_EQUAL_ORDERED |
|
cmp_word_equal_any cmp_word_equal_any cmp_uword_ranges cmp_sword_ranges cmp_word_equal_each cmp_word_equal_each cmp_word_equal_ordered | _ (CMP_SBYTE_EQUAL_ORDERED, cmp_byte_equal_ordered) _(CMP_SWORD_EQUAL_ORDERED |
|
FORCE_INLINE uint16_t | _sse2neon_sido_negative (int res, int lb, int imm8, int bound) |
|
FORCE_INLINE int | _sse2neon_clz (unsigned int x) |
|
FORCE_INLINE int | _sse2neon_ctz (unsigned int x) |
|
FORCE_INLINE int | _sse2neon_ctzll (unsigned long long x) |
|
FORCE_INLINE int | _mm_cmpestra (__m128i a, int la, __m128i b, int lb, const int imm8) |
|
FORCE_INLINE int | _mm_cmpestrc (__m128i a, int la, __m128i b, int lb, const int imm8) |
|
FORCE_INLINE int | _mm_cmpestri (__m128i a, int la, __m128i b, int lb, const int imm8) |
|
FORCE_INLINE __m128i | _mm_cmpestrm (__m128i a, int la, __m128i b, int lb, const int imm8) |
|
FORCE_INLINE int | _mm_cmpestro (__m128i a, int la, __m128i b, int lb, const int imm8) |
|
FORCE_INLINE int | _mm_cmpestrs (__m128i a, int la, __m128i b, int lb, const int imm8) |
|
FORCE_INLINE int | _mm_cmpestrz (__m128i a, int la, __m128i b, int lb, const int imm8) |
|
FORCE_INLINE int | _mm_cmpistra (__m128i a, __m128i b, const int imm8) |
|
FORCE_INLINE int | _mm_cmpistrc (__m128i a, __m128i b, const int imm8) |
|
FORCE_INLINE int | _mm_cmpistri (__m128i a, __m128i b, const int imm8) |
|
FORCE_INLINE __m128i | _mm_cmpistrm (__m128i a, __m128i b, const int imm8) |
|
FORCE_INLINE int | _mm_cmpistro (__m128i a, __m128i b, const int imm8) |
|
FORCE_INLINE int | _mm_cmpistrs (__m128i a, __m128i b, const int imm8) |
|
FORCE_INLINE int | _mm_cmpistrz (__m128i a, __m128i b, const int imm8) |
|
FORCE_INLINE __m128i | _mm_cmpgt_epi64 (__m128i a, __m128i b) |
|
FORCE_INLINE uint32_t | _mm_crc32_u16 (uint32_t crc, uint16_t v) |
|
FORCE_INLINE uint32_t | _mm_crc32_u32 (uint32_t crc, uint32_t v) |
|
FORCE_INLINE uint64_t | _mm_crc32_u64 (uint64_t crc, uint64_t v) |
|
FORCE_INLINE __m128i | _mm_aesenc_si128 (__m128i a, __m128i RoundKey) |
|
FORCE_INLINE __m128i | _mm_aesdec_si128 (__m128i a, __m128i RoundKey) |
|
FORCE_INLINE __m128i | _mm_aesenclast_si128 (__m128i a, __m128i RoundKey) |
|
FORCE_INLINE __m128i | _mm_aesdeclast_si128 (__m128i a, __m128i RoundKey) |
|
FORCE_INLINE __m128i | _mm_aesimc_si128 (__m128i a) |
|
FORCE_INLINE __m128i | _mm_aeskeygenassist_si128 (__m128i a, const int rcon) |
|
FORCE_INLINE __m128i | _mm_clmulepi64_si128 (__m128i _a, __m128i _b, const int imm) |
|
FORCE_INLINE unsigned int | _sse2neon_mm_get_denormals_zero_mode (void) |
|
FORCE_INLINE int | _mm_popcnt_u32 (unsigned int a) |
|
FORCE_INLINE int64_t | _mm_popcnt_u64 (uint64_t a) |
|
FORCE_INLINE void | _sse2neon_mm_set_denormals_zero_mode (unsigned int flag) |
|
FORCE_INLINE uint64_t | _rdtsc (void) |
|