src/nfa/limex_internal.h
src/nfa/limex_runtime.h
src/nfa/limex_runtime_impl.h
+ src/nfa/limex_shuffle.h
src/nfa/limex_state_impl.h
src/nfa/mpv.h
src/nfa/mpv.c
src/util/pqueue.h
src/util/scatter.h
src/util/scatter_runtime.h
- src/util/shuffle.h
- src/util/shuffle_ssse3.h
src/util/simd_utils.h
- src/util/simd_utils_ssse3.h
- src/util/simd_utils_ssse3.c
+ src/util/simd_utils.c
src/util/state_compress.h
src/util/state_compress.c
src/util/unaligned.h
src/util/report_manager.cpp
src/util/report_manager.h
src/util/simd_utils.h
- src/util/simd_utils_ssse3.h
src/util/target_info.cpp
src/util/target_info.h
src/util/ue2_containers.h
#include "teddy.h"
#include "teddy_internal.h"
#include "util/simd_utils.h"
-#include "util/simd_utils_ssse3.h"
/** \brief number of bytes processed in each iteration */
#define ITER_BYTES 16
#include "teddy_internal.h"
#include "teddy_runtime_common.h"
#include "util/simd_utils.h"
-#include "util/simd_utils_ssse3.h"
const u8 ALIGN_DIRECTIVE p_mask_arr[17][32] = {
{0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00,
#include "teddy_internal.h"
#include "teddy_runtime_common.h"
#include "util/simd_utils.h"
-#include "util/simd_utils_ssse3.h"
#if defined(__AVX2__)
#include "util/compare.h"
#include "util/masked_move.h"
#include "util/simd_utils.h"
-#include "util/simd_utils_ssse3.h"
#include <ctype.h>
#include <stdbool.h>
#include "accel.h"
#include "limex_internal.h"
#include "limex_limits.h"
+#include "limex_shuffle.h"
#include "nfa_internal.h"
#include "shufti.h"
#include "truffle.h"
#include "ue2common.h"
#include "vermicelli.h"
#include "util/bitutils.h"
-#include "util/shuffle.h"
#include "util/simd_utils.h"
-#include "util/simd_utils_ssse3.h"
-#include "util/shuffle_ssse3.h"
static really_inline
size_t accelScanWrapper(const u8 *accelTable, const union AccelAux *aux,
size_t doAccel32(u32 s, u32 accel, const u8 *accelTable,
const union AccelAux *aux, const u8 *input, size_t i,
size_t end) {
- u32 idx = shuffleDynamic32(s, accel);
+ u32 idx = packedExtract32(s, accel);
return accelScanWrapper(accelTable, aux, input, idx, i, end);
}
DEBUG_PRINTF("using PSHUFB for 128-bit shuffle\n");
m128 accelPerm = limex->accelPermute;
m128 accelComp = limex->accelCompare;
- idx = shufflePshufb128(s, accelPerm, accelComp);
+ idx = packedExtract128(s, accelPerm, accelComp);
return accelScanWrapper(accelTable, aux, input, idx, i, end);
}
m256 accelPerm = limex->accelPermute;
m256 accelComp = limex->accelCompare;
#if !defined(__AVX2__)
- u32 idx1 = shufflePshufb128(s.lo, accelPerm.lo, accelComp.lo);
- u32 idx2 = shufflePshufb128(s.hi, accelPerm.hi, accelComp.hi);
-#else
- // TODO: learn you some avx2 shuffles for great good
- u32 idx1 = shufflePshufb128(movdq_lo(s), movdq_lo(accelPerm),
- movdq_lo(accelComp));
- u32 idx2 = shufflePshufb128(movdq_hi(s), movdq_hi(accelPerm),
- movdq_hi(accelComp));
-#endif
+ u32 idx1 = packedExtract128(s.lo, accelPerm.lo, accelComp.lo);
+ u32 idx2 = packedExtract128(s.hi, accelPerm.hi, accelComp.hi);
assert((idx1 & idx2) == 0); // should be no shared bits
idx = idx1 | idx2;
+#else
+ idx = packedExtract256(s, accelPerm, accelComp);
+#endif
return accelScanWrapper(accelTable, aux, input, idx, i, end);
}
DEBUG_PRINTF("using PSHUFB for 384-bit shuffle\n");
m384 accelPerm = limex->accelPermute;
m384 accelComp = limex->accelCompare;
- u32 idx1 = shufflePshufb128(s.lo, accelPerm.lo, accelComp.lo);
- u32 idx2 = shufflePshufb128(s.mid, accelPerm.mid, accelComp.mid);
- u32 idx3 = shufflePshufb128(s.hi, accelPerm.hi, accelComp.hi);
+ u32 idx1 = packedExtract128(s.lo, accelPerm.lo, accelComp.lo);
+ u32 idx2 = packedExtract128(s.mid, accelPerm.mid, accelComp.mid);
+ u32 idx3 = packedExtract128(s.hi, accelPerm.hi, accelComp.hi);
assert((idx1 & idx2 & idx3) == 0); // should be no shared bits
idx = idx1 | idx2 | idx3;
return accelScanWrapper(accelTable, aux, input, idx, i, end);
m512 accelPerm = limex->accelPermute;
m512 accelComp = limex->accelCompare;
#if !defined(__AVX2__)
- u32 idx1 = shufflePshufb128(s.lo.lo, accelPerm.lo.lo, accelComp.lo.lo);
- u32 idx2 = shufflePshufb128(s.lo.hi, accelPerm.lo.hi, accelComp.lo.hi);
- u32 idx3 = shufflePshufb128(s.hi.lo, accelPerm.hi.lo, accelComp.hi.lo);
- u32 idx4 = shufflePshufb128(s.hi.hi, accelPerm.hi.hi, accelComp.hi.hi);
-#else
- u32 idx1 = shufflePshufb128(movdq_lo(s.lo), movdq_lo(accelPerm.lo),
- movdq_lo(accelComp.lo));
- u32 idx2 = shufflePshufb128(movdq_hi(s.lo), movdq_hi(accelPerm.lo),
- movdq_hi(accelComp.lo));
- u32 idx3 = shufflePshufb128(movdq_lo(s.hi), movdq_lo(accelPerm.hi),
- movdq_lo(accelComp.hi));
- u32 idx4 = shufflePshufb128(movdq_hi(s.hi), movdq_hi(accelPerm.hi),
- movdq_hi(accelComp.hi));
-#endif
+ u32 idx1 = packedExtract128(s.lo.lo, accelPerm.lo.lo, accelComp.lo.lo);
+ u32 idx2 = packedExtract128(s.lo.hi, accelPerm.lo.hi, accelComp.lo.hi);
+ u32 idx3 = packedExtract128(s.hi.lo, accelPerm.hi.lo, accelComp.hi.lo);
+ u32 idx4 = packedExtract128(s.hi.hi, accelPerm.hi.hi, accelComp.hi.hi);
assert((idx1 & idx2 & idx3 & idx4) == 0); // should be no shared bits
idx = idx1 | idx2 | idx3 | idx4;
+#else
+ u32 idx1 = packedExtract256(s.lo, accelPerm.lo, accelComp.lo);
+ u32 idx2 = packedExtract256(s.hi, accelPerm.hi, accelComp.hi);
+ assert((idx1 & idx2) == 0); // should be no shared bits
+ idx = idx1 | idx2;
+#endif
return accelScanWrapper(accelTable, aux, input, idx, i, end);
}
/*
- * Copyright (c) 2015, Intel Corporation
+ * Copyright (c) 2015-2016, Intel Corporation
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are met:
* be faster and actually correct if these assumptions don't hold true.
*/
-#ifndef SHUFFLE_H
-#define SHUFFLE_H
+#ifndef LIMEX_SHUFFLE_H
+#define LIMEX_SHUFFLE_H
-#include "config.h"
-#include "bitutils.h"
-#include "simd_utils.h"
#include "ue2common.h"
+#include "util/bitutils.h"
+#include "util/simd_utils.h"
#if defined(__BMI2__) || (defined(_WIN32) && defined(__AVX2__))
#define HAVE_PEXT
#endif
static really_inline
-u32 shuffleDynamic32(u32 x, u32 mask) {
+u32 packedExtract32(u32 x, u32 mask) {
#if defined(HAVE_PEXT)
// Intel BMI2 can do this operation in one instruction.
return _pext_u32(x, mask);
}
static really_inline
-u32 shuffleDynamic64(u64a x, u64a mask) {
+u32 packedExtract64(u64a x, u64a mask) {
#if defined(HAVE_PEXT) && defined(ARCH_64_BIT)
// Intel BMI2 can do this operation in one instruction.
return _pext_u64(x, mask);
#undef HAVE_PEXT
-#endif // SHUFFLE_H
+static really_inline
+u32 packedExtract128(m128 s, const m128 permute, const m128 compare) {
+ m128 shuffled = pshufb(s, permute);
+ m128 compared = and128(shuffled, compare);
+ u16 rv = ~cmpmsk8(compared, shuffled);
+ return (u32)rv;
+}
+
+#if defined(__AVX2__)
+static really_inline
+u32 packedExtract256(m256 s, const m256 permute, const m256 compare) {
+ // vpshufb doesn't cross lanes, so this is a bit of a cheat
+ m256 shuffled = vpshufb(s, permute);
+ m256 compared = and256(shuffled, compare);
+ u32 rv = ~movemask256(eq256(compared, shuffled));
+ // stitch the lane-wise results back together
+ return (u32)((rv >> 16) | (rv & 0xffffU));
+}
+#endif // AVX2
+
+#endif // LIMEX_SHUFFLE_H
/*
- * Copyright (c) 2015, Intel Corporation
+ * Copyright (c) 2015-2016, Intel Corporation
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are met:
#include "ue2common.h"
#include "util/bitutils.h"
#include "util/simd_utils.h"
-#include "util/simd_utils_ssse3.h"
static really_inline
const u8 *JOIN(MATCH_ALGO, fwdBlock)(m256 mask_lo, m256 mask_hi, m256 chars,
/*
- * Copyright (c) 2015, Intel Corporation
+ * Copyright (c) 2015-2016, Intel Corporation
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are met:
#include "ue2common.h"
#include "util/bitutils.h"
#include "util/simd_utils.h"
-#include "util/simd_utils_ssse3.h"
/* Normal SSSE3 shufti */
/*
- * Copyright (c) 2015, Intel Corporation
+ * Copyright (c) 2015-2016, Intel Corporation
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are met:
#include "multitruffle.h"
#include "util/bitutils.h"
#include "util/simd_utils.h"
-#include "util/simd_utils_ssse3.h"
#include "multiaccel_common.h"
/*
- * Copyright (c) 2015, Intel Corporation
+ * Copyright (c) 2015-2016, Intel Corporation
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are met:
#include "shufti_common.h"
-#include "util/simd_utils_ssse3.h"
/** \brief Naive byte-by-byte implementation. */
static really_inline
/*
- * Copyright (c) 2015, Intel Corporation
+ * Copyright (c) 2015-2016, Intel Corporation
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are met:
#include "util/bitutils.h"
#include "util/simd_utils.h"
#include "util/unaligned.h"
-#include "util/simd_utils_ssse3.h"
/*
* Common stuff for all versions of shufti (single, multi and multidouble)
/*
- * Copyright (c) 2015, Intel Corporation
+ * Copyright (c) 2015-2016, Intel Corporation
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are met:
#include "truffle.h"
#include "util/bitutils.h"
#include "util/simd_utils.h"
-#include "util/simd_utils_ssse3.h"
#include "truffle_common.h"
/*
- * Copyright (c) 2015, Intel Corporation
+ * Copyright (c) 2015-2016, Intel Corporation
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are met:
#include "util/bitutils.h"
#include "util/simd_utils.h"
-#include "util/simd_utils_ssse3.h"
/*
* Common stuff for all versions of truffle (single, multi and multidouble)
/*
- * Copyright (c) 2015, Intel Corporation
+ * Copyright (c) 2015-2016, Intel Corporation
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are met:
#include "rose_internal.h"
#include "nfa/nfa_api_queue.h"
#include "util/simd_utils.h"
-#include "util/simd_utils_ssse3.h"
/** \brief Maximum number of bytes to scan when looking for a "counting miracle"
* stop character. */
/*
- * Copyright (c) 2015, Intel Corporation
+ * Copyright (c) 2015-2016, Intel Corporation
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are met:
#include "unaligned.h"
#include "simd_utils.h"
-#include "simd_utils_ssse3.h"
#ifdef __cplusplus
extern "C" {
+++ /dev/null
-/*
- * Copyright (c) 2015, Intel Corporation
- *
- * Redistribution and use in source and binary forms, with or without
- * modification, are permitted provided that the following conditions are met:
- *
- * * Redistributions of source code must retain the above copyright notice,
- * this list of conditions and the following disclaimer.
- * * Redistributions in binary form must reproduce the above copyright
- * notice, this list of conditions and the following disclaimer in the
- * documentation and/or other materials provided with the distribution.
- * * Neither the name of Intel Corporation nor the names of its contributors
- * may be used to endorse or promote products derived from this software
- * without specific prior written permission.
- *
- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
- * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
- * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
- * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
- * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
- * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
- * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
- * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
- * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
- * POSSIBILITY OF SUCH DAMAGE.
- */
-
-#ifndef SHUFFLE_SSSE3_H
-#define SHUFFLE_SSSE3_H
-
-#include "simd_utils_ssse3.h"
-
-#ifdef DEBUG
-#include "compare.h"
-static really_inline void shufDumpMsk(m128 msk) {
- u8 * mskAsU8 = (u8 *)&msk;
- for (int i = 0; i < 16; i++) {
- u8 c = mskAsU8[i];
- for (int j = 0; j < 8; j++) {
- if ((c >> (7-j)) & 0x1)
- printf("1");
- else
- printf("0");
- }
- printf(" ");
- }
-}
-
-static really_inline void shufDumpMskAsChars(m128 msk) {
- u8 * mskAsU8 = (u8 *)&msk;
- for (int i = 0; i < 16; i++) {
- u8 c = mskAsU8[i];
- if (ourisprint(c))
- printf("%c",c);
- else
- printf(".");
- }
-}
-#endif
-
-#if !defined(NO_SSSE3)
-static really_inline
-u32 shufflePshufb128(m128 s, const m128 permute, const m128 compare) {
- m128 shuffled = pshufb(s, permute);
- m128 compared = and128(shuffled, compare);
-#ifdef DEBUG
- printf("State: "); shufDumpMsk(s); printf("\n");
- printf("Permute: "); shufDumpMsk(permute); printf("\n");
- printf("Compare: "); shufDumpMsk(compare); printf("\n");
- printf("Shuffled: "); shufDumpMsk(shuffled); printf("\n");
- printf("Compared: "); shufDumpMsk(compared); printf("\n");
-#endif
- u16 rv = ~cmpmsk8(compared, shuffled);
- return (u32)rv;
-}
-#endif // NO_SSSE3
-
-#endif // SHUFFLE_SSSE3_H
* POSSIBILITY OF SUCH DAMAGE.
*/
-#include "simd_utils_ssse3.h"
+#include "simd_utils.h"
const char vbs_mask_data[] ALIGN_CL_DIRECTIVE = {
0xf0, 0xf0, 0xf0, 0xf0, 0xf0, 0xf0, 0xf0, 0xf0,
#ifndef SIMD_UTILS
#define SIMD_UTILS
+#if !defined(_WIN32) && !defined(__SSSE3__)
+#error SSSE3 instructions must be enabled
+#endif
+
#include "config.h"
#include <string.h> // for memcpy
#define assume_aligned(x, y) (x)
#endif
+#ifdef __cplusplus
+extern "C" {
+#endif
+extern const char vbs_mask_data[];
+#ifdef __cplusplus
+}
+#endif
+
static really_inline m128 ones128(void) {
#if !defined(NO_ASM)
// trick from Intel's optimization guide to generate all-ones. We have to
#define eq128(a, b) _mm_cmpeq_epi8((a), (b))
#define movemask128(a) ((u32)_mm_movemask_epi8((a)))
-
// We found that this generated better code with gcc-4.1 and with the default
// tuning settings on gcc-4.4 than just using the _mm_set1_epi8() instrinsic.
static really_inline m128 set16x8(u8 c) {
return !!(bytes[n / 8] & (1 << (n % 8)));
}
+// offset must be an immediate
+#define palignr(r, l, offset) _mm_alignr_epi8(r, l, offset)
+
+static really_inline
+m128 pshufb(m128 a, m128 b) {
+ m128 result;
+ result = _mm_shuffle_epi8(a, b);
+ return result;
+}
+
+static really_inline
+m256 vpshufb(m256 a, m256 b) {
+#if defined(__AVX2__)
+ return _mm256_shuffle_epi8(a, b);
+#else
+ m256 rv;
+ rv.lo = pshufb(a.lo, b.lo);
+ rv.hi = pshufb(a.hi, b.hi);
+ return rv;
+#endif
+}
+
+static really_inline
+m128 variable_byte_shift_m128(m128 in, s32 amount) {
+ assert(amount >= -16 && amount <= 16);
+ m128 shift_mask = loadu128(vbs_mask_data + 16 - amount);
+ return pshufb(in, shift_mask);
+}
+
+
/****
**** 256-bit Primitives
****/
#define extractlow32from256(a) movd(cast256to128(a))
#define interleave256hi(a, b) _mm256_unpackhi_epi8(a, b);
#define interleave256lo(a, b) _mm256_unpacklo_epi8(a, b);
+#define vpalignr(r, l, offset) _mm256_alignr_epi8(r, l, offset)
#endif //AVX2
+++ /dev/null
-/*
- * Copyright (c) 2015-2016, Intel Corporation
- *
- * Redistribution and use in source and binary forms, with or without
- * modification, are permitted provided that the following conditions are met:
- *
- * * Redistributions of source code must retain the above copyright notice,
- * this list of conditions and the following disclaimer.
- * * Redistributions in binary form must reproduce the above copyright
- * notice, this list of conditions and the following disclaimer in the
- * documentation and/or other materials provided with the distribution.
- * * Neither the name of Intel Corporation nor the names of its contributors
- * may be used to endorse or promote products derived from this software
- * without specific prior written permission.
- *
- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
- * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
- * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
- * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
- * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
- * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
- * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
- * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
- * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
- * POSSIBILITY OF SUCH DAMAGE.
- */
-
-/** \file
- * \brief SIMD primitives specifically for Intel SSSE3 platforms.
- */
-
-#ifndef SIMD_UTILS_SSSE3_H_E27DF795C9AA02
-#define SIMD_UTILS_SSSE3_H_E27DF795C9AA02
-
-#if !defined(_WIN32) && !defined(__SSSE3__)
-#error SSSE3 instructions must be enabled
-#endif
-
-#include "simd_utils.h"
-#include "ue2common.h"
-
-// we may already have x86intrin.h
-#if !defined(USE_X86INTRIN_H)
-#if defined(HAVE_C_INTRIN_H)
-#include <intrin.h>
-#elif defined(HAVE_TMMINTRIN_H)
-#include <tmmintrin.h> // SSSE3 intrinsics
-#else
-#define I_HAVE_BROKEN_INTRINSICS
-#endif
-#endif
-
-
-#if !defined(I_HAVE_BROKEN_INTRINSICS)
-// newish compilers get this right
-#define palignr(r, l, offset) _mm_alignr_epi8(r, l, offset)
-#else
-// must be inline, even in weak-sauce debug builds.
-// oldish compilers either don't have the intrinsic, or force one arg through memory
-static really_really_inline
-m128 palignr(m128 r, m128 l, const int offset) {
- __asm__ ("palignr %2,%1,%0" : "+x"(r) : "x"(l), "i"(offset));
- return r;
-}
-#endif
-
-
-static really_inline
-m128 pshufb(m128 a, m128 b) {
- m128 result;
-#if !defined(I_HAVE_BROKEN_INTRINSICS)
- result = _mm_shuffle_epi8(a, b);
-#else
- __asm__("pshufb\t%1,%0" : "=x"(result) : "xm"(b), "0"(a));
-#endif
- return result;
-}
-
-#ifdef __cplusplus
-extern "C" {
-#endif
-extern const char vbs_mask_data[];
-#ifdef __cplusplus
-}
-#endif
-
-static really_inline
-m128 variable_byte_shift_m128(m128 in, s32 amount) {
- assert(amount >= -16 && amount <= 16);
- m128 shift_mask = loadu128(vbs_mask_data + 16 - amount);
- return pshufb(in, shift_mask);
-}
-
-#if defined(__AVX2__)
-
-static really_inline
-m256 vpshufb(m256 a, m256 b) {
- return _mm256_shuffle_epi8(a, b);
-}
-
-#if defined(USE_GCC_COMPOUND_STATEMENTS)
-#define vpalignr(r, l, offset) ({ \
- m256 res = _mm256_alignr_epi8(r, l, offset); \
- res; \
-})
-#else
-#define vpalignr(r, l, offset) _mm256_alignr_epi8(r, l, offset)
-#endif
-
-#else // not __AVX2__
-
-static really_inline
-m256 vpshufb(m256 a, m256 b) {
- m256 rv;
- rv.lo = pshufb(a.lo, b.lo);
- rv.hi = pshufb(a.hi, b.hi);
- return rv;
-}
-
-/* palignr requires the offset to be an immediate, which we can do with a
- * compound macro, otherwise we have to enumerate the offsets and hope the
- * compiler can throw the rest away. */
-#if defined(USE_GCC_COMPOUND_STATEMENTS)
-#define vpalignr(r, l, offset) ({ \
- m256 res; \
- res.lo = palignr(r.lo, l.lo, offset); \
- res.hi = palignr(r.hi, l.hi, offset); \
- res; \
-})
-#else
-#define VPALIGN_CASE(N) case N: \
- res.lo = palignr(r.lo, l.lo, N); \
- res.hi = palignr(r.hi, l.hi, N); \
- return res;
-static really_inline
-m256 vpalignr(m256 r, m256 l, const int offset) {
- m256 res;
- switch (offset) {
- VPALIGN_CASE(0)
- VPALIGN_CASE(1)
- VPALIGN_CASE(2)
- VPALIGN_CASE(3)
- VPALIGN_CASE(4)
- VPALIGN_CASE(5)
- VPALIGN_CASE(6)
- VPALIGN_CASE(7)
- VPALIGN_CASE(8)
- VPALIGN_CASE(9)
- VPALIGN_CASE(10)
- VPALIGN_CASE(11)
- VPALIGN_CASE(12)
- VPALIGN_CASE(13)
- VPALIGN_CASE(14)
- VPALIGN_CASE(15)
- default:
- assert(0);
- return zeroes256();
- }
-}
-#undef VPALIGN_CASE
-#endif
-#endif // __AVX2__
-
-#endif /* SIMD_UTILS_SSSE3_H_E27DF795C9AA02 */
-
/*
- * Copyright (c) 2015, Intel Corporation
+ * Copyright (c) 2015-2016, Intel Corporation
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are met:
#include "gtest/gtest.h"
#include "util/simd_utils.h"
-#include "util/shuffle.h"
-#include "util/shuffle_ssse3.h"
+#include "nfa/limex_shuffle.h"
namespace {
return cf.simd;
}
-TEST(Shuffle, ShuffleDynamic32_1) {
+TEST(Shuffle, PackedExtract32_1) {
// Try all possible one-bit masks
for (unsigned int i = 0; i < 32; i++) {
// shuffle a single 1 bit to the front
u32 mask = 1U << i;
- EXPECT_EQ(1U, shuffleDynamic32(mask, mask));
- EXPECT_EQ(1U, shuffleDynamic32(~0U, mask));
+ EXPECT_EQ(1U, packedExtract32(mask, mask));
+ EXPECT_EQ(1U, packedExtract32(~0U, mask));
// we should get zero out of these cases
- EXPECT_EQ(0U, shuffleDynamic32(0, mask));
- EXPECT_EQ(0U, shuffleDynamic32(~mask, mask));
+ EXPECT_EQ(0U, packedExtract32(0, mask));
+ EXPECT_EQ(0U, packedExtract32(~mask, mask));
// we should get zero out of all the other bit positions
for (unsigned int j = 0; (j != i && j < 32); j++) {
- EXPECT_EQ(0U, shuffleDynamic32((1U << j), mask));
+ EXPECT_EQ(0U, packedExtract32((1U << j), mask));
}
}
}
-TEST(Shuffle, ShuffleDynamic32_2) {
+TEST(Shuffle, PackedExtract32_2) {
// All 32 bits in mask are on
u32 mask = ~0U;
- EXPECT_EQ(0U, shuffleDynamic32(0, mask));
- EXPECT_EQ(mask, shuffleDynamic32(mask, mask));
+ EXPECT_EQ(0U, packedExtract32(0, mask));
+ EXPECT_EQ(mask, packedExtract32(mask, mask));
for (unsigned int i = 0; i < 32; i++) {
- EXPECT_EQ(1U << i, shuffleDynamic32(1U << i, mask));
+ EXPECT_EQ(1U << i, packedExtract32(1U << i, mask));
}
}
-TEST(Shuffle, ShuffleDynamic32_3) {
+TEST(Shuffle, PackedExtract32_3) {
// Try setting every second bit
u32 mask = 0;
for (unsigned int i = 0; i < 32; i += 2) {
}
// Test both cases (all even bits, all odd bits)
- EXPECT_EQ((1U << 16) - 1, shuffleDynamic32(mask, mask));
- EXPECT_EQ((1U << 16) - 1, shuffleDynamic32(~mask, ~mask));
- EXPECT_EQ(0U, shuffleDynamic32(~mask, mask));
- EXPECT_EQ(0U, shuffleDynamic32(mask, ~mask));
+ EXPECT_EQ((1U << 16) - 1, packedExtract32(mask, mask));
+ EXPECT_EQ((1U << 16) - 1, packedExtract32(~mask, ~mask));
+ EXPECT_EQ(0U, packedExtract32(~mask, mask));
+ EXPECT_EQ(0U, packedExtract32(mask, ~mask));
for (unsigned int i = 0; i < 32; i += 2) {
- EXPECT_EQ(1U << (i/2), shuffleDynamic32(1U << i, mask));
- EXPECT_EQ(0U, shuffleDynamic32(1U << i, ~mask));
- EXPECT_EQ(1U << (i/2), shuffleDynamic32(1U << (i+1), ~mask));
- EXPECT_EQ(0U, shuffleDynamic32(1U << (i+1), mask));
+ EXPECT_EQ(1U << (i/2), packedExtract32(1U << i, mask));
+ EXPECT_EQ(0U, packedExtract32(1U << i, ~mask));
+ EXPECT_EQ(1U << (i/2), packedExtract32(1U << (i+1), ~mask));
+ EXPECT_EQ(0U, packedExtract32(1U << (i+1), mask));
}
}
-TEST(Shuffle, ShuffleDynamic64_1) {
+TEST(Shuffle, PackedExtract64_1) {
// Try all possible one-bit masks
for (unsigned int i = 0; i < 64; i++) {
// shuffle a single 1 bit to the front
u64a mask = 1ULL << i;
- EXPECT_EQ(1U, shuffleDynamic64(mask, mask));
- EXPECT_EQ(1U, shuffleDynamic64(~0ULL, mask));
+ EXPECT_EQ(1U, packedExtract64(mask, mask));
+ EXPECT_EQ(1U, packedExtract64(~0ULL, mask));
// we should get zero out of these cases
- EXPECT_EQ(0U, shuffleDynamic64(0, mask));
- EXPECT_EQ(0U, shuffleDynamic64(~mask, mask));
+ EXPECT_EQ(0U, packedExtract64(0, mask));
+ EXPECT_EQ(0U, packedExtract64(~mask, mask));
// we should get zero out of all the other bit positions
for (unsigned int j = 0; (j != i && j < 64); j++) {
- EXPECT_EQ(0U, shuffleDynamic64((1ULL << j), mask));
+ EXPECT_EQ(0U, packedExtract64((1ULL << j), mask));
}
}
}
-TEST(Shuffle, ShuffleDynamic64_2) {
+TEST(Shuffle, PackedExtract64_2) {
// Fill first half of mask
u64a mask = 0x00000000ffffffffULL;
- EXPECT_EQ(0U, shuffleDynamic64(0, mask));
- EXPECT_EQ(0xffffffffU, shuffleDynamic64(mask, mask));
+ EXPECT_EQ(0U, packedExtract64(0, mask));
+ EXPECT_EQ(0xffffffffU, packedExtract64(mask, mask));
for (unsigned int i = 0; i < 32; i++) {
- EXPECT_EQ(1U << i, shuffleDynamic64(1ULL << i, mask));
+ EXPECT_EQ(1U << i, packedExtract64(1ULL << i, mask));
}
// Fill second half of mask
mask = 0xffffffff00000000ULL;
- EXPECT_EQ(0U, shuffleDynamic64(0, mask));
- EXPECT_EQ(0xffffffffU, shuffleDynamic64(mask, mask));
+ EXPECT_EQ(0U, packedExtract64(0, mask));
+ EXPECT_EQ(0xffffffffU, packedExtract64(mask, mask));
for (unsigned int i = 32; i < 64; i++) {
- EXPECT_EQ(1U << (i - 32), shuffleDynamic64(1ULL << i, mask));
+ EXPECT_EQ(1U << (i - 32), packedExtract64(1ULL << i, mask));
}
// Try one in the middle
mask = 0x0000ffffffff0000ULL;
- EXPECT_EQ(0U, shuffleDynamic64(0, mask));
- EXPECT_EQ(0xffffffffU, shuffleDynamic64(mask, mask));
+ EXPECT_EQ(0U, packedExtract64(0, mask));
+ EXPECT_EQ(0xffffffffU, packedExtract64(mask, mask));
for (unsigned int i = 16; i < 48; i++) {
- EXPECT_EQ(1U << (i - 16), shuffleDynamic64(1ULL << i, mask));
+ EXPECT_EQ(1U << (i - 16), packedExtract64(1ULL << i, mask));
}
}
-TEST(Shuffle, ShuffleDynamic64_3) {
+TEST(Shuffle, PackedExtract64_3) {
// Try setting every second bit (note: 32 bits, the max we can shuffle)
u64a mask = 0;
for (unsigned int i = 0; i < 64; i += 2) {
}
// Test both cases (all even bits, all odd bits)
- EXPECT_EQ(0xffffffffU, shuffleDynamic64(mask, mask));
- EXPECT_EQ(0xffffffffU, shuffleDynamic64(~mask, ~mask));
- EXPECT_EQ(0U, shuffleDynamic64(~mask, mask));
- EXPECT_EQ(0U, shuffleDynamic64(mask, ~mask));
+ EXPECT_EQ(0xffffffffU, packedExtract64(mask, mask));
+ EXPECT_EQ(0xffffffffU, packedExtract64(~mask, ~mask));
+ EXPECT_EQ(0U, packedExtract64(~mask, mask));
+ EXPECT_EQ(0U, packedExtract64(mask, ~mask));
for (unsigned int i = 0; i < 64; i += 2) {
- EXPECT_EQ(1U << (i/2), shuffleDynamic64(1ULL << i, mask));
- EXPECT_EQ(0U, shuffleDynamic64(1ULL << i, ~mask));
- EXPECT_EQ(1U << (i/2), shuffleDynamic64(1ULL << (i+1), ~mask));
- EXPECT_EQ(0U, shuffleDynamic64(1ULL << (i+1), mask));
+ EXPECT_EQ(1U << (i/2), packedExtract64(1ULL << i, mask));
+ EXPECT_EQ(0U, packedExtract64(1ULL << i, ~mask));
+ EXPECT_EQ(1U << (i/2), packedExtract64(1ULL << (i+1), ~mask));
+ EXPECT_EQ(0U, packedExtract64(1ULL << (i+1), mask));
}
}
+template<typename T>
static
-void build_pshufb_masks_onebit(unsigned int bit, m128 *permute, m128 *compare) {
+void build_pshufb_masks_onebit(unsigned int bit, T *permute, T *compare) {
+ static_assert(sizeof(T) == sizeof(m128) || sizeof(T) == sizeof(m256),
+ "should be valid type");
// permute mask has 0x80 in all bytes except the one we care about
memset(permute, 0x80, sizeof(*permute));
memset(compare, 0, sizeof(*compare));
char *pmsk = (char *)permute;
char *cmsk = (char *)compare;
- pmsk[0] = bit/8;
- cmsk[0] = ~(1 << (bit % 8));
+ u8 off = (bit >= 128) ? 0x10 : 0;
+ pmsk[off] = bit/8;
+ cmsk[off] = ~(1 << (bit % 8));
}
-TEST(Shuffle, ShufflePshufb128_1) {
+TEST(Shuffle, PackedExtract128_1) {
// Try all possible one-bit masks
for (unsigned int i = 0; i < 128; i++) {
// shuffle a single 1 bit to the front
m128 permute, compare;
build_pshufb_masks_onebit(i, &permute, &compare);
- EXPECT_EQ(1U, shufflePshufb128(setbit<m128>(i), permute, compare));
- EXPECT_EQ(1U, shufflePshufb128(ones128(), permute, compare));
+ EXPECT_EQ(1U, packedExtract128(setbit<m128>(i), permute, compare));
+ EXPECT_EQ(1U, packedExtract128(ones128(), permute, compare));
// we should get zero out of these cases
- EXPECT_EQ(0U, shufflePshufb128(zeroes128(), permute, compare));
- EXPECT_EQ(0U, shufflePshufb128(not128(setbit<m128>(i)), permute, compare));
+ EXPECT_EQ(0U, packedExtract128(zeroes128(), permute, compare));
+ EXPECT_EQ(0U, packedExtract128(not128(setbit<m128>(i)), permute, compare));
// we should get zero out of all the other bit positions
for (unsigned int j = 0; (j != i && j < 128); j++) {
- EXPECT_EQ(0U, shufflePshufb128(setbit<m128>(j), permute, compare));
+ EXPECT_EQ(0U, packedExtract128(setbit<m128>(j), permute, compare));
}
}
}
+#if defined(__AVX2__)
+TEST(Shuffle, PackedExtract256_1) {
+ // Try all possible one-bit masks
+ for (unsigned int i = 0; i < 256; i++) {
+ // shuffle a single 1 bit to the front
+ m256 permute, compare;
+ build_pshufb_masks_onebit(i, &permute, &compare);
+ EXPECT_EQ(1U, packedExtract256(setbit<m256>(i), permute, compare));
+ EXPECT_EQ(1U, packedExtract256(ones256(), permute, compare));
+ // we should get zero out of these cases
+ EXPECT_EQ(0U, packedExtract256(zeroes256(), permute, compare));
+ EXPECT_EQ(0U, packedExtract256(not256(setbit<m256>(i)), permute, compare));
+ // we should get zero out of all the other bit positions
+ for (unsigned int j = 0; (j != i && j < 256); j++) {
+ EXPECT_EQ(0U, packedExtract256(setbit<m256>(j), permute, compare));
+ }
+ }
+}
+#endif
} // namespace
#include "util/alloc.h"
#include "util/make_unique.h"
#include "util/simd_utils.h"
-#include "util/simd_utils_ssse3.h"
using namespace std;
using namespace ue2;