mirror of
https://github.com/sockspls/badfish
synced 2025-07-12 12:09:14 +00:00
Add AVX512 support.
bench: 3909820
This commit is contained in:
parent
7a13d4ed60
commit
961a4dad5c
2 changed files with 70 additions and 5 deletions
26
src/Makefile
26
src/Makefile
|
@ -87,6 +87,7 @@ endif
|
||||||
# sse42 = yes/no --- -msse4.2 --- Use Intel Streaming SIMD Extensions 4.2
|
# sse42 = yes/no --- -msse4.2 --- Use Intel Streaming SIMD Extensions 4.2
|
||||||
# avx2 = yes/no --- -mavx2 --- Use Intel Advanced Vector Extensions 2
|
# avx2 = yes/no --- -mavx2 --- Use Intel Advanced Vector Extensions 2
|
||||||
# pext = yes/no --- -DUSE_PEXT --- Use pext x86_64 asm-instruction
|
# pext = yes/no --- -DUSE_PEXT --- Use pext x86_64 asm-instruction
|
||||||
|
# avx512 = yes/no --- -mavx512vbmi --- Use Intel Advanced Vector Extensions 512
|
||||||
#
|
#
|
||||||
# Note that Makefile is space sensitive, so when adding new architectures
|
# Note that Makefile is space sensitive, so when adding new architectures
|
||||||
# or modifying existing flags, you have to make sure there are no extra spaces
|
# or modifying existing flags, you have to make sure there are no extra spaces
|
||||||
|
@ -105,6 +106,7 @@ sse41 = no
|
||||||
sse42 = no
|
sse42 = no
|
||||||
avx2 = no
|
avx2 = no
|
||||||
pext = no
|
pext = no
|
||||||
|
avx512 = no
|
||||||
|
|
||||||
### 2.2 Architecture specific
|
### 2.2 Architecture specific
|
||||||
ifeq ($(ARCH),general-32)
|
ifeq ($(ARCH),general-32)
|
||||||
|
@ -183,6 +185,20 @@ ifeq ($(ARCH),x86-64-bmi2)
|
||||||
pext = yes
|
pext = yes
|
||||||
endif
|
endif
|
||||||
|
|
||||||
|
ifeq ($(ARCH),x86-64-avx512)
|
||||||
|
arch = x86_64
|
||||||
|
bits = 64
|
||||||
|
prefetch = yes
|
||||||
|
popcnt = yes
|
||||||
|
sse = yes
|
||||||
|
ssse3 = yes
|
||||||
|
sse41 = yes
|
||||||
|
sse42 = yes
|
||||||
|
avx2 = yes
|
||||||
|
pext = yes
|
||||||
|
avx512 = yes
|
||||||
|
endif
|
||||||
|
|
||||||
ifeq ($(ARCH),armv7)
|
ifeq ($(ARCH),armv7)
|
||||||
arch = armv7
|
arch = armv7
|
||||||
prefetch = yes
|
prefetch = yes
|
||||||
|
@ -411,6 +427,13 @@ ifeq ($(avx2),yes)
|
||||||
endif
|
endif
|
||||||
endif
|
endif
|
||||||
|
|
||||||
|
ifeq ($(avx512),yes)
|
||||||
|
CXXFLAGS += -DUSE_AVX512
|
||||||
|
ifeq ($(comp),$(filter $(comp),gcc clang mingw msys2))
|
||||||
|
CXXFLAGS += -mavx512vbmi
|
||||||
|
endif
|
||||||
|
endif
|
||||||
|
|
||||||
ifeq ($(sse42),yes)
|
ifeq ($(sse42),yes)
|
||||||
CXXFLAGS += -DUSE_SSE42
|
CXXFLAGS += -DUSE_SSE42
|
||||||
ifeq ($(comp),$(filter $(comp),gcc clang mingw msys2))
|
ifeq ($(comp),$(filter $(comp),gcc clang mingw msys2))
|
||||||
|
@ -493,6 +516,7 @@ help:
|
||||||
@echo ""
|
@echo ""
|
||||||
@echo "Supported archs:"
|
@echo "Supported archs:"
|
||||||
@echo ""
|
@echo ""
|
||||||
|
@echo "x86-64-avx512 > x86 64-bit with avx512 support"
|
||||||
@echo "x86-64-bmi2 > x86 64-bit with bmi2 support"
|
@echo "x86-64-bmi2 > x86 64-bit with bmi2 support"
|
||||||
@echo "x86-64-avx2 > x86 64-bit with avx2 support"
|
@echo "x86-64-avx2 > x86 64-bit with avx2 support"
|
||||||
@echo "x86-64-sse42 > x86 64-bit with sse42 support"
|
@echo "x86-64-sse42 > x86 64-bit with sse42 support"
|
||||||
|
@ -599,6 +623,7 @@ config-sanity:
|
||||||
@echo "sse42: '$(sse42)'"
|
@echo "sse42: '$(sse42)'"
|
||||||
@echo "avx2: '$(avx2)'"
|
@echo "avx2: '$(avx2)'"
|
||||||
@echo "pext: '$(pext)'"
|
@echo "pext: '$(pext)'"
|
||||||
|
@echo "avx512: '$(avx512)'"
|
||||||
@echo ""
|
@echo ""
|
||||||
@echo "Flags:"
|
@echo "Flags:"
|
||||||
@echo "CXX: $(CXX)"
|
@echo "CXX: $(CXX)"
|
||||||
|
@ -622,6 +647,7 @@ config-sanity:
|
||||||
@test "$(sse42)" = "yes" || test "$(sse42)" = "no"
|
@test "$(sse42)" = "yes" || test "$(sse42)" = "no"
|
||||||
@test "$(avx2)" = "yes" || test "$(avx2)" = "no"
|
@test "$(avx2)" = "yes" || test "$(avx2)" = "no"
|
||||||
@test "$(pext)" = "yes" || test "$(pext)" = "no"
|
@test "$(pext)" = "yes" || test "$(pext)" = "no"
|
||||||
|
@test "$(avx512)" = "yes" || test "$(avx512)" = "no"
|
||||||
@test "$(comp)" = "gcc" || test "$(comp)" = "icc" || test "$(comp)" = "mingw" || test "$(comp)" = "clang"
|
@test "$(comp)" = "gcc" || test "$(comp)" = "icc" || test "$(comp)" = "mingw" || test "$(comp)" = "clang"
|
||||||
|
|
||||||
$(EXE): $(OBJS)
|
$(EXE): $(OBJS)
|
||||||
|
|
|
@ -82,7 +82,11 @@ class AffineTransform {
|
||||||
const auto input = previous_layer_.Propagate(
|
const auto input = previous_layer_.Propagate(
|
||||||
transformed_features, buffer + kSelfBufferSize);
|
transformed_features, buffer + kSelfBufferSize);
|
||||||
const auto output = reinterpret_cast<OutputType*>(buffer);
|
const auto output = reinterpret_cast<OutputType*>(buffer);
|
||||||
#if defined(USE_AVX2)
|
#if defined(USE_AVX512)
|
||||||
|
constexpr IndexType kNumChunks = kPaddedInputDimensions / (kSimdWidth * 2);
|
||||||
|
const __m512i kOnes = _mm512_set1_epi16(1);
|
||||||
|
const auto input_vector = reinterpret_cast<const __m512i*>(input);
|
||||||
|
#elif defined(USE_AVX2)
|
||||||
constexpr IndexType kNumChunks = kPaddedInputDimensions / kSimdWidth;
|
constexpr IndexType kNumChunks = kPaddedInputDimensions / kSimdWidth;
|
||||||
const __m256i kOnes = _mm256_set1_epi16(1);
|
const __m256i kOnes = _mm256_set1_epi16(1);
|
||||||
const auto input_vector = reinterpret_cast<const __m256i*>(input);
|
const auto input_vector = reinterpret_cast<const __m256i*>(input);
|
||||||
|
@ -96,8 +100,43 @@ class AffineTransform {
|
||||||
#endif
|
#endif
|
||||||
for (IndexType i = 0; i < kOutputDimensions; ++i) {
|
for (IndexType i = 0; i < kOutputDimensions; ++i) {
|
||||||
const IndexType offset = i * kPaddedInputDimensions;
|
const IndexType offset = i * kPaddedInputDimensions;
|
||||||
#if defined(USE_AVX2)
|
#if defined(USE_AVX512)
|
||||||
__m256i sum = _mm256_set_epi32(0, 0, 0, 0, 0, 0, 0, biases_[i]);
|
__m512i sum = _mm512_setzero_si512();
|
||||||
|
const auto row = reinterpret_cast<const __m512i*>(&weights_[offset]);
|
||||||
|
for (IndexType j = 0; j < kNumChunks; ++j) {
|
||||||
|
#if defined(__MINGW32__) || defined(__MINGW64__)
|
||||||
|
__m512i product = _mm512_maddubs_epi16(_mm512_loadu_si512(&input_vector[j]), _mm512_load_si512(&row[j]));
|
||||||
|
#else
|
||||||
|
__m512i product = _mm512_maddubs_epi16(_mm512_load_si512(&input_vector[j]), _mm512_load_si512(&row[j]));
|
||||||
|
#endif
|
||||||
|
product = _mm512_madd_epi16(product, kOnes);
|
||||||
|
sum = _mm512_add_epi32(sum, product);
|
||||||
|
}
|
||||||
|
output[i] = _mm512_reduce_add_epi32(sum) + biases_[i];
|
||||||
|
|
||||||
|
// Note: Changing kMaxSimdWidth from 32 to 64 breaks loading existing networks.
|
||||||
|
// As a result kPaddedInputDimensions may not be an even multiple of 64(512bit)
|
||||||
|
// and we have to do one more 256bit chunk.
|
||||||
|
if (kPaddedInputDimensions != kNumChunks * kSimdWidth * 2)
|
||||||
|
{
|
||||||
|
const auto iv_256 = reinterpret_cast<const __m256i*>(input);
|
||||||
|
const auto row_256 = reinterpret_cast<const __m256i*>(&weights_[offset]);
|
||||||
|
int j = kNumChunks * 2;
|
||||||
|
#if defined(__MINGW32__) || defined(__MINGW64__) // See HACK comment below in AVX2.
|
||||||
|
__m256i sum256 = _mm256_maddubs_epi16(_mm256_loadu_si256(&iv_256[j]), _mm256_load_si256(&row_256[j]));
|
||||||
|
#else
|
||||||
|
__m256i sum256 = _mm256_maddubs_epi16(_mm256_load_si256(&iv_256[j]), _mm256_load_si256(&row_256[j]));
|
||||||
|
#endif
|
||||||
|
sum256 = _mm256_madd_epi16(sum256, _mm256_set1_epi16(1));
|
||||||
|
|
||||||
|
sum256 = _mm256_hadd_epi32(sum256, sum256);
|
||||||
|
sum256 = _mm256_hadd_epi32(sum256, sum256);
|
||||||
|
const __m128i lo = _mm256_extracti128_si256(sum256, 0);
|
||||||
|
const __m128i hi = _mm256_extracti128_si256(sum256, 1);
|
||||||
|
output[i] += _mm_cvtsi128_si32(lo) + _mm_cvtsi128_si32(hi);
|
||||||
|
}
|
||||||
|
#elif defined(USE_AVX2)
|
||||||
|
__m256i sum = _mm256_setzero_si256();
|
||||||
const auto row = reinterpret_cast<const __m256i*>(&weights_[offset]);
|
const auto row = reinterpret_cast<const __m256i*>(&weights_[offset]);
|
||||||
for (IndexType j = 0; j < kNumChunks; ++j) {
|
for (IndexType j = 0; j < kNumChunks; ++j) {
|
||||||
__m256i product = _mm256_maddubs_epi16(
|
__m256i product = _mm256_maddubs_epi16(
|
||||||
|
@ -117,7 +156,7 @@ class AffineTransform {
|
||||||
sum = _mm256_hadd_epi32(sum, sum);
|
sum = _mm256_hadd_epi32(sum, sum);
|
||||||
const __m128i lo = _mm256_extracti128_si256(sum, 0);
|
const __m128i lo = _mm256_extracti128_si256(sum, 0);
|
||||||
const __m128i hi = _mm256_extracti128_si256(sum, 1);
|
const __m128i hi = _mm256_extracti128_si256(sum, 1);
|
||||||
output[i] = _mm_cvtsi128_si32(lo) + _mm_cvtsi128_si32(hi);
|
output[i] = _mm_cvtsi128_si32(lo) + _mm_cvtsi128_si32(hi) + biases_[i];
|
||||||
#elif defined(USE_SSSE3)
|
#elif defined(USE_SSSE3)
|
||||||
__m128i sum = _mm_cvtsi32_si128(biases_[i]);
|
__m128i sum = _mm_cvtsi32_si128(biases_[i]);
|
||||||
const auto row = reinterpret_cast<const __m128i*>(&weights_[offset]);
|
const auto row = reinterpret_cast<const __m128i*>(&weights_[offset]);
|
||||||
|
|
Loading…
Add table
Reference in a new issue