Message ID | 1634290206-251913-2-git-send-email-vladimir.medvedkin@intel.com (mailing list archive) |
---|---|
State | Superseded, archived |
Delegated to: | Thomas Monjalon |
Headers | show |
Series | optimized Toeplitz hash implementation | expand |
Context | Check | Description |
---|---|---|
ci/checkpatch | success | coding style OK |
On Fri, 15 Oct 2021 10:30:02 +0100 Vladimir Medvedkin <vladimir.medvedkin@intel.com> wrote: > + m[i * 8 + j] = (rss_key[i] << j)| > + (uint8_t)((uint16_t)(rss_key[i + 1]) >> > + (8 - j)); > + } This ends up being harder than necessary to read. Maybe split into multiple statements and/or use temporary variable. > +RTE_INIT(rte_thash_gfni_init) > +{ > + rte_thash_gfni_supported = 0; Not necessary in C globals are initialized to zero by default. By removing that the constructor can be totally behind #ifdef > +__rte_internal > +static inline __m512i > +__rte_thash_gfni(const uint64_t *mtrx, const uint8_t *tuple, > + const uint8_t *secondary_tuple, int len) > +{ > + __m512i permute_idx = _mm512_set_epi8(7, 6, 5, 4, 7, 6, 5, 4, > + 6, 5, 4, 3, 6, 5, 4, 3, > + 5, 4, 3, 2, 5, 4, 3, 2, > + 4, 3, 2, 1, 4, 3, 2, 1, > + 3, 2, 1, 0, 3, 2, 1, 0, > + 2, 1, 0, -1, 2, 1, 0, -1, > + 1, 0, -1, -2, 1, 0, -1, -2, > + 0, -1, -2, -3, 0, -1, -2, -3); NAK Please don't put the implementation in an inline. This makes it harder to support (API/ABI) and blocks other architectures from implementing same thing with different instructions.
> On Fri, 15 Oct 2021 10:30:02 +0100 > Vladimir Medvedkin <vladimir.medvedkin@intel.com> wrote: > > > + m[i * 8 + j] = (rss_key[i] << j)| > > + (uint8_t)((uint16_t)(rss_key[i + 1]) >> > > + (8 - j)); > > + } > > This ends up being harder than necessary to read. Maybe split into > multiple statements and/or use temporary variable. > > > +RTE_INIT(rte_thash_gfni_init) > > +{ > > + rte_thash_gfni_supported = 0; > > Not necessary in C globals are initialized to zero by default. > > By removing that the constructor can be totally behind #ifdef > > > +__rte_internal > > +static inline __m512i > > +__rte_thash_gfni(const uint64_t *mtrx, const uint8_t *tuple, > > + const uint8_t *secondary_tuple, int len) > > +{ > > + __m512i permute_idx = _mm512_set_epi8(7, 6, 5, 4, 7, 6, 5, 4, > > + 6, 5, 4, 3, 6, 5, 4, 3, > > + 5, 4, 3, 2, 5, 4, 3, 2, > > + 4, 3, 2, 1, 4, 3, 2, 1, > > + 3, 2, 1, 0, 3, 2, 1, 0, > > + 2, 1, 0, -1, 2, 1, 0, -1, > > + 1, 0, -1, -2, 1, 0, -1, -2, > > + 0, -1, -2, -3, 0, -1, -2, -3); > > NAK > > Please don't put the implementation in an inline. This makes it harder > to support (API/ABI) and blocks other architectures from implementing > same thing with different instructions. I don't really understand your reasoning here. rte_thash_gfni.h is an arch-specific header, which provides arch-specific optimizations for RSS hash calculation (Vladimir pls correct me if I am wrong here). We do have dozens of inline functions that do use arch-specific instructions (both x86 and arm) for different purposes: sync primitives, memory-ordering, cache manipulations, LPM lookup, TSX, power-saving, etc. That's a usual trade-off taken for performance reasons, when extra function call costs too much comparing to the operation itself. Why it suddenly became a problem for that particular case and how exactly it blocks other architectures? Also I don't understand how it makes things harder in terms of API/ABI stability. As I can see this patch doesn't introduce any public structs/unions. All functions take as arguments just raw data buffers and length. To summarize - in general, I don't see any good reason why this patch shouldn't be allowed. Konstantin
Hi Stephen, Thanks for reviewing On 15/10/2021 18:58, Stephen Hemminger wrote: > On Fri, 15 Oct 2021 10:30:02 +0100 > Vladimir Medvedkin <vladimir.medvedkin@intel.com> wrote: > >> + m[i * 8 + j] = (rss_key[i] << j)| >> + (uint8_t)((uint16_t)(rss_key[i + 1]) >> >> + (8 - j)); >> + } > > This ends up being harder than necessary to read. Maybe split into > multiple statements and/or use temporary variable. > >> +RTE_INIT(rte_thash_gfni_init) >> +{ >> + rte_thash_gfni_supported = 0; > > Not necessary in C globals are initialized to zero by default. > > By removing that the constructor can be totally behind #ifdef > >> +__rte_internal >> +static inline __m512i >> +__rte_thash_gfni(const uint64_t *mtrx, const uint8_t *tuple, >> + const uint8_t *secondary_tuple, int len) >> +{ >> + __m512i permute_idx = _mm512_set_epi8(7, 6, 5, 4, 7, 6, 5, 4, >> + 6, 5, 4, 3, 6, 5, 4, 3, >> + 5, 4, 3, 2, 5, 4, 3, 2, >> + 4, 3, 2, 1, 4, 3, 2, 1, >> + 3, 2, 1, 0, 3, 2, 1, 0, >> + 2, 1, 0, -1, 2, 1, 0, -1, >> + 1, 0, -1, -2, 1, 0, -1, -2, >> + 0, -1, -2, -3, 0, -1, -2, -3); > > NAK > > Please don't put the implementation in an inline. This makes it harder > to support (API/ABI) and blocks other architectures from implementing > same thing with different instructions. > By making this function not inline, its performance drops by about 2 times. Compiler optimization (at least with respect to the len argument) helps a lot in the implementation.
On Mon, 18 Oct 2021 10:40:00 +0000 "Ananyev, Konstantin" <konstantin.ananyev@intel.com> wrote: > > On Fri, 15 Oct 2021 10:30:02 +0100 > > Vladimir Medvedkin <vladimir.medvedkin@intel.com> wrote: > > > > > + m[i * 8 + j] = (rss_key[i] << j)| > > > + (uint8_t)((uint16_t)(rss_key[i + 1]) >> > > > + (8 - j)); > > > + } > > > > This ends up being harder than necessary to read. Maybe split into > > multiple statements and/or use temporary variable. > > > > > +RTE_INIT(rte_thash_gfni_init) > > > +{ > > > + rte_thash_gfni_supported = 0; > > > > Not necessary in C globals are initialized to zero by default. > > > > By removing that the constructor can be totally behind #ifdef > > > > > +__rte_internal > > > +static inline __m512i > > > +__rte_thash_gfni(const uint64_t *mtrx, const uint8_t *tuple, > > > + const uint8_t *secondary_tuple, int len) > > > +{ > > > + __m512i permute_idx = _mm512_set_epi8(7, 6, 5, 4, 7, 6, 5, 4, > > > + 6, 5, 4, 3, 6, 5, 4, 3, > > > + 5, 4, 3, 2, 5, 4, 3, 2, > > > + 4, 3, 2, 1, 4, 3, 2, 1, > > > + 3, 2, 1, 0, 3, 2, 1, 0, > > > + 2, 1, 0, -1, 2, 1, 0, -1, > > > + 1, 0, -1, -2, 1, 0, -1, -2, > > > + 0, -1, -2, -3, 0, -1, -2, -3); > > > > NAK > > > > Please don't put the implementation in an inline. This makes it harder > > to support (API/ABI) and blocks other architectures from implementing > > same thing with different instructions. > > I don't really understand your reasoning here. > rte_thash_gfni.h is an arch-specific header, which provides > arch-specific optimizations for RSS hash calculation > (Vladimir pls correct me if I am wrong here). Ok, but rte_thash_gfni.h is included on all architectures. > We do have dozens of inline functions that do use arch-specific instructions (both x86 and arm) > for different purposes: > sync primitives, memory-ordering, cache manipulations, LPM lookup, TSX, power-saving, etc. > That's a usual trade-off taken for performance reasons, when extra function call > costs too much comparing to the operation itself. > Why it suddenly became a problem for that particular case and how exactly it blocks other architectures? > Also I don't understand how it makes things harder in terms of API/ABI stability. > As I can see this patch doesn't introduce any public structs/unions. > All functions take as arguments just raw data buffers and length. > To summarize - in general, I don't see any good reason why this patch shouldn't be allowed. > Konstantin The comments about rte_thash_gfni_supported initialization still apply. Why not: #ifdef __GFNI__ RTE_INIT(rte_thash_gfni_init) { if (rte_cpu_get_flag_enabled(RTE_CPUFLAG_GFNI)) rte_thash_gfni_supported = 1; } #endif
Hi Stephen, On 19/10/2021 03:15, Stephen Hemminger wrote: > On Mon, 18 Oct 2021 10:40:00 +0000 > "Ananyev, Konstantin" <konstantin.ananyev@intel.com> wrote: > >>> On Fri, 15 Oct 2021 10:30:02 +0100 >>> Vladimir Medvedkin <vladimir.medvedkin@intel.com> wrote: >>> >>>> + m[i * 8 + j] = (rss_key[i] << j)| >>>> + (uint8_t)((uint16_t)(rss_key[i + 1]) >> >>>> + (8 - j)); >>>> + } >>> >>> This ends up being harder than necessary to read. Maybe split into >>> multiple statements and/or use temporary variable. >>> >>>> +RTE_INIT(rte_thash_gfni_init) >>>> +{ >>>> + rte_thash_gfni_supported = 0; >>> >>> Not necessary in C globals are initialized to zero by default. >>> >>> By removing that the constructor can be totally behind #ifdef >>> >>>> +__rte_internal >>>> +static inline __m512i >>>> +__rte_thash_gfni(const uint64_t *mtrx, const uint8_t *tuple, >>>> + const uint8_t *secondary_tuple, int len) >>>> +{ >>>> + __m512i permute_idx = _mm512_set_epi8(7, 6, 5, 4, 7, 6, 5, 4, >>>> + 6, 5, 4, 3, 6, 5, 4, 3, >>>> + 5, 4, 3, 2, 5, 4, 3, 2, >>>> + 4, 3, 2, 1, 4, 3, 2, 1, >>>> + 3, 2, 1, 0, 3, 2, 1, 0, >>>> + 2, 1, 0, -1, 2, 1, 0, -1, >>>> + 1, 0, -1, -2, 1, 0, -1, -2, >>>> + 0, -1, -2, -3, 0, -1, -2, -3); >>> >>> NAK >>> >>> Please don't put the implementation in an inline. This makes it harder >>> to support (API/ABI) and blocks other architectures from implementing >>> same thing with different instructions. >> >> I don't really understand your reasoning here. >> rte_thash_gfni.h is an arch-specific header, which provides >> arch-specific optimizations for RSS hash calculation >> (Vladimir pls correct me if I am wrong here). > > Ok, but rte_thash_gfni.h is included on all architectures. > Ok, I'll rework the patch to move x86 + avx512 related things into x86 arch specific header. Would that suit? >> We do have dozens of inline functions that do use arch-specific instructions (both x86 and arm) >> for different purposes: >> sync primitives, memory-ordering, cache manipulations, LPM lookup, TSX, power-saving, etc. >> That's a usual trade-off taken for performance reasons, when extra function call >> costs too much comparing to the operation itself. >> Why it suddenly became a problem for that particular case and how exactly it blocks other architectures? >> Also I don't understand how it makes things harder in terms of API/ABI stability. >> As I can see this patch doesn't introduce any public structs/unions. >> All functions take as arguments just raw data buffers and length. >> To summarize - in general, I don't see any good reason why this patch shouldn't be allowed. >> Konstantin > > The comments about rte_thash_gfni_supported initialization still apply. > Why not: > > #ifdef __GFNI__ > RTE_INIT(rte_thash_gfni_init) > { > if (rte_cpu_get_flag_enabled(RTE_CPUFLAG_GFNI)) > rte_thash_gfni_supported = 1; > } > #endif > Agree, I'll reflect this changes in v3.
diff --git a/doc/api/doxy-api-index.md b/doc/api/doxy-api-index.md index 1992107..7549477 100644 --- a/doc/api/doxy-api-index.md +++ b/doc/api/doxy-api-index.md @@ -139,6 +139,7 @@ The public API headers are grouped by topics: [hash] (@ref rte_hash.h), [jhash] (@ref rte_jhash.h), [thash] (@ref rte_thash.h), + [thash_gfni] (@ref rte_thash_gfni.h), [FBK hash] (@ref rte_fbk_hash.h), [CRC hash] (@ref rte_hash_crc.h) diff --git a/lib/hash/meson.build b/lib/hash/meson.build index 9bc5ef9..40444ac 100644 --- a/lib/hash/meson.build +++ b/lib/hash/meson.build @@ -7,6 +7,7 @@ headers = files( 'rte_hash.h', 'rte_jhash.h', 'rte_thash.h', + 'rte_thash_gfni.h', ) indirect_headers += files('rte_crc_arm64.h') diff --git a/lib/hash/rte_thash.c b/lib/hash/rte_thash.c index 696a112..59a8b8e 100644 --- a/lib/hash/rte_thash.c +++ b/lib/hash/rte_thash.c @@ -13,6 +13,7 @@ #include <rte_eal_memconfig.h> #include <rte_log.h> #include <rte_malloc.h> +#include <rte_thash_gfni.h> #define THASH_NAME_LEN 64 #define TOEPLITZ_HASH_LEN 32 @@ -90,6 +91,24 @@ struct rte_thash_ctx { uint8_t hash_key[0]; }; +/** Flag indicating GFNI support */ +uint8_t rte_thash_gfni_supported; + +void +rte_thash_complete_matrix(uint64_t *matrixes, const uint8_t *rss_key, int size) +{ + int i, j; + uint8_t *m = (uint8_t *)matrixes; + + for (i = 0; i < size; i++) { + for (j = 0; j < 8; j++) { + m[i * 8 + j] = (rss_key[i] << j)| + (uint8_t)((uint16_t)(rss_key[i + 1]) >> + (8 - j)); + } + } +} + static inline uint32_t get_bit_lfsr(struct thash_lfsr *lfsr) { @@ -761,3 +780,12 @@ rte_thash_adjust_tuple(struct rte_thash_ctx *ctx, return ret; } + +RTE_INIT(rte_thash_gfni_init) +{ + rte_thash_gfni_supported = 0; +#ifdef __GFNI__ + if (rte_cpu_get_flag_enabled(RTE_CPUFLAG_GFNI)) + rte_thash_gfni_supported = 1; +#endif +} diff --git a/lib/hash/rte_thash.h b/lib/hash/rte_thash.h index 76109fc..e4f14a5 100644 --- a/lib/hash/rte_thash.h +++ b/lib/hash/rte_thash.h @@ -28,6 +28,7 @@ extern "C" { #include <rte_config.h> #include <rte_ip.h> #include <rte_common.h> +#include <rte_thash_gfni.h> #if defined(RTE_ARCH_X86) || defined(__ARM_NEON) #include <rte_vect.h> @@ -113,6 +114,9 @@ union rte_thash_tuple { }; #endif +/** Flag indicating GFNI support */ +extern uint8_t rte_thash_gfni_supported; + /** * Prepare special converted key to use with rte_softrss_be() * @param orig @@ -223,6 +227,26 @@ rte_softrss_be(uint32_t *input_tuple, uint32_t input_len, return ret; } +/** + * Converts Toeplitz hash key (RSS key) into matrixes required + * for GFNI implementation + * + * @warning + * @b EXPERIMENTAL: this API may change without prior notice. + * + * @param matrixes + * pointer to the memory where matrices will be written. + * Note: the size of this memory must be equal to size * 8 + * @param rss_key + * pointer to the Toeplitz hash key + * @param size + * Size of the rss_key in bytes. + */ +__rte_experimental +void +rte_thash_complete_matrix(uint64_t *matrixes, const uint8_t *rss_key, + int size); + /** @internal Logarithm of minimum size of the RSS ReTa */ #define RTE_THASH_RETA_SZ_MIN 2U /** @internal Logarithm of maximum size of the RSS ReTa */ diff --git a/lib/hash/rte_thash_gfni.h b/lib/hash/rte_thash_gfni.h new file mode 100644 index 0000000..2e5de0d --- /dev/null +++ b/lib/hash/rte_thash_gfni.h @@ -0,0 +1,232 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(c) 2021 Intel Corporation + */ + +#ifndef _RTE_THASH_GFNI_H_ +#define _RTE_THASH_GFNI_H_ + +/** + * @file + * + * Optimized Toeplitz hash functions implementation + * using Galois Fields New Instructions. + */ + +#include <rte_vect.h> + +#ifdef __cplusplus +extern "C" { +#endif + +#ifdef __GFNI__ + +#define RTE_THASH_FIRST_ITER_MSK 0x0f0f0f0f0f0e0c08 +#define RTE_THASH_PERM_MSK 0x0f0f0f0f0f0f0f0f +#define RTE_THASH_FIRST_ITER_MSK_2 0xf0f0f0f0f0e0c080 +#define RTE_THASH_PERM_MSK_2 0xf0f0f0f0f0f0f0f0 +#define RTE_THASH_REWIND_MSK 0x0000000000113377 + +__rte_internal +static inline void +__rte_thash_xor_reduce(__m512i xor_acc, uint32_t *val_1, uint32_t *val_2) +{ + __m256i tmp_256_1, tmp_256_2; + __m128i tmp128_1, tmp128_2; + uint64_t tmp_1, tmp_2; + + tmp_256_1 = _mm512_castsi512_si256(xor_acc); + tmp_256_2 = _mm512_extracti32x8_epi32(xor_acc, 1); + tmp_256_1 = _mm256_xor_si256(tmp_256_1, tmp_256_2); + + tmp128_1 = _mm256_castsi256_si128(tmp_256_1); + tmp128_2 = _mm256_extracti32x4_epi32(tmp_256_1, 1); + tmp128_1 = _mm_xor_si128(tmp128_1, tmp128_2); + + tmp_1 = _mm_extract_epi64(tmp128_1, 0); + tmp_2 = _mm_extract_epi64(tmp128_1, 1); + tmp_1 ^= tmp_2; + + *val_1 = (uint32_t)tmp_1; + *val_2 = (uint32_t)(tmp_1 >> 32); +} + +__rte_internal +static inline __m512i +__rte_thash_gfni(const uint64_t *mtrx, const uint8_t *tuple, + const uint8_t *secondary_tuple, int len) +{ + __m512i permute_idx = _mm512_set_epi8(7, 6, 5, 4, 7, 6, 5, 4, + 6, 5, 4, 3, 6, 5, 4, 3, + 5, 4, 3, 2, 5, 4, 3, 2, + 4, 3, 2, 1, 4, 3, 2, 1, + 3, 2, 1, 0, 3, 2, 1, 0, + 2, 1, 0, -1, 2, 1, 0, -1, + 1, 0, -1, -2, 1, 0, -1, -2, + 0, -1, -2, -3, 0, -1, -2, -3); + + const __m512i rewind_idx = _mm512_set_epi8(0, 0, 0, 0, 0, 0, 0, 0, + 0, 0, 0, 0, 0, 0, 0, 0, + 0, 0, 0, 0, 0, 0, 0, 0, + 0, 0, 0, 0, 0, 0, 0, 0, + 0, 0, 0, 0, 0, 0, 0, 0, + 0, 0, 0, 59, 0, 0, 0, 59, + 0, 0, 59, 58, 0, 0, 59, 58, + 0, 59, 58, 57, 0, 59, 58, 57); + const __mmask64 rewind_mask = RTE_THASH_REWIND_MSK; + const __m512i shift_8 = _mm512_set1_epi8(8); + __m512i xor_acc = _mm512_setzero_si512(); + __m512i perm_bytes = _mm512_setzero_si512(); + __m512i vals, matrixes, tuple_bytes, tuple_bytes_2; + __mmask64 load_mask, permute_mask, permute_mask_2; + int chunk_len = 0, i = 0; + uint8_t mtrx_msk; + const int prepend = 3; + + for (; len > 0; len -= 64, tuple += 64) { + if (i == 8) + perm_bytes = _mm512_maskz_permutexvar_epi8(rewind_mask, + rewind_idx, perm_bytes); + + permute_mask = RTE_THASH_FIRST_ITER_MSK; + load_mask = (len >= 64) ? UINT64_MAX : ((1ULL << len) - 1); + tuple_bytes = _mm512_maskz_loadu_epi8(load_mask, tuple); + if (secondary_tuple) { + permute_mask_2 = RTE_THASH_FIRST_ITER_MSK_2; + tuple_bytes_2 = _mm512_maskz_loadu_epi8(load_mask, + secondary_tuple); + } + + chunk_len = __builtin_popcountll(load_mask); + for (i = 0; i < ((chunk_len + prepend) / 8); i++, mtrx += 8) { + perm_bytes = _mm512_mask_permutexvar_epi8(perm_bytes, + permute_mask, permute_idx, tuple_bytes); + + if (secondary_tuple) + perm_bytes = + _mm512_mask_permutexvar_epi8(perm_bytes, + permute_mask_2, permute_idx, + tuple_bytes_2); + + matrixes = _mm512_maskz_loadu_epi64(UINT8_MAX, mtrx); + vals = _mm512_gf2p8affine_epi64_epi8(perm_bytes, + matrixes, 0); + + xor_acc = _mm512_xor_si512(xor_acc, vals); + permute_idx = _mm512_add_epi8(permute_idx, shift_8); + permute_mask = RTE_THASH_PERM_MSK; + if (secondary_tuple) + permute_mask_2 = RTE_THASH_PERM_MSK_2; + } + } + + int rest_len = (chunk_len + prepend) % 8; + if (rest_len != 0) { + mtrx_msk = (1 << (rest_len % 8)) - 1; + matrixes = _mm512_maskz_loadu_epi64(mtrx_msk, mtrx); + if (i == 8) { + perm_bytes = _mm512_maskz_permutexvar_epi8(rewind_mask, + rewind_idx, perm_bytes); + } else { + perm_bytes = _mm512_mask_permutexvar_epi8(perm_bytes, + permute_mask, permute_idx, tuple_bytes); + + if (secondary_tuple) + perm_bytes = + _mm512_mask_permutexvar_epi8( + perm_bytes, permute_mask_2, + permute_idx, tuple_bytes_2); + } + + vals = _mm512_gf2p8affine_epi64_epi8(perm_bytes, matrixes, 0); + xor_acc = _mm512_xor_si512(xor_acc, vals); + } + + return xor_acc; +} + +/** + * Calculate Toeplitz hash. + * + * @warning + * @b EXPERIMENTAL: this API may change without prior notice. + * + * @param m + * Pointer to the matrices generated from the corresponding + * RSS hash key using rte_thash_complete_matrix(). + * @param tuple + * Pointer to the data to be hashed. Data must be in network byte order. + * @param len + * Length of the data to be hashed. + * @return + * Calculated Toeplitz hash value. + */ +__rte_experimental +static inline uint32_t +rte_thash_gfni(const uint64_t *m, const uint8_t *tuple, int len) +{ + uint32_t val, val_zero; + + __m512i xor_acc = __rte_thash_gfni(m, tuple, NULL, len); + __rte_thash_xor_reduce(xor_acc, &val, &val_zero); + + return val; +} + +/** + * Calculate Toeplitz hash for two independent data buffers. + * + * @warning + * @b EXPERIMENTAL: this API may change without prior notice. + * + * @param m + * Pointer to the matrices generated from the corresponding + * RSS hash key using rte_thash_complete_matrix(). + * @param tuple_1 + * Pointer to the data to be hashed. Data must be in network byte order. + * @param tuple_2 + * Pointer to the data to be hashed. Data must be in network byte order. + * @param len + * Length of the largest data buffer to be hashed. + * @param val_1 + * Pointer to uint32_t where to put calculated Toeplitz hash value for + * the first tuple. + * @param val_2 + * Pointer to uint32_t where to put calculated Toeplitz hash value for + * the second tuple. + */ +__rte_experimental +static inline void +rte_thash_gfni_x2(const uint64_t *mtrx, const uint8_t *tuple_1, + const uint8_t *tuple_2, int len, + uint32_t *val_1, uint32_t *val_2) +{ + __m512i xor_acc = __rte_thash_gfni(mtrx, tuple_1, tuple_2, len); + __rte_thash_xor_reduce(xor_acc, val_1, val_2); +} + +#else /* __GFNI__ */ + +static inline uint32_t +rte_thash_gfni(const uint64_t *mtrx __rte_unused, + const uint8_t *key __rte_unused, int len __rte_unused) +{ + return 0; +} + +static inline void +rte_thash_gfni_x2(const uint64_t *mtrx __rte_unused, + const uint8_t *tuple_1 __rte_unused, + const uint8_t *tuple_2 __rte_unused, int len __rte_unused, + uint32_t *val_1 __rte_unused, uint32_t *val_2 __rte_unused) +{ + *val_1 = 0; + *val_2 = 0; +} + +#endif + +#ifdef __cplusplus +} +#endif + +#endif /* _RTE_THASH_GFNI_H_ */ diff --git a/lib/hash/version.map b/lib/hash/version.map index ce4309a..cecf922 100644 --- a/lib/hash/version.map +++ b/lib/hash/version.map @@ -39,10 +39,12 @@ EXPERIMENTAL { rte_hash_rcu_qsbr_add; rte_thash_add_helper; rte_thash_adjust_tuple; + rte_thash_complete_matrix; rte_thash_find_existing; rte_thash_free_ctx; rte_thash_get_complement; rte_thash_get_helper; rte_thash_get_key; + rte_thash_gfni_supported; rte_thash_init_ctx; };
This patch add a new Toeplitz hash implementation using Galios Fields New Instructions (GFNI). Signed-off-by: Vladimir Medvedkin <vladimir.medvedkin@intel.com> --- doc/api/doxy-api-index.md | 1 + lib/hash/meson.build | 1 + lib/hash/rte_thash.c | 28 ++++++ lib/hash/rte_thash.h | 24 +++++ lib/hash/rte_thash_gfni.h | 232 ++++++++++++++++++++++++++++++++++++++++++++++ lib/hash/version.map | 2 + 6 files changed, 288 insertions(+) create mode 100644 lib/hash/rte_thash_gfni.h