On 2025-08-21 22:35, Stephen Hemminger wrote:
Add new compare functions for common small key sizes.
Bugzilla ID: 1775
Suggested-by: Morten Brørup <[email protected]>
Reported-by: Mattias Rönnblom <[email protected]>
Signed-off-by: Stephen Hemminger <[email protected]>
---
lib/hash/rte_cuckoo_hash.c | 54 ++++++++++++++++++++++++++++++++++++++
1 file changed, 54 insertions(+)
diff --git a/lib/hash/rte_cuckoo_hash.c b/lib/hash/rte_cuckoo_hash.c
index 3212695d92..825889c320 100644
--- a/lib/hash/rte_cuckoo_hash.c
+++ b/lib/hash/rte_cuckoo_hash.c
@@ -49,6 +49,11 @@ RTE_LOG_REGISTER_DEFAULT(hash_logtype, INFO);
*/
enum cmp_jump_table_case {
KEY_CUSTOM = 0,
+ KEY_2_BYTES,
+ KEY_4_BYTES,
+ KEY_6_BYTES,
+ KEY_8_BYTES,
+ KEY_12_BYTES,
KEY_16_BYTES,
KEY_32_BYTES,
KEY_48_BYTES,
@@ -86,6 +91,50 @@ rte_hash_k32_cmp_eq(const void *key1, const void *key2,
size_t key_len)
}
#endif
+static inline int
+rte_hash_k2_cmp_eq(const void *key1, const void *key2, size_t key_len
__rte_unused)
+{
+ const uint16_t *k1 = key1;
+ const uint16_t *k2 = key2;
+
What we do now is to require the keys are 16-bit aligned (which wasn't
the case before).
You could
uint16_t k1;
memcpy(&k1, key1, sizeof(uint16_t));
instead.
Would generate the same code, but be safe from any future alignment issues.
Anyway, maybe it's safe to assume the keys are aligned, so this is not
an issue.
+ return k1[0] ^ k2[0];
+}
Haven't you implemented "neq" rather than "eq" here? If the keys are
equal, the result is 0. Should be != 0.
Would it be worth adding a comment like "use XOR to make this
branch-free"? It may not be obvious to all readers.
That said, I’m not sure this trick will actually change the generated
object code - especially if the result of the eq function is still used
in a conditional afterward. Anyway, keeping it seems like a good
conservative approach.
+
+static inline int
+rte_hash_k4_cmp_eq(const void *key1, const void *key2, size_t key_len
__rte_unused)
+{
+ const uint32_t *k1 = key1;
+ const uint32_t *k2 = key2;
+
+ return k1[0] ^ k2[0];
+}
+
+static inline int
+rte_hash_k6_cmp_eq(const void *key1, const void *key2, size_t key_len
__rte_unused)
+{
+ const uint16_t *k1 = key1;
+ const uint16_t *k2 = key2;
+
+ return (k1[0] ^ k2[0]) | (k1[1] ^ k2[1]) | (k1[2] ^ k2[2]);
+}
+
+static inline int
+rte_hash_k8_cmp_eq(const void *key1, const void *key2, size_t key_len
__rte_unused)
+{
+ const uint64_t *k1 = key1;
+ const uint64_t *k2 = key2;
+
+ return k1[0] ^ k2[0];
+}
+
+static inline int
+rte_hash_k12_cmp_eq(const void *key1, const void *key2, size_t key_len
__rte_unused)
+{
+ const uint32_t *k1 = key1;
+ const uint32_t *k2 = key2;
+
+ return (k1[0] ^ k2[0]) | (k1[1] ^ k2[1]) | (k1[2] ^ k2[2]);
+}
static inline int
rte_hash_k48_cmp_eq(const void *key1, const void *key2, size_t key_len)
@@ -228,6 +277,11 @@ void rte_hash_set_cmp_func(struct rte_hash *h,
rte_hash_cmp_eq_t func)
*/
static const rte_hash_cmp_eq_t cmp_jump_table[NUM_KEY_CMP_CASES] = {
[KEY_CUSTOM] = NULL,
+ [KEY_2_BYTES] = rte_hash_k2_cmp_eq,
+ [KEY_4_BYTES] = rte_hash_k4_cmp_eq,
+ [KEY_6_BYTES] = rte_hash_k6_cmp_eq,
+ [KEY_8_BYTES] = rte_hash_k8_cmp_eq,
+ [KEY_12_BYTES] = rte_hash_k12_cmp_eq,
[KEY_16_BYTES] = rte_hash_k16_cmp_eq,
[KEY_32_BYTES] = rte_hash_k32_cmp_eq,
[KEY_48_BYTES] = rte_hash_k48_cmp_eq,