17 #undef BIHASH_KVP_PER_PAGE 18 #undef BIHASH_32_64_SVM 19 #undef BIHASH_ENABLE_STATS 20 #undef BIHASH_KVP_AT_BUCKET_LEVEL 21 #undef BIHASH_LAZY_INSTANTIATE 22 #undef BIHASH_BUCKET_PREFETCH_CACHE_LINES 24 #define BIHASH_TYPE _48_8 25 #define BIHASH_KVP_PER_PAGE 4 26 #define BIHASH_KVP_AT_BUCKET_LEVEL 0 27 #define BIHASH_LAZY_INSTANTIATE 1 28 #define BIHASH_BUCKET_PREFETCH_CACHE_LINES 1 30 #ifndef __included_bihash_48_8_h__ 31 #define __included_bihash_48_8_h__ 49 if (v->
key[0] == ~0ULL && v->
value == ~0ULL)
57 #ifdef clib_crc32c_uses_intrinsics 58 return clib_crc32c ((
u8 *) v->
key, 48);
71 s =
format (s,
"key %llu %llu %llu %llu %llu %llu value %llu", v->
key[0],
80 #if defined (CLIB_HAVE_VEC512) 81 u64x8 v = u64x8_load_unaligned (a) ^ u64x8_load_unaligned (b);
82 return (u64x8_is_zero_mask (v) & 0x3f) == 0;
83 #elif defined (CLIB_HAVE_VEC256) 85 v = u64x4_insert_lo (v, u64x2_load_unaligned (a + 4) ^
86 u64x2_load_unaligned (b + 4));
87 v |= u64x4_load_unaligned (a) ^ u64x4_load_unaligned (b);
88 return u64x4_is_all_zero (v);
89 #elif defined(CLIB_HAVE_VEC128) && defined(CLIB_HAVE_VEC128_UNALIGNED_LOAD_STORE) 91 v = u64x2_load_unaligned (a) ^ u64x2_load_unaligned (b);
92 v |= u64x2_load_unaligned (a + 2) ^ u64x2_load_unaligned (b + 2);
93 v |= u64x2_load_unaligned (a + 4) ^ u64x2_load_unaligned (b + 4);
94 return u64x2_is_all_zero (v);
96 return ((a[0] ^ b[0]) | (a[1] ^ b[1]) | (a[2] ^ b[2]) | (a[3] ^ b[3])
97 | (a[4] ^ b[4]) | (a[5] ^ b[5])) == 0;
101 #undef __included_bihash_template_h__ static u64 clib_bihash_hash_48_8(const clib_bihash_kv_48_8_t *v)
Fixed length block allocator.
static int clib_bihash_key_compare_48_8(u64 *a, u64 *b)
static u64 clib_xxhash(u64 key)
static u8 * format_bihash_kvp_48_8(u8 *s, va_list *args)
static int clib_bihash_is_free_48_8(const clib_bihash_kv_48_8_t *v)
epu8_epi32 epu16_epi32 u64x2