Previous strategy was to resize the hash table when the length of any
one bucket exceeded a certain size, with some logic for intelligently
stopping resizes when the gains from doing so weren't sufficient. While
this was a good idea that attempted to optimize both space and lookup
time, unfortunately under transient degenerate conditions this led to
some issues with the tables not resizing when they should have,
harming performance. The resizing restriction was lifted, but this had
the result of exacerbating degenerate behavior and caused out of memory
conditions.
This patch changes the hash expansion criterion to be based on the
number of elements in the table. Once the # of elements in the table
exceeds the number of buckets, the table size is doubled. While the
space efficiency of this method decreases relative to the perfectness of
the hash function, at least this strategy puts the table performance
squarely in the hands of the hash function.
Signed-off-by: Quentin Young <qlyoung@cumulusnetworks.com>
unsigned int key;
unsigned int index;
void *newdata;
- unsigned int len;
struct hash_backet *backet;
key = (*hash->hash_key)(data);
index = key & (hash->size - 1);
- len = 0;
for (backet = hash->index[index]; backet != NULL;
backet = backet->next) {
if (backet->key == key && (*hash->hash_cmp)(backet->data, data))
return backet->data;
- ++len;
}
if (alloc_func) {
if (newdata == NULL)
return NULL;
- if (len > HASH_THRESHOLD) {
+ if (HASH_THRESHOLD(hash->count + 1, hash->size)) {
hash_expand(hash);
index = key & (hash->size - 1);
}
DECLARE_MTYPE(HASH_BACKET)
/* Default hash table size. */
-#define HASH_INITIAL_SIZE 256 /* initial number of backets. */
-#define HASH_THRESHOLD 10 /* expand when backet. */
+#define HASH_INITIAL_SIZE 256
+/* Expansion threshold */
+#define HASH_THRESHOLD(used, size) ((used) > (size))
#define HASHWALK_CONTINUE 0
#define HASHWALK_ABORT -1