Hash Table Internals - Part 11 - Implementing Hash Maps
A map, when implemented with Hash Tables, is called a Hash Map; and just like a regular map it supports operations like put, get, and iterate.
Key Implementation Details
Hash Maps are required to store the application keys of any type, but the Hash Table understands only integers. We first pass the application keys through a hash function to map it to a 32-bit integer, and then the usual Hash Table operation takes over.
Given that the application key to hash key is a frequent operation, we would keep it handy by storing it alongside the application key. This helps us save repeated computations.
While looking up a key in the hash map, we first reach the slot, and then compare if the key present there is really the key we are looking for, as we cannot just rely on the equality of the hash keys. The Hash Map, hence, accepts a key comparator function.
With Chained Hash Tables
Each node of the linked list in the chained hash table has the following structure
struct node {
int32 hash_key;
void *key;
void *value;
struct node *next;
}
void * key and void * value allows us to hold a key and a value of any type, while int32 hash_key enables us to hold a pre-computed hash.
At the Hash Table level, we hold
Instead of having a contains function, we have a lookup function that returns the value for the matching key, and NULL if the key does not exist. Thus, the lookup function doubles as a contains function.
When duplicate keys are inserted, we can do one of the following
With Hash Tables having open addressing
Each slot of the hash table has the following structure
struct node {
int32 hash_key;
void *key;
void *value;
bool is_empty;
bool is_deleted;
}
void *key and void *value allows us to hold a key and a value of any type, while int32 hash_key enables us to hold a pre-computed hash, saving a lot of runtime computations. is_empty tells us if the slot is empty, while is_deleted represents a soft deleted slot.
At the Hash Table level, we hold
Here the load factor will be computed as number of used slots/size of the array because the soft deleted keys also affect the performance of the hash table.
During insert, lookup, and delete when we find the matching hash, and explicitly compare the application keys, as multiple keys can hash to the same location, and we cannot rely on just hash key equivalence.
Here's the video of my explaining this in-depth ?? do check it out
Thank you so much for reading ?? If you found this helpful, do spread the word about it on social media; it would mean the world to me.
If you liked this short essay, you might also like my courses on
I teach an interactive course on System Design where you'll learn how to intuitively design scalable systems. The course will help you
I have compressed my ~10 years of work experience into this course, and aim to accelerate your engineering growth 100x. To date, the course is trusted by 800+ engineers from 11 different countries and here you can find what they say about the course.
Together, we will dissect and build some amazing systems and understand the intricate details. You can find the week-by-week curriculum and topics, testimonials, and other information at https://arpitbhayani.me/masterclass.
Senior Manager at Ciena
2 年Can you make video on Bloom filters, and why bloom filters are recommended for lookup than using a traditional Hash table/map ?
SDE @ Jio Plaforms | 2+ years exp | Backend Development | Node.js | Java | API Design & Security | Microservices & REST APIs | System Integration | Problem Solving |
2 年#AsliEngineers :)
More about me: arpitbhayani.me Newsletter: arpitbhayani.me/newsletter Subscribe #AsliEngineering for such in-depth engineering concepts: https://www.youtube.com/c/ArpitBhayani System Design course: arpitbhayani.me/masterclass Microservices: https://courses.arpitbhayani.me/designing-microservices All GitHub Outages: https://courses.arpitbhayani.me/github-outage-dissections/