As a programmer, you use hash functions every day. They’re used in databases
to optimise queries, they’re used in data structures to make things faster,
they’re used in security to keep data safe. Almost every interaction you have
with technology will involve hash functions in one way or another.

Hash functions are foundational, and they are everywhere.

But what is a hash function, and how do they work?

In this post, we’re going to demystify hash functions. We’re going to start by
looking at a simple hash function, then we’re going to learn how to test if a
hash function is good or not, and then we’re going to look at a real-world use
of hash functions: the hash map.

This article has visualisations that can be clicked.

What is a hash function?

Hash functions are functions that take an input, usually a string, and produce a
number. If you were to call a hash function multiple times with the same input,
it will always return the same number, and that number returned will always be
within a promised range. What that range is will depend on the hash function,
some use 32-bit integers (so 0 to 4 billion), others go much larger.

If we were to write a dummy hash function in JavaScript, it might look like

function hash(input) {
  return 0;

Even without knowing how hash functions are used, it’s probably no surprise
that this hash function is useless. Let’s see how we can measure how good a
hash function is, and after that we’ll do a deep dive on how they’re used
within hash maps.

What makes a hash function good?

Because input can be any string, but the number returned is within some
promised range, it’s possible that two different inputs can return the same
number. This is called a “collision,” and good hash functions try to minimise
how many collisions they produce.

It’s not possible to completely eliminate collisions, though. If we wrote a hash
function that returned a number in the range 0 to 7, and we gave it 9 unique
inputs, we’re guaranteed at least 1 collision.

hash("to") == 3
hash("the") == 2
hash("café") == 0
hash("de") == 6
hash("versailles") == 4
hash("for") == 5
hash("coffee") == 0
hash("we") == 7
hash("had") == 1

Output values from a well-known hash function, modulo 8. No matter what 9
values we pass, there are only 8 unique numbers and so collisions are
inevitable. The goal is to have as few as possible.

To visualise collisions, I’m going to use a grid. Each square of the grid is
going to represent a number output by a hash function. Here’s an example 8×2
grid. Click on the grid to increment the
example hash output value and see how we map it to a grid square. See what
happens when you get a number larger than the number of grid squares.

13 % 16 == 13
















Every time we hash a value, we’re going to make its corresponding square on the
grid a bit darker. The idea is to create an easy way to see how well a hash
function avoids collisions. What we’re looking for is a nice, even distribution.
We’ll know that the hash function isn’t good if we have clumps or patterns of
dark squares.

You said that when a hash function outputs the same value for 2 different
inputs, that’s a collision. But if we have a hash function that outputs values
in a big range, and we mapped those to a small grid, aren’t we going to create
lots of collisions on the grid that aren’t actually collisions? On our 8×2
grid, 1 and 17 both map to the 2nd square.

This is a great observation. You’re absolutely right, we’re going to be creating
“pseudo-collisions” on our grid. It’s okay, though, because if the hash function
is good we will still see an even distribution. Incrementing every square by 100
is just as good a distribution as incrementing every square by 1. If we have a
bad hash function that collides a lot, that will still stand out. We’ll see
this shortly.

Let’s take a larger grid and hash 1,000 randomly-generated strings. You can
click on the grid to hash a new set of random
inputs, and the grid will animate to show you each input being hashed and placed
on the grid.

The values are nice and evenly distributed because we’re using a good,
well-known hash function called murmur3. This hash
is widely used in the real-world because it has great distribution while also
being really, really fast.

What would our grid look like if we used a bad hash function?

function hash(input) {
  let hash = 0;
  for (let c of input) {
    hash += c.charCodeAt(0);
  return hash % 1000000;

This hash function loops through the string that we’re given and sums the
numeric values of each character. It then makes sure that the value is between 0
and 1000000 by using the modulus operator (%). Let’s call this hash function

Here it is on the grid. Reminder, this is 1,000 randomly generated strings that
we’re hashing.

This doesn’t look all that different from murmur3.
What gives?

The problem is that the strings we’re giving to be hashed are random. Let’s see
how each function performs when given input that is not random: the numbers from
1 to 1000 converted to strings.

Now the problem is more clear. When the input isn’t random, the output of stringSum forms a pattern. Our murmur3 grid, however, looks the same as how it looked with
random values.

How about if we hash the top 1,000 most common English words:

It’s more subtle, but we do see a pattern on the stringSum grid. As usual, murmur3
looks the same as it always does.

This is the power of a good hash function: no matter the input,
the output is evenly distributed. Let’s talk about one more way to visualise
this and then talk about why it matters.

The avalanche effect

Another way hash functions get evaluated is on something called the “avalanche
effect.” This refers to how many bits in the output value change when just a
single bit of the input changes. To say that a hash function has a good
avalanche effect, a single bit flip in the input should result in an average of
50% the output bits flipping.

It’s this property that helps hash functions avoid forming patterns in the grid.
If small changes in the input result in small changes in the output, you get
patterns. Patterns indicate poor distribution, and a higher rate of collisions.

Below, we are visualising the avalanche effect by showing two 8-bit binary
numbers. The top number is the input value, and the bottom number is the murmur3 output value. Click on it to flip a single bit
in the input. Bits that change in the output will be green, bits that stay the same will be red.

murmur3 does well, though you will notice that
sometimes fewer than 50% of the bits flip and sometimes more. This is okay,
provided that it is 50% on average.

Let’s see how stringSum performs.

Well this is embarassing. The output is equal to the input, and so only a single
bit flips each time. This does make sense, because stringSum just sums the numeric value of each character in the
string. This example only hashes the equivalent of a single character, which
means the output will always be the same as the input.

Why all of this matters

We’ve taken the time to understand some of the ways to determine if a hash
function is good, but we’ve not spent any time talking about why it matters.
Let’s fix that by talking about hash maps.

To understand hash maps, we first must understand what a map is. A map is a data
structure that allows you to store key-value pairs. Here’s an example in

let map = new Map();
map.set("hello", "world");

Here we take a key-value pair ("hello""world") and store it in the map.
Then we print out the value associated with the key "hello", which will be

A more fun real-world use-case would be to find anagrams. An anagram is when two
different words contain the same letters, for example “antlers” and “rentals”
or “article” and “recital.” If you have a list of words and you want to find
all of the anagrams, you can sort the letters in each word alphabetically and
use that as a key in a map.

let words = [

let map = new Map();

for (let word of words) {
  let key = word

  if (!map.has(key)) {
    map.set(key, []);

This code results in a map with the following structure:

  "aelnrst": [
  "aceilrt": [
  "aabflmnoty": [

Implementing our own simple hash map

Hash maps are one of many map implementations, and there are many ways to
implement hash maps. The simplest way, and the way we’re going to demonstrate,
is to use a list of lists. The inner lists are often referred to as “buckets” in
the real-world, so that’s what we’ll call them here. A hash function is used on
the key to determine which bucket to store the key-value pair in, then the
key-value pair is added to that bucket.

Let’s walk through a simple hash map implementation in JavaScript. We’re going
to go through it bottom-up, so we’ll see some utility methods before getting to
the set and get implementations.

class HashMap {
  constructor() { = [[], [], []];

We start off by creating a HashMap class with a constructor that sets up 3
buckets. We use 3 buckets and the short variable name bs so that this code
displays nicely on devices with smaller screens. In reality, you could have
however many buckets you want (and better variable names).

class HashMap {
  // ...
  bucket(key) {
    let h = murmur3(key);
      h %

The bucket method uses murmur3 on the key
passed in to find a bucket to use. This is the only place in our hash map code
that a hash function is used.

class HashMap {
  // ...
  entry(bucket, key) {
    for (let e of bucket) {
      if (e.key === key) {
        return e;
    return null;

The entry method takes a bucket and a key and scans the bucket until it
finds an entry with the given key. If no entry is found, null is returned.

class HashMap {
  // ...
  set(key, value) {
    let b = this.bucket(key);
    let e = this.entry(b, key);
    if (e) {
      e.value = value;
    b.push({ key, value });

The set method is the first one we should recognise from our earlier
JavaScript Map examples. It takes a key-value pair and stores it in our hash
map. It does this by using the bucket and entry methods we created earlier.
If an entry is found, its value is overwritten. If no entry is found, the
key-value pair is added to the map. In JavaScript, { key, value } is
shorthand for { key: key, value: value }.

class HashMap {
  // ...
  get(key) {
    let b = this.bucket(key);
    let e = this.entry(b, key);
    if (e) {
      return e.value;
    return null;

The get method is very similar to set. It uses bucket and entry to find
the entry related to the key passed in, just like set does. If an entry is
found, its value is returned. If one isn’t found, null is returned.

That was quite a lot of code. What you should take away from it is that our
hash map is a list of lists, and a hash function is used to know which of the
lists to store and retrieve a given key from.

Here’s a visual representation of this hash map in action. Click anywhere on the buckets to add a new key-value pair using our
set method. To keep the visualisation simple, if a bucket were to “overflow”,
the buckets are all reset.

Because we’re using murmur3 as our hash
function, you should see good distribution between the buckets. It’s expected
you’ll see some imbalance, but it should generally be quite even.

To get a value out of the hash map, we first hash the key to figure out which
bucket the value will be in. Then we have to compare the key we’re searching for
against all of the keys in the bucket.

It’s this search step that we minimise through hashing, and why murmur3 is optimised for speed. The faster the hash
function, the faster we find the right bucket to search, the faster our hash
map is overall.

This is also why reducing collisions is so crucial. If we did decide to use that
dummy hash function from all the way at the start of this article, the one that
returns 0 all the time, we’ll put all of our key-value pairs into the first
bucket. Finding anything could mean we have to check all of the values in the
hash map. With a good hash function, with good distribution, we reduce the
amount of searching we have to do to 1/N, where N is the number of buckets.

Let’s see how stringSum does.

Interestingly, stringSum seems to distribute
values quite well. You notice a pattern, but the overall distribution looks

Finally! A win for stringSum. I knew it would
be good for something.

Not so fast, Haskie. We need to talk about a serious problem. The distribution
looks okay on these sequential numbers, but we’ve seen that stringSum doesn’t have a good avalanche effect. This doesn’t end

Real-world collisions

Let’s look at 2 real-world data sets: IP addresses and English words. What I’m
going to do is take 100,000,000 random IP addresses and 466,550 English
, hash all of them with both murmur3
and stringSum, and see how many collisions we get.

IP Addresses

murmur3 stringSum
Collisions 1,156,959 99,999,566
1.157% 99.999%

English words

murmur3 stringSum
Collisions 25 464,220
0.005% 99.5%

When we use hash maps for real, we aren’t usually storing random values in them.
We can imagine counting the number of times we’ve seen an IP address in rate
limiting code for a server. Or code that counts the occurrences of words in
books throughout history to track their origin and popularity. stringSum sucks for these applications because of it’s extremely
high collision rate.

Manufactured collisions

Now it’s murmur3‘s turn for some bad news.
It’s not just collisions caused by similarity in the input we have to worry
about. Check this out.

What’s happening here? Why do all of these jibberish strings hash to the same

I hashed 141 trillion random strings to find values that hash to the number
1228476406 when using murmur3. Hash functions
have to always return the same output for a specific input, so it’s possible to
find collisions by brute force.

I’m sorry, 141 trillion? Like… 141 and then 12 zeroes?

Yes, and it only took me 25 minutes. Computers are fast.

Bad actors having easy access to collisions can be devastating if your software
builds hash maps out of user input. Take HTTP headers, for example. An HTTP
request looks like this:

GET / HTTP/1.1
Accept: */*
Accept-Encoding: gzip, deflate
Connection: keep-alive

You don’t have to understand all of the words, just that the first line is the
path being requested and all of the other lines are headers. Headers are Key: Value pairs, so HTTP servers tend to use maps to store them. Nothing stops us
from passing any headers we want, so we can be really mean and pass headers we
know will cause collisions. This can significantly slow down the server.

This isn’t
theoretical, either
If you search “HashDoS” you’ll find a lot more examples of this. It was a really
big deal in the mid-2000s.

There are a few ways to mitigate this specific to HTTP servers: ignoring
jibberish header keys and limiting the number of headers you store, for
example. But modern hash functions like murmur3
offer a more generalised solution: randomisation.

Earlier in this post we showed some examples of hash function implementations.
Those implementations took a single argument: input. Lots of modern hash
functions take a 2nd parameter: seed (sometimes called salt). In the case
of murmur3, this seed is a number.

So far, we’ve been using 0 as the seed. Let’s see what happens with the
collisions I’ve collected when we use a seed of 1.

Just like that, 0 to 1, the collisions are gone. This is the purpose of the
seed: it randomises the output of the hash function in an unpredictable way.
How it achieves this is beyond the scope of this article, all hash functions do
this in their own way.

The hash function still returns the same output for the same input, it’s just
that the input is a combination of input and seed. Things that collide with
one seed shouldn’t collide when using another. Programming languages often
generate a random number to use as the seed when the process starts, so that
every time you run your program the seed is different. As a bad guy, not knowing
the seed, it is now impossible for me to reliably cause harm.

If you look closely in the above visualisation and the one before it, they’re
the same values being hashed but they produce different hash values. The
implication of this is that if you hash a value with one seed, and want to be
able to compare against it in the future, you need to make sure you use the same

Having different values for different seeds doesn’t affect the hash map
use-case, because hash maps only live for the duration the program is running.
Provided you use the same seed for the lifetime of the program, your hash maps
will continue to work just fine. If you ever store hash values outside of your
program, in a file for example, you need to be careful you know what seed has
been used.


As is tradition, I’ve made a playground for you to write your own hash functions
and see them visualised with the grids seen in this article. Click
here to try it!


We’ve covered what a hash function is, some ways to measure how good it is,
what happens when it’s not good, and some of the ways they can be broken by
bad actors.

The universe of hash functions is a large one, and we’ve really only scratched
the surface in this post. We haven’t spoken about cryptographic vs
non-cryptographic hashing, we’ve touched on only 1 of the thousands of use-cases
for hash functions, and we haven’t talked about how exactly modern hash
functions actually work.

Some further reading I recommend if you’re really enthusiastic about this topic
and want to learn more:

  • this repository is the gold standard
    for testing how good hash functions are. They run a tonne of tests against
    a wide number of hash functions and present the results in a big table. It
    will be difficult to understand what all of the tests are for, but this is
    where the state of the art of hash testing lives.
  • this is an interactive piece
    on a data structure called HyperLogLog. It’s used to efficiently count the
    number of unique elements in very, very large sets. It uses
    hashing to do it in a really clever way.
  • is a piece of software that, when given
    the expected set of things you want to hash, can generate a “perfect” hash
    function automatically.

Feel free to join the discussion on Hacker News!


Thanks to everyone who read early drafts and provided invaluable feedback.

And everyone who helped me find murmur3 hash


After the success of Load Balancing and Memory
, I have decided to set up a Patreon page: For all of these articles going forward, I am
going to post a Patreon-exclusive behind-the-scenes post talking about
decisions, difficulties, and lessons learned from each post. It will give you
a deep look in to how these articles evolve, and I’m really stoked about
the one I’ve written for this one.

If you enjoy my writing, and want to support it going forward, I’d really
appreciate you becoming a Patreon. ❤️

Read More