This publish continues Behind the Faucet, a collection exploring the hidden mechanics of on a regular basis tech — from Uber to Spotify to search engines like google and yahoo. I’ll dive underneath the hood to demystify the programs shaping your digital world.
first relationship with music listening began at 6, rotating by means of the albums in the lounge’s Onkyo 6-disc participant. Cat Stevens, Groove Armada, Sade. There was all the time one music I saved rewinding to, although I didn’t know its identify. 10 years on, moments of the music returned to reminiscence. I searched by means of boards, ‘previous saxophone melody’, ‘classic music about sand dunes’, in search of years with no success. Then, in the future at college, I used to be in my pal Pegler’s dorm room when he performed it:
That lengthy search taught me how necessary it’s to have the ability to discover the music you like.
Earlier than streaming and good assistants, music discovery relied on reminiscence, luck, or a pal with good music style. That one catchy refrain might be misplaced to the ether.
Then got here a music-lover’s miracle.
A number of seconds of sound. A button press. And a reputation in your display screen.
Shazam made music recognisable.
The Origin: 2580
Shazam launched in 2002, lengthy earlier than apps had been a factor. Again then it labored like this:
You’d dial 2580# in your cellular (UK solely).
Maintain your cellphone as much as the speaker.
…Wait in silence…
And obtain a SMS telling you the identify of the music.
It felt like magic. The founding group, Chris Barton, Philip Inghelbrecht, Avery Wang, and Dhiraj Mukherjee, spent years constructing that phantasm.
To construct its first database, Shazam hired 30 young workers to run 18-hour shifts, manually loading 100,000 CDs into computer systems and utilizing customized software program. As a result of CD’s don’t comprise metadata they needed to sort the names of the songs manually, referring to the CD sleeve, to finally create the corporate’s first million audio fingerprints — a painstaking course of that took months.
In an period earlier than smartphones or apps, when Nokia’s and Blackberry’s couldn’t deal with the processing or reminiscence calls for, Shazam needed to keep alive lengthy sufficient for the know-how to catch as much as their concept. This was a lesson in market timing.
This publish is about what occurs within the second between the faucet and the title, the sign processing, hashing, indexing, and sample matching that lets Shazam hear what you possibly can’t fairly identify.
The Algorithm: Audio Fingerprinting
In 2003, Shazam co-founder Avery Wang published the blueprint for an algorithm that also powers the app as we speak. The paper’s central concept: If people can perceive music by superimposing layers of sound, a machine may do it too.
Let’s stroll by means of how Shazam breaks sound all the way down to one thing a machine can recognise immediately.
1. Capturing Audio Pattern
It begins with a faucet.
If you hit the Shazam button, the app data a 5–10 second snippet of the audio round you. That is lengthy sufficient to determine most songs, although we’ve all waited minutes holding our telephones within the air (or hiding in our pockets) for the ID.
However Shazam doesn’t retailer that recording. As a substitute, it reduces it to one thing far smaller and smarter: a fingerprint.
2. Producing the Spectrogram
Earlier than Shazam can recognise a music, it wants to grasp what frequencies are within the sound and once they happen. To do that, it makes use of a mathematical instrument referred to as the Fast Fourier Transform (FFT).
The FFT breaks an audio sign into its element frequencies, revealing which notes or tones make up the sound at any second.
Why it issues: Waveforms are fragile, delicate to noise, pitch modifications, and system compression. However frequency relationships over time stay steady. That’s the gold.
In the event you studied Arithmetic at Uni, you’ll keep in mind the struggles of studying the Discrete Fourier Transform process.Quick Fourier Rework (FFT) is a extra environment friendly model that lets us decompose a fancy sign into its frequency parts, like listening to all of the notes in a chord.
Music isn’t static. Notes and harmonics change over time. So Shazam doesn’t simply run FFT as soon as, it runs it repeatedly over small, overlapping home windows of the sign. This course of is named the Brief-Time Fourier Rework (STFT) and types the idea of the spectrogram.
The ensuing spectrogram is a change of sound from the amplitude-time area (waveform) into the frequency-time area.
Consider this as turning a messy audio waveform right into a musical heatmap.
As a substitute of displaying how loud the sound is, a spectrogram reveals what frequencies are current at what instances.

A spectrogram strikes evaluation from the amplitude-time area to frequency-time area. It shows time on the horizontal axis, frequency on the vertical axis, and makes use of brightness to point the amplitude (or quantity) of every frequency at every second. This lets you see not simply which frequencies are current, but additionally how their depth evolves, making it potential to determine patterns, transient occasions, or modifications within the sign that aren’t seen in an ordinary time-domain waveform.
Spectrograms are extensively utilized in fields similar to audio evaluation, speech processing, seismology, and music, offering a strong instrument for understanding the temporal and spectral traits of alerts.
3. From Spectrogram to Constellation Map
Spectrograms are dense and comprise an excessive amount of information to check throughout thousands and thousands of songs. Shazam filters out low-intensity frequencies, leaving simply the loudest peaks.
This creates a constellation map, a visible scatterplot of standout frequencies over time, just like sheet music, though it jogs my memory of a mechanical music-box.

4. Creating the Audio Fingerprint
Now comes the magic, turning factors right into a signature.
Shazam takes every anchor level (a dominant peak) and pairs it with goal peaks in a small time window forward — forming a connection that encodes each frequency pair and timing distinction.
Every of those turns into a hash tuple:
(anchor_frequency, target_frequency, time_delta)

What’s a Hash?
A hash is the output of a mathematical operate, referred to as a hash operate, that transforms enter information right into a fixed-length string of numbers and/or characters. It’s a manner of turning complicated information into a brief, distinctive identifier.
Hashing is extensively utilized in laptop science and cryptography, particularly for duties like information lookup, verification, and indexing.

For Shazam, a typical hash is 32 bits lengthy, and it would possibly be structured like this:
- 10 bits for the anchor frequency
- 10 bits for the goal frequency
- 12 bits for the time delta between them

This tiny fingerprint captures the connection between two sound peaks and the way far aside they’re in time, and is powerful sufficient to determine the music and sufficiently small to transmit rapidly, even on low-bandwidth connections.
5. Matching In opposition to the Database
As soon as Shazam creates a fingerprint out of your snippet, it must rapidly discover a match in its database containing thousands and thousands of songs.
Though Shazam has no concept the place within the music your clip got here from — intro, verse, refrain, bridge — doesn’t matter, it appears to be like for relative timing between hash pairs. This makes the system strong to time offsets within the enter audio.

Shazam compares your recording’s hashes towards its database and identifies the music with the best variety of matches, the fingerprint that greatest traces up together with your pattern, even when it’s not an actual match resulting from background noise.
The way it Searches So Quick
To make this lightning-fast, Shazam makes use of a hashmap, an information construction that enables for near-instant lookup.
A hashmap can discover a match in O(1) time, meaning the lookup time stays fixed, even when there are thousands and thousands of entries.
In distinction, a sorted index (like B-tree on disk) takes O(log n) time, which grows slowly because the database grows.
This stability of time and house complexity is named Big O Notation, concept I’m not ready of bothered to show. Please seek advice from a Pc Scientist.
6. Scaling the System
To keep up this pace at world scale, Shazam does extra than simply use quick information buildings, it optimises how and the place the info lives:
- Shards the database — dividing it by time vary, hash prefix, or geography
- Retains scorching shards in reminiscence (RAM) for fast entry
- Offloads colder information to disk, which is slower however cheaper to retailer
- Distributes the system by area (e.g., US East, Europe, Asia ) so recognition is quick regardless of the place you might be
This design helps 23,000+ recognitions per minute, even at world scale.
Influence & Future Purposes
The plain utility is music discovery in your cellphone, however there may be one other main utility of Shazam’s course of.
Shazam facilitates Market Insights. Each time a person tags a music, Shazam collects anonymised, geo-temporal metadata (the place, when, and the way usually a music is being ID’d.)
Labels, artists, and promoters use this to:
- Spot breakout tracks earlier than they hit the charts.
- Establish regional traits (a remix gaining traction in Tokyo earlier than LA).
- Information advertising spend primarily based on natural attraction.
In contrast to Spotify, which makes use of person listening behaviour to refine suggestions, Shazam supplies real-time information on songs folks actively determine, providing the music trade early insights into rising traits and common tracks.
What Spotify Hears Before You Do
The Data Science of Music Recommendationmedium.com
On December 2017, Apple bought Shazam for a reported $400 million. Apple reportedly uses Shazam’s data to augment Apple Music’s recommendation engine, and record labels now monitor Shazam trends like they used to monitor radio spins.

Sooner or later, there may be anticipated evolution in areas like:
- Visual Shazam: Already piloted, level you digicam at an object or art work to determine it, helpful for an Augmented Actuality future.
- Live performance Mode: Establish songs stay throughout gigs and sync to a real-time setlist.
- Hyper-local trends: Floor what’s trending ‘on this avenue’ or ‘on this venue’, increasing community-shared music style.
- Generative AI integration: Pair audio snippets with lyric technology, remix recommendations, or visible accompaniment.
Outro: The Algorithm That Endures
In a world of ever-shifting tech stacks, it’s uncommon for an algorithm to remain related for over 20 years.
However Shazam’s fingerprinting technique hasn’t simply endured, it’s scaled, advanced, and turn into a blueprint for audio recognition programs throughout industries.
The magic isn’t simply that Shazam can identify a music. It’s the way it does it, turning messy sound into elegant math, and doing it reliably, immediately, and globally.
So subsequent time you’re in a loud, trashy bar holding your cellphone as much as the speaker taking part in Lola Younger’s ‘Messy’ simply keep in mind: behind that faucet is a good looking stack of sign processing, hashing, and search, designed so nicely it barely needed to change.