Module 2.3 · Foundation Track · How You Hear
How You Hear Stereo
Stereo isn't a setting on your gear. It's a 3D world you step into — where every instrument, every voice, every breath finds its place around you.
Put on a good pair of headphones. Close your eyes. Press play on a record you love. Now notice: you're not listening to the music — you're inside it. The vocal sits a few inches from your face. The drums are behind your eyes. A shaker rattles past your right shoulder; a guitar pad expands like fog around the room. That sensation, of being placed inside a sonic space, is what stereo actually is. Your ears are receiving two slightly different versions of a recording, and your brain is reconstructing a 3D world from the difference.
Mono is one ear of information. Stereo is two ears of information, plus the physical fact that your brain evolved to figure out where sounds come from in 3D space — through millions of years of "is that a tiger or just leaves?" — using the tiny differences between what your left and right ears pick up. When recording engineers send you a stereo mix, they're hijacking that ancient location-finding circuitry. They're putting elements somewhere in your auditory world: this guitar to the left, that vocal harmony just behind the lead, the kick anchored dead center, the cymbals expanding outward into the air around you.
This is why the best mixes feel spatial and not just loud. Every instrument has a place. Nothing is fighting for the same square inch. Nothing is invisible. As a mixer, you're an architect of an invisible 3D room, and your tools — panning, delay, reverb, stereo widening, mid-side processing — are how you build the rooms walls, doors, and windows. The widget below introduces the two cues your brain uses to construct that space: level differences between L and R, and timing differences between L and R. Drag the source, switch modes, hit Sum to mono. Your headphones will reveal what's possible — and what costs come with each technique.
Tap to start
Headphones strongly recommended
This widget teaches stereo perception by playing pink noise through different L/R cues. The lesson lives in the difference between left and right — laptop speakers and Bluetooth speakers blur that difference. Use headphones or proper stereo monitors for the demo to work.
Foundation · Module 2.3 · How You Hear
How You Hear Space
Drag the source. The brain decodes its position from L/R level differences AND L/R timing — and the two cues behave very differently in mono.
Level Panning · Center
Drag the source left or right. In Level Panning mode, the L and R speakers play the same sound at different volumes — the louder side feels closer. This is the standard "pan pot" most DAWs use. Both ears hear the source at the same time; only the level differs.
Going deeper
Two ears, one world
You have two ears, spaced about 17 cm apart on opposite sides of a head that blocks high frequencies. When a sound comes from your left, it reaches your left ear first (a few hundred microseconds before the right) and at a slightly higher level (because your head shadows the high frequencies on the far side). Your brain has been decoding these tiny differences since you were a fetus listening to your mother's heartbeat. It's not a learned skill; it's hardware.
Stereo recording exploits this. By giving your left ear and your right ear slightly different signals, engineers can place sounds anywhere in space — in front, beside, even slightly behind — without you ever moving. The widget you just used shows the two cues at work in isolation, but real records use both at once, and they layer reverb, delay, and pan on top. Every element has a place.
The two cues your brain uses
Lord Rayleigh, in 1907, formalized what's now called the Duplex Theory of localization. He proposed that the brain uses two distinct mechanisms for placing a sound in space:
- Interaural Level Difference (ILD) — the higher frequencies are louder on the closer ear. Your head shadows the far side. The ear-to-ear level difference is the cue your brain uses for high-frequency sources.
- Interaural Time Difference (ITD) — sound waves reach the closer ear earlier. For low frequencies (below ~1.5 kHz), the wavelengths are bigger than your head, so level differences are tiny — but timing differences (microseconds) are perceptible. The ear-to-ear time difference is the cue your brain uses for low-frequency sources.
In real life, both cues fire simultaneously — that's the "Combined" mode in the widget. Recording engineers can choose to use only ILD (Level Panning, the standard pan pot in every DAW), only ITD (Haas / Time Panning, used for widening tricks), or both (which sounds most natural). Each technique has trade-offs, especially in mono.
Helmut Haas and the precedence effect
In 1949, German researcher Helmut Haas published a doctoral dissertation that established what we now call the Haas effect (also "the precedence effect" or "the law of the first wavefront"). He found that when two identical sounds arrive at the ears with a delay between 1 and 35 milliseconds, the brain perceives them as a single sound located on the side of the earlier one — even when both are at the same level.
Haas's research had practical consequences for live sound: PA systems with delayed speakers can reinforce volume across a large venue without splitting the source's apparent location. Mix engineers later borrowed the principle for stereo widening — duplicate a mono track, pan hard L and R, delay one side ~15-25 ms, and the source "becomes wide" with a single performance. The Haas trick.
The catch is in mono. Two identical signals offset by 15 ms create comb filtering: certain frequencies (multiples of 1/2t Hz, where t is the time delta) cancel, others reinforce. The notch frequencies for a 15 ms delay land at 33 Hz, 100 Hz, 167 Hz, 233 Hz, 300 Hz... a series spaced 67 Hz apart, getting denser as you go up. That's why Haas-widened sources sound thin when summed to mono — the low end is hollow and the mids feel swooshy.
"The greatest mixes don't ask you to listen — they put you inside something. Stereo is the door." — FTM, paraphrasing the spirit of Bob Clearmountain interviews
Why headphones reveal what speakers can hide
On speakers, the L and R signals leak across to both ears (the L speaker reaches your right ear too, just delayed by ~0.3 ms and slightly attenuated by your head). This "crosstalk" partially blends the stereo image — speakers reveal less of the L/R difference than headphones do. Headphones, by contrast, deliver the L signal only to your left ear and the R signal only to your right. The stereo cues land cleanly. That's why this widget begs for headphones — the lesson is in the difference, and headphones don't blur it.
It's also why headphone-mixed tracks sometimes don't translate to speakers: an engineer who only checks on headphones may push the stereo image too wide because the headphone delivery is so clean, and on speakers the image either crowds or collapses. The reverse can happen too — a mix that sounds "narrow" on headphones may feel right on monitors. Always check both.
Beyond stereo: HRTF and binaural
Your outer ears (the pinna) are weirdly shaped, and that's intentional. The folds and ridges filter sound differently depending on the direction it comes from — front, back, above, below. Combined with the head shadow and ear-to-ear time differences, this gives your brain enough information to localize sound in three dimensions, not just left and right.
Audio engineers can now simulate these directional filters digitally — a process called binaural rendering, based on the Head-Related Transfer Function (HRTF). When you hear a "binaural" recording on headphones, sounds appear in front of you, behind you, even above. Apple's Spatial Audio, Dolby Atmos for headphones, and game-audio engines all use HRTF-based processing to create immersive 3D sound on standard stereo headphones. We won't go deep on this in the Foundation track, but know that the principles you're learning here — L/R level, L/R time — are the foundation. HRTF is just stereo with directional filters added.
Mono compatibility: the elephant in every mix
Roughly half of all music listening happens in mono. Phone speakers (most phones have a single forward-facing speaker). Single Bluetooth speakers. Kitchen radios. Voice-assistant devices. Club PAs with a malfunctioning side. Even some "stereo" earbuds collapse to mono when one bud loses connection. If your mix sounds great in stereo but falls apart in mono, half your listeners will hear the broken version.
This is why every working mix engineer checks mono regularly during a mix. The check reveals:
- Phase issues — out-of-polarity stereo content disappears when summed.
- Haas-induced comb filtering — wide sources go thin.
- Stereo widener overuse — bass drops out, vocal sounds hollow.
- Mic-pair phase problems — overheads, room mics, snare top/bottom can fight in mono.
The fix isn't to mix in mono — that loses the spatial richness — but to periodically check mono. Insert a mono utility on the master bus, listen for 30 seconds, fix any obvious problems, then return to stereo. Tchad Blake mixes in mono more than half the time as a discipline; many other engineers mix in stereo and check mono every 10-15 minutes. Either approach works; what matters is the habit.
The art of placement
Once you understand the mechanics — level cues, time cues, mono caveats — the creative work begins: placing every instrument in the 3D world. There's no universal rule book, but there are widely-shared instincts:
- The kick, snare, bass, and lead vocal sit dead center. They carry the most energy, and centered elements are mono-safe.
- Hi-hat and overheads spread wide. Cymbals provide stereo width.
- Doubled guitars or BGV stacks pan hard left and right. Each take is its own performance, so even in mono they sum constructively (no Haas comb filter).
- Keys, pads, and ambient elements fill the middle ground. Sometimes wide stereo, sometimes lightly off-center.
- Lead instruments like a guitar solo or featured horn often pan slightly off-center — say L 30 — to give the lead vocal its own central pocket.
- Reverbs and ambience returns spread wide to create depth.
- FX sends, ear candy, ad libs, occasional details go anywhere. The "stereo seasoning" that keeps a mix interesting.
We'll cover specific placement decisions in detail in Module 9 — Mixing Vocals, Module 10 — Mixing Drums, and the rest of Track 2. For now, just internalize: every instrument has a place. The mix is a 3D world. You're the architect.
In your DAW — stereo tools you'll actually use
| DAW | Pan / Stereo tools | Mono check |
|---|---|---|
| GarageBand | Track pan knob (constant-power). Limited stereo manipulation beyond panning. | Master output → set channel mode to "Mono" temporarily, or use a third-party mono utility plugin. |
| Logic Pro | Track pan knob. Direction Mixer for stereo width adjustment. Stereo Spread for frequency-dependent widening. Mid/Side EQ via Channel EQ in M/S mode. | Click the master channel's stereo→mono toggle (the small icon near the master fader), or insert Direction Mixer with width = 0%. |
| Ableton Live | Track pan knob. Utility plugin: Width slider (0% = mono, 100% = full stereo, 200% = exaggerated). Spread on Audio Effects for stereo widening with frequency tilt. | Insert Utility on master bus → set Width = 0%. Or use the Audio Effect "Mono". |
| Pro Tools | Pan knob (stereo channels have two pans for L/R). Trim plugin for level adjustments. S1 stereo image processor (Pro Tools HD). | Insert Trim plugin on master, click the M/S button. Or use a third-party mono utility (Voxengo MSED is free). |
| Reaper | Pan knob with multiple modes (constant-power, balance, dual-pan). JS plugins for stereo width manipulation. | Master track → insert JS: Channel Tools → set pan law. Or use ReaPlugs ReaJS. |
| FL Studio | Pan knob. Fruity Stereo Shaper (free with FL) for width manipulation. Fruity Stereo Enhancer for M/S processing. | Master → insert Fruity Stereo Shaper, set Stereo Separation to 0%, or use Patcher-based mono utility. |
Free mono-checking tools across all DAWs: Voxengo MSED (free, M/S encoder/decoder, can listen to mid or side in solo, sum to mono with one click), Voxengo Correlometer (free, shows stereo correlation in real time — anything below 0 is a phase problem). Drop one of these on your master bus and you've got pro-grade mono compatibility checking for free.
Common mistakes
- Pulling everything to center. A mix with no stereo decisions sounds flat and "phone-speaker-shaped." Even a centered kick, snare, bass, and lead vocal benefit from the rest of the band having stereo placement.
- Pulling everything wide. If every element is at hard L or hard R, there's no center, no anchor — and lead vocals get lost. Save the wide positions for atmospheric and supporting elements.
- Haas-widening everything. Seductive in solo on headphones, but stack 5+ Haas-widened sources and the mono sum becomes a comb-filter mess. Use Haas sparingly — one or two elements per mix.
- Never checking mono. 50% of listening is mono. If you don't check, you don't know. Build it into your workflow: every 10-15 minutes, hit mono for 30 seconds.
- Mixing only on headphones for stereo decisions. Headphones over-emphasize the L/R difference (no crosstalk). What sounds "wide enough" on speakers may feel claustrophobic on headphones, and vice versa. Reference both.
- Trusting visual width over audible width. Some plugins offer "stereo widener" controls that look impressive on a meter but introduce phase issues you don't hear until it's too late. Check mono. Check correlation. Trust your ears.
Next up · Module 2.4
Critical Bands & Masking