Skip to content
Huberman LabHuberman Lab

Dr. Jennifer Groh on Huberman Lab: How Gaze Reshapes Hearing

Groh shows thoughts are multi-sensory simulations running in sensory cortex; the superior colliculus links gaze direction to sound maps, altering what you hear.

Dr. Jennifer GrohguestAndrew Hubermanhost
Nov 9, 20252h 16mWatch on YouTube ↗

At a glance

WHAT IT’S REALLY ABOUT

Shape Your Thoughts: Sensory Simulations, Attention, and Brain Design Revealed

  1. Neuroscientist Dr. Jennifer Groh explains how the brain builds thoughts by running multi-sensory simulations—using the same visual, auditory, and motor circuits we use to perceive and act in the world.
  2. She describes how the brain integrates sight and sound for spatial awareness, why eye movements literally change what we hear (down to the eardrum), and how environments and devices like phones hijack or support our attention.
  3. The conversation explores music, rhythm, and emotion, the physics and biology of sound localization, and why focus behaves more like interval training or endurance sport than a simple on/off switch.
  4. Throughout, Groh and Huberman connect lab findings to practical tools for better focus, learning, and self-management of brain states, including visual techniques, environment design, and realistic expectations about concentration.

IDEAS WORTH REMEMBERING

5 ideas

Thoughts Are Built as Multi-Sensory Simulations in Sensory Cortex

Groh proposes that thinking is the brain running internal simulations in the same sensory and motor circuits used for perception and action. For example, thinking “cat” likely triggers visual cortex activity (appearance), auditory cortex (meow, purr), and even smell or tactile representations. This explains why we can instantly specify a cat’s color or smell when asked, and why competing sensory tasks (like conversation) can impair visually demanding behavior (like merging in heavy traffic). Practically, recognizing thought as simulation suggests that controlling sensory input—what you see, hear, and feel—is a direct lever on what and how you think.

Eye Position Literally Changes What You Hear, Starting at the Eardrum

Classically, audio processing was thought to flow bottom-up from the ear. Groh’s work shows that eye position information is injected very early into the auditory system: auditory neurons in the superior colliculus change responses depending on where the eyes are pointed, and even the eardrum moves in a precise pattern with each saccade. When you move your eyes left or right, top-down signals from the brain drive tiny, anti-phase eardrum motions in the two ears, encoding eye movement amplitude and direction. This suggests that spatial alignment of sight and sound begins at or near the periphery, not just in “high” brain centers.

Sound Localization Relies on Microsecond Timing and Ear-Specific Filtering

To locate sound in space, the brain exploits tiny differences in arrival time and loudness between the two ears—up to only about 0.5 milliseconds for sounds from the left vs. right. This is shorter than a single neural spike, so specialized pathways with highly precise synapses and population coding are required. Outer ear folds also filter frequencies differently depending on sound direction, creating an individual ‘acoustic fingerprint’ you learn over development (and must relearn as your head grows, or if your ear anatomy changes). This has implications for hearing loss in one ear, for the design of 3D audio, and for understanding why some spatial audio tricks (like ventriloquism) work so well.

Perception of Sound Sources Is Highly Constructed and Often Misleading

The brain routinely overrides raw ear input to create coherent sound sources. In movies, dialogue and explosions seem to come from shifting locations on the screen, even though the physical sound may come from fixed speakers or earbuds. In ventriloquism, visual cues about a puppet’s moving mouth can reassign the perceived sound source away from the actual speaker. The brain uses ‘most likely source’ rules that weigh visual information heavily, and constantly resolves multiple delayed echoes into a single auditory event. Understanding this construction process highlights why acoustic environments (rooms, materials, ceilings) matter for clarity and comfort, and why audio illusions are so convincing.

Music and Rhythm Likely Evolved to Coordinate Group Action and Emotion

Music appears in every culture, but its evolutionary function is less obvious than language. Groh highlights a theory that rhythm and group sound-making (stomping, chanting, singing together) evolved to synchronize group behavior and amplify collective presence—for example, to scare off predators or competitors around a carcass. Rhythmic coordination feels good, reinforcing cooperation. This framing helps explain military bands, war chants, pre-game haka rituals, and the social bonding power of concerts. It also reframes music as a powerful tool for state control—energizing vigor, signaling unity, or regulating mood—rather than mere entertainment.

WORDS WORTH SAVING

5 quotes

What goes on in our brains when we think might be that we're running simulations related to the thought using that sensory, sensory motor infrastructure of the brain.

Jennifer Groh

If you moved the eyes, the neurons' receptive field... would shift as the eyes moved. And that blew my mind.

Jennifer Groh

Half a millisecond is less than the duration of a single action potential.

Andrew Huberman

Your ears are making sounds, folks.

Jennifer Groh

Being blocked can mean you don't know yet what needs to come next.

Jennifer Groh

Sensory integration of vision and audition in the brainNeural basis of sound localization and spatial hearingTop-down control of the ear and eye: eye movements, eardrum motion, and attentionMusic, rhythm, emotion, and evolutionary roles of group soundThought as multi-sensory simulation in sensory-motor cortexAttention, flow states, and practical focus strategies in a phone-saturated worldEnvironmental and architectural effects on sound perception and cognition

High quality AI-generated summary created from speaker-labeled transcript.

Get more out of YouTube videos.

High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.

Add to Chrome