• Welcome to ASR. There are many reviews of audio hardware and expert members to help answer your questions. Click here to have your audio equipment measured for free!

Insights Into Sensory Integration, Sound Localization, and the Neuroscience of Thought

  • Thread starter Thread starter DonM
  • Start date Start date
  • Featured

DonM

Member
Joined
Feb 5, 2021
Messages
90
Likes
99
Location
Ontario, Canada
I recently watched a fascinating discussion between Dr. Jennifer Groh (Duke University, Psychology & Neuroscience) and Andrew Huberman that offers deep insights into how our brains encode and integrate sights and sounds, and what this means for perception, attention, and cognition. The conversation goes far beyond typical pop neuroscience, drilling into topics directly relevant to audio science: dynamic sensory maps in the brain, how the superior colliculus merges auditory and visual space, mechanisms behind sound localization (including milliseconds-scale interaural time differences), and the profound role of physical environment and resonant frequencies in shaping perception. Dr. Groh also tackles what “thought” is from a neurobiological perspective, with practical advice on focus, brain state optimization, and the impact of sound environments—including music, noise, and silence—on our ability to learn and concentrate.

For anyone interested in technical audio reproduction, psychoacoustics, or the neuroscience behind how we experience sound (YouTube - safe link), this interview contains some of the clearest explanations I’ve encountered. As Audiophiles we tend to focus on the source of the sound, the amplifiers, the speakers and the effect of the room. Dr Groh during this interview explains how our brains interpret that sound. Feel free to add your own perspectives!
 
Dr. Jennifer Groh (Duke University) is a neuroscientist whose research fundamentally changes how we think about sound localization and auditory processing in the brain.

Her key finding, crucial for audio science, is that the brain does not have a purely "auditory" map of space. Instead, it constantly translates and aligns sound location information with visual and eye movement signals.

Her lab has shown:

  • Vision Calibrates Hearing: Brain areas thought to be purely visual (like the visual cortex) are influenced by sound, and vice-versa. This cross-talk is critical for creating a single, coherent perceptual space.
  • Eye Position Remaps Sound: The brain's interpretation of where a sound is located changes based on the position of your eyes. When you move your eyes, your auditory spatial map is updated to match. This means hearing is not an independent process but is deeply integrated with our visual and motor systems.
Implication for Audio Science: Her work demonstrates that "accurate" sound perception is not just a function of the ear or an audio system's fidelity, but a active, multi-sensory computational process in the brain. This has profound implications for understanding how listeners integrate what they see (like a speaker's position) with what they hear, and for designing audio systems that account for this natural brain function.
 
Dr. Jennifer Groh (Duke University) is a neuroscientist whose research fundamentally changes how we think about sound localization and auditory processing in the brain.

Her key finding, crucial for audio science, is that the brain does not have a purely "auditory" map of space. Instead, it constantly translates and aligns sound location information with visual and eye movement signals.

Her lab has shown:

  • Vision Calibrates Hearing: Brain areas thought to be purely visual (like the visual cortex) are influenced by sound, and vice-versa. This cross-talk is critical for creating a single, coherent perceptual space.
  • Eye Position Remaps Sound: The brain's interpretation of where a sound is located changes based on the position of your eyes. When you move your eyes, your auditory spatial map is updated to match. This means hearing is not an independent process but is deeply integrated with our visual and motor systems.
Implication for Audio Science: Her work demonstrates that "accurate" sound perception is not just a function of the ear or an audio system's fidelity, but a active, multi-sensory computational process in the brain. This has profound implications for understanding how listeners integrate what they see (like a speaker's position) with what they hear, and for designing audio systems that account for this natural brain function.
Is this a summary of her comments in the video you posted or her published research?
 
Is this a summary of her comments in the video you posted or her published research?
It is primarily a summary of her research and Andrew H draws out highlights of that research with his questions during the video interview.
 
Cool stuff, thanks for sharing! Most of what I've heard about Huberman is about his influence on the "wellness" space, but it sounds like his guest had some interesting stuff to share on audio. Of course we know we listen with our eyes first (to paraphrase Rachel Ray) and it's great to get more detail on that general impression.
 
Cool stuff, thanks for sharing! Most of what I've heard about Huberman is about his influence on the "wellness" space, but it sounds like his guest had some interesting stuff to share on audio. Of course we know we listen with our eyes first (to paraphrase Rachel Ray) and it's great to get more detail on that general impression.
Yes we do listen with our eyes. That is why speakers are normally hidden behind a curtain for listener testing.
Thanks. For my own practical purpose, do I need to watch the video or would listening be enough? I mean, does the information presented depend on visuals in the video?
No, you do not need to watch the video interview between Huberman and Dr. Groh. There are no illustrations shown in the interview that must be viewed. (Edited response Nov. 11)
 
Last edited:
I didn't watch the video..

Implication for Audio Science: Her work demonstrates that "accurate" sound perception is not just a function of the ear or an audio system's fidelity, but a active, multi-sensory computational process in the brain.
I think most of us are aware of that. ;) It's why measurements are helpful. And it's one of the reasons that controlled-blind listening tests are important... To minimize the number of variables and influences. You can't eliminate our state-of-mind or emotions but we can try to minimize those effects by making the test quick, and if the results are statistically repeatable that can remove-randomize some of these "unknown" and uncontrollable variables and biases.

The whole "stereo soundstage" is obviously an illusion with the sound coming from a pair of speakers. And I ASSUME different people experience it differently (in the same room with the same recording and same speakers, etc.). Some people hear front-to back of the stage "depth'. I only hear left-to right. And the music I listen to is produced in a studio and "artificially" mixed and panned anyway. There are supposed to be "tricks" with timing to bring the sounds forward or back but I don't know how much it's done and in any case, I just don't hear it...

BTW - Floyd Toole says:
The important localization and soundstage information is the responsibility of the recording engineer, not the loudspeaker.
But he doesn't say the recording is the only important thing and I'm sure the speakers, room acoustics, and our brain also have an influence.


I KNOW people perceive headphone soundstage differently (Headphone soundstage survey).

Even in the "real world" we often over-estimate our ability to locate the source of a sound. But if you've ever tried to find a cricket or a squeak or rattle in a car it's usually not so easy.

Live amplified music it's usually mono with the exception of the sounds that might be coming directly from the stage (drums & guitar amps, etc.). If you are at a stadium or arena concert you're probably hearing mono from the PA speakers. In smaller venues you are more likely to hear at least some sounds from the stage.

Concert halls have lots of reverb and reflected sounds. And the angle-width of the stage is narrower than at home. I've wondered if we can audibly locate the instruments and if I ever hear a live symphony again I plan on closing my eyes to see if I can locate them, although I can't "forget" where the different sections are so I might fool myself...
 
Thank you for this post. She is very interesting. It’s a somewhat odd feeling when the brain tries to understand itself.
 
Try listening to a Scotish backpipe through gear in an untreated room (specially front wall and ceiling ) without following it with your eyes , left-right and up and down :facepalm:
That's eye-muscle and brain exercise!

Really interesting video, thanks for that.
 
I didn't watch the video..


I think most of us are aware of that. ;) It's why measurements are helpful. And it's one of the reasons that controlled-blind listening tests are important... To minimize the number of variables and influences. You can't eliminate our state-of-mind or emotions but we can try to minimize those effects by making the test quick, and if the results are statistically repeatable that can remove-randomize some of these "unknown" and uncontrollable variables and biases.

The whole "stereo soundstage" is obviously an illusion with the sound coming from a pair of speakers. And I ASSUME different people experience it differently (in the same room with the same recording and same speakers, etc.). Some people hear front-to back of the stage "depth'. I only hear left-to right. And the music I listen to is produced in a studio and "artificially" mixed and panned anyway. There are supposed to be "tricks" with timing to bring the sounds forward or back but I don't know how much it's done and in any case, I just don't hear it...

BTW - Floyd Toole says:

But he doesn't say the recording is the only important thing and I'm sure the speakers, room acoustics, and our brain also have an influence.


I KNOW people perceive headphone soundstage differently (Headphone soundstage survey).

Even in the "real world" we often over-estimate our ability to locate the source of a sound. But if you've ever tried to find a cricket or a squeak or rattle in a car it's usually not so easy.

Live amplified music it's usually mono with the exception of the sounds that might be coming directly from the stage (drums & guitar amps, etc.). If you are at a stadium or arena concert you're probably hearing mono from the PA speakers. In smaller venues you are more likely to hear at least some sounds from the stage.

Concert halls have lots of reverb and reflected sounds. And the angle-width of the stage is narrower than at home. I've wondered if we can audibly locate the instruments and if I ever hear a live symphony again I plan on closing my eyes to see if I can locate them, although I can't "forget" where the different sections are so I might fool myself...
Thanks for the detailed summary and links—these are great points and align closely with both Dr. Groh’s research and Floyd Toole’s longstanding observations about perception vs. physical stimulus.

Audio system measurements, controlled tests, and room corrections all matter enormously—but Dr. Groh’s neuroscience reminds us that subjective experience will always play a role, because perception is a multi-stage computational process, not just the end-product of accurate reproduction. Rather than dismissing individual differences as irrelevant, perhaps we can view them as part of the challenge (and beauty) of both audio engineering and listening.

By the way, if the video’s full length is a barrier, Dr. Huberman typically provides chapter links and timestamps in the YouTube description so you can jump directly to specific topics of interest:

Timestamps

00:00:00 Jennifer Groh

00:03:41 Sounds & Vision, Sensory Integration; Dynamic Maps

00:07:42 Context & Mapping; Screens, Projection & Perception, Ventriloquists

00:13:52 Sound Localization

00:19:50 Hearing Loss & Sound Localization, Ear Folds

00:21:56 Unfamiliarity of Hearing Your Own Voice; Tool: Bone Conduction Headphones

00:26:16 Tool: Headphone Volume & Protecting Hearing

00:28:57 3D Sound, Sound Distance, Thunder, Earthquakes

00:37:24 Sound Integration; Sound Frequency & Distance, Warning Signals

00:47:39 Music, Rhythm, Community & Emotion

00:57:00 Music, Military; Courtship; Evolution of Music & Language

01:02:37 Ears, Visual & Auditory Integration, Sound Localization

01:09:48 Evolution of Visual & Auditory Systems, Music; Brain Controlling Vision

01:16:45 Physical Space & Sounds; Cathedrals, Sound Delay

01:22:37 Music, Emotion & Community; Science & Admitting Weakness

01:27:01 Thinking & Sensory Simulations; Forming Thoughts

01:33:18 Attention, Attractor States, Flow States, Tool: Changing Environment

01:37:38 Sounds & Environment for Focus, Attention, Tool: Mental Interval Training

01:45:58 Endurance & Interval Mental Work; Mental Rest, Music

01:50:37 Musician, Rehearsal & Performance; Pressure

Special Note: At 7:30pm on Day 2 there have been over 55,000 persons who have viewed that interview.
 
Last edited:
Thank you for this post. She is very interesting. It’s a somewhat odd feeling when the brain tries to understand itself.
It really is fascinating—and a bit strange—that the brain can try to understand itself. Dr. Groh talks about how we use our own brain circuits to think about our thoughts and feelings, including how we listen and react to sound. That’s part of what makes neuroscience and audio so interesting: every time we listen or analyze sound, we’re also reflecting on our own mind at work. Thanks for pointing that out!
 
Try listening to a Scotish backpipe through gear in an untreated room (specially front wall and ceiling ) without following it with your eyes , left-right and up and down :facepalm:
That's eye-muscle and brain exercise!

Really interesting video, thanks for that.
Great example! In Scottish history, bagpipes weren’t just musical instruments—they played a powerful role in battle. Scottish warriors grew up during childhood hearing bagpipes at gatherings and important events, so their brains “learned” to associate that sound with action, unity, and even courage. By the time they became fighters, the sound of the pipes could trigger strong emotional and physical responses—almost like a call to arms. You likely did not grow up in a family with that same exposure to bagpipe music. Accordingly you react differently to that sound.

Thanks for raising this personal experience; it really brings Dr. Groh’s science to life!
 
- that the brain can try to understand itself.
Yes the book "Gödel, Escher, Bach: an Eternal Golden Braid" (GEB) does explore, in depth, the paradoxes and mysteries that arise when a system (like the brain) turns its powers of analysis onto itself. Dr. Groh in the video interview explains how the brain completes this.

This is no different than when we listen to sound and sit back trying to evaluate the sound source, the amplifier, the speaker and finally the room interaction.

I posted that interview to bring awareness to everyone about how their brains process the sounds they are listening to with it's impact. And since everyone grows up during childhood under different scenarios this can explain in part why their preference for sound systems can be different.
 
Last edited:
And since everyone grows up during childhood under different scenarios this can explain in part why their preference for sound systems can be different.
This is wildly speculative and at odds with blind studies showing that people's audio preferences are almost shockingly homogeneous
 
This is wildly speculative and at odds with blind studies showing that people's audio preferences are almost shockingly homogeneous
Thanks for pointing that out. Yes you are absolutely correct. My comment however was directed towards consumers (not those carefully screened and trained for blind tests) who both view and listen to speakers (not blindly) before they are purchased and listening to them after. This comment can not be substantiated though by a recognized study or survey which provides an actual percentage. If we just looked at a sample of actual blind speaker tests posted on Audio Science and then looked at the actual sales, you would still see a wide range of speakers which did not win the blind tests still being purchased by consumers.

Dr. Grohs information on the brain processing included both visual and auditory and is supported by scientific studies.
 
This is wildly speculative and at odds with blind studies showing that people's audio preferences are almost shockingly homogeneous
People's experiences can certainly influence their preference for content. But there's no logical reason why there would be any influence on what one would like for reproducing said content (ingrained brand or appearance preferences aside). You want to hear what was produced, therefore you logically want as neutral and transparent a reproduction chain as possible.
 
People's experiences can certainly influence their preference for content. But there's no logical reason why there would be any influence on what one would like for reproducing said content (ingrained brand or appearance preferences aside). You want to hear what was produced, therefore you logically want as neutral and transparent a reproduction chain as possible.
I’m curious. Have you watched the full interview between Huberman and Dr Groh? Where did Dr Groh state that persons would always prefer a speaker or sound that was neutral and transparent?
 
Back
Top Bottom