Lex Fridman Podcast

Edward Gibson: Human Language, Psycholinguistics, Syntax, Grammar & LLMs | Lex Fridman Podcast #426

Lex Fridman and Edward (Ted) Gibson on mIT linguist dissects language, thought, LLMs, and why legalese fails.

Edward (Ted) GibsonguestLex Fridmanhost
Apr 17, 20242h 50m
Dependency grammar, syntax, and word order universalsCognitive cost of long-distance dependencies and center-embeddingLanguage vs. thought: brain networks, aphasia, and modularityCultural variation: Pirahã and Tsimane number and color systemsLegalese: why contracts are uniquely hard to understandLarge language models: form vs meaning, and construction grammarEvolution, learnability, and innateness of language

In this episode of Lex Fridman Podcast, featuring Edward (Ted) Gibson and Lex Fridman, Edward Gibson: Human Language, Psycholinguistics, Syntax, Grammar & LLMs | Lex Fridman Podcast #426 explores mIT linguist dissects language, thought, LLMs, and why legalese fails Lex Fridman interviews MIT psycholinguist Edward (Ted) Gibson about how human language is structured, processed, and learned, and how this contrasts with large language models. Gibson argues for a dependency-grammar view of syntax and shows that across languages, people strongly prefer short dependencies because long-distance links are cognitively costly. He distinguishes language (a communication system) from thought, citing brain-imaging and neuropsychology showing that high-level language and non‑linguistic reasoning use different neural systems. The conversation ranges from Pirahã number words and Amazonian color terms to the pathology of legalese, the limits of LLM “understanding,” and speculative ideas about communicating with animals and aliens.

At a glance

WHAT IT’S REALLY ABOUT

MIT linguist dissects language, thought, LLMs, and why legalese fails

  1. Lex Fridman interviews MIT psycholinguist Edward (Ted) Gibson about how human language is structured, processed, and learned, and how this contrasts with large language models. Gibson argues for a dependency-grammar view of syntax and shows that across languages, people strongly prefer short dependencies because long-distance links are cognitively costly. He distinguishes language (a communication system) from thought, citing brain-imaging and neuropsychology showing that high-level language and non‑linguistic reasoning use different neural systems. The conversation ranges from Pirahã number words and Amazonian color terms to the pathology of legalese, the limits of LLM “understanding,” and speculative ideas about communicating with animals and aliens.

IDEAS WORTH REMEMBERING

7 ideas

Human languages strongly minimize dependency length.

Across ~60 typologically diverse languages with parsed corpora, actual sentences consistently have much shorter word-to-word dependency distances than randomized but grammatically plausible alternatives, indicating a universal pressure to keep related words close for easier production and comprehension.

Center-embedding is universally hard for humans and LLMs.

Nested structures like “The boy who the cat that the dog chased scratched cried” massively increase dependency distances and working-memory load; both humans and large language models struggle to complete or process such sentences, suggesting shared constraints on processing form.

Legalese is difficult primarily because of extreme center-embedding, not jargon or passives.

Corpus and behavioral studies on contracts show unusually high rates of center-embedded clauses (e.g., definitions wedged between subject and verb), which severely hurt comprehension and recall for both laypeople and lawyers; low-frequency vocabulary matters somewhat, while passive voice has negligible effect.

Language and thought are neurally dissociable systems.

fMRI work (Fedorenko et al.) finds a stable, left-lateralized “language network” activated by sentences (spoken or written) but not by math, music, programming, or other demanding cognitive tasks, while patients with severe aphasia can still reason, play chess, and do arithmetic—showing that high-level thinking doesn’t require language.

Words people invent reflect communicative needs, not perceptual limits.

Groups like the Tsimane and Pirahã see the same colors and numerosities we do but have far fewer basic color terms and lack exact number words (even for ‘one’); experiments show they use approximate quantifiers (‘few/some/many’) and can match sets perceptually but can’t perform exact counting tasks, highlighting that lexical systems track what must be talked about, not what can be perceived.

Movement-based syntactic theories are harder to learn than lexicalized, dependency-based ones.

Chomsky’s classic movement analyses (e.g., moving auxiliaries to form questions) create learnability problems and require innate “universal grammar” assumptions, whereas alternative accounts with word-specific argument-structure patterns (lexical copying) and dependency trees are simpler, more compatible with data irregularities, and more learnable from input.

Current LLMs model form extremely well but show brittle grasp of meaning.

They approximate construction-grammar-like patterns and even internally recover something akin to dependency structure, yet fail on small semantic twists (e.g., Monty Hall variants where probabilities are explicitly given), indicating that high-quality surface prediction doesn’t equate to human-like understanding or reasoning about underlying situations.

WORDS WORTH SAVING

5 quotes

Language is an invented system by humans for communicating their ideas.

Edward Gibson

I don’t see any limits to their form. Their form is perfect.

Edward Gibson (on large language models)

We don’t think in language.

Edward Gibson

Legalese is massively center-embedded. About 70 percent of sentences have a center-embedded clause.

Edward Gibson

Naively, I certainly thought that all humans would have words for exact counting. And the Pirahã don’t.

Edward Gibson

QUESTIONS ANSWERED IN THIS EPISODE

5 questions

If language and thought are distinct systems in the brain, what exactly are the representations and mechanisms that underlie non-linguistic thought?

Lex Fridman interviews MIT psycholinguist Edward (Ted) Gibson about how human language is structured, processed, and learned, and how this contrasts with large language models. Gibson argues for a dependency-grammar view of syntax and shows that across languages, people strongly prefer short dependencies because long-distance links are cognitively costly. He distinguishes language (a communication system) from thought, citing brain-imaging and neuropsychology showing that high-level language and non‑linguistic reasoning use different neural systems. The conversation ranges from Pirahã number words and Amazonian color terms to the pathology of legalese, the limits of LLM “understanding,” and speculative ideas about communicating with animals and aliens.

How far can large language models go toward genuine understanding if they are only ever directly trained on form—do we eventually need grounded perception and action?

Could we quantify a ‘cognitive cost function’ for dependency length well enough to automatically simplify complex legal or technical texts without changing their meaning?

What social, economic, or technological conditions tend to trigger the invention of exact number systems and other major lexical innovations in a culture?

If we eventually decode whale or bird communication, what criteria should we use to decide whether what they have counts as ‘language’ in the human sense?

EVERY SPOKEN WORD

Install uListen for AI-powered chat & search across the full episode — Get Full Transcript

Get more out of YouTube videos.

High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.

Add to Chrome