Skip to content
Lex Fridman PodcastLex Fridman Podcast

Jay McClelland: Neural Networks and the Emergence of Cognition | Lex Fridman Podcast #222

Jay McClelland is a cognitive scientist at Stanford. Please support this podcast by checking out our sponsors: - Paperspace: https://gradient.run/lex to get $15 credit - Skiff: https://skiff.org/lex to get early access - Uprising Food: https://uprisingfood.com/lex to get $10 off 1st starter bundle - Four Sigmatic: https://foursigmatic.com/lex and use code LexPod to get up to 60% off - Onnit: https://lexfridman.com/onnit to get up to 10% off EPISODE LINKS: Jay's Website: https://stanford.edu/~jlmcc/ PODCAST INFO: Podcast website: https://lexfridman.com/podcast Apple Podcasts: https://apple.co/2lwqZIr Spotify: https://spoti.fi/2nEwCF8 RSS: https://lexfridman.com/feed/podcast/ Full episodes playlist: https://www.youtube.com/playlist?list=PLrAXtmErZgOdP_8GztsuKi9nrraNbKKp4 Clips playlist: https://www.youtube.com/playlist?list=PLrAXtmErZgOeciFP3CBCIEElOJeitOr41 OUTLINE: 0:00 - Introduction 0:43 - Beauty in neural networks 5:02 - Darwin and evolution 10:47 - The origin of intelligence 17:29 - Explorations in cognition 23:33 - Learning representations by back-propagating errors 29:58 - Dave Rumelhart and cognitive modeling 43:01 - Connectionism 1:05:54 - Geoffrey Hinton 1:07:49 - Learning in a neural network 1:24:42 - Mathematics & reality 1:31:50 - Modeling intelligence 1:42:28 - Noam Chomsky and linguistic cognition 1:56:49 - Advice for young people 2:07:56 - Psychiatry and exploring the mind 2:20:35 - Legacy 2:26:24 - Meaning of life SOCIAL: - Twitter: https://twitter.com/lexfridman - LinkedIn: https://www.linkedin.com/in/lexfridman - Facebook: https://www.facebook.com/lexfridman - Instagram: https://www.instagram.com/lexfridman - Medium: https://medium.com/@lexfridman - Reddit: https://reddit.com/r/lexfridman - Support on Patreon: https://www.patreon.com/lexfridman

Lex FridmanhostJay McClellandguest
Sep 20, 20212h 31mWatch on YouTube ↗

At a glance

WHAT IT’S REALLY ABOUT

Neural Networks, Emergent Mind, and the Limits of Human Understanding

  1. Jay McClelland traces how neural networks bridge biology and cognition, arguing that thought and meaning emerge from massively parallel, connectionist systems rather than explicit symbolic rules. He recounts the early days of connectionism with Rumelhart and Hinton, including the birth of backpropagation and interactive activation models, and how these reshaped cognitive science and AI.
  2. The conversation explores emergence across evolution, brain development, language, and mathematics, contrasting intuitive, distributed representations with formal logic and symbolic AI. McClelland also discusses neurological conditions like semantic dementia as windows into how meaning is represented and lost in the brain.
  3. He reflects on the influence of formal training on how experts misunderstand ordinary cognition, the interplay of intuition and proof in mathematics, and what large neural networks suggest about creativity and insight. The discussion ends with personal themes: intrinsic motivation, legacy, degeneration versus death, and the idea that humans create meaning rather than discover a pre-given one.

IDEAS WORTH REMEMBERING

5 ideas

Neural networks offer a mechanistic link between brain biology and thought.

By modeling simple neuron-like units connected in parallel and layers, neural networks show how high-level cognition can emerge from low-level biological processes, dissolving the old Cartesian separation between body and mind.

Connectionism encodes knowledge in weights, not in explicit symbols or rules.

In McClelland’s view, systems like interactive activation models and modern CNNs demonstrate that what we call ‘knowledge’ is distributed across connections; there is no internal dictionary, only patterns of connectivity that shape input–output behavior.

Backpropagation arose from reframing learning as optimization, not biology mimicry.

Hinton’s push to ‘define an objective function and minimize error’ led Rumelhart to generalize the single-layer delta rule into backpropagation, propagating error signals backward through layers—now the core of deep learning.

Emergent phenomena can be real and important without being explicitly represented.

From evolution and developmental stages to insights in math and language, McClelland argues for a ‘radical emergentist’ stance: high-level structures (thoughts, concepts, sand dunes) are real patterns arising from lower-level dynamics, even if they aren’t discrete symbols inside the system.

Semantic dementia reveals how distributed meaning representations degrade.

Patients who progressively lose semantic distinctions (e.g., calling all four-legged animals ‘dog’ or ‘horse’) illustrate how overlapping, graded representations in the brain can collapse, matching predictions from parallel distributed processing models.

WORDS WORTH SAVING

5 quotes

If I think about the mind in terms of a neural network, it will help me answer the questions about the mind that I'm trying to answer.

Jay McClelland

The unit for the word ‘time’ isn’t a unit for the word ‘time’ for any other reason than it’s got the connections to the letters that make up the word ‘time.’

Jay McClelland

It is by logic that we prove, but by intuition that we discover.

Henri Poincaré (quoted by Jay McClelland)

I used to sometimes tell people I was a radical eliminative connectionist… I don’t like the word ‘eliminative’ anymore… I would call myself a radical emergentist connectionist.

Jay McClelland

Meaning is what we make of it… we are an emergent result of a process that happened naturally without guidance.

Jay McClelland

Neural networks as bridges between biology and cognitionHistory of connectionism, PDP, and the invention of backpropagationEmergence: evolution, development, cognition, and complex systemsDistributed representation, semantic cognition, and semantic dementiaIntuition vs. formalism in mathematics and languageSymbolic AI vs. connectionism and modern deep learningPersonal meaning, intrinsic motivation, and scientific legacy

High quality AI-generated summary created from speaker-labeled transcript.

Get more out of YouTube videos.

High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.

Add to Chrome