Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5

Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5

Lex Fridman PodcastNov 16, 201854m

Lex Fridman (host), Vladimir Vapnik (guest), Narrator

Instrumentalism vs realism in science and machine learningRole and limits of mathematics in understanding reality and learningHuman intuition, axioms, and the discovery of simple underlying principlesTwo mechanisms of learning: strong convergence vs weak convergencePredicates, invariants, and the role of the teacher in learningVC theory, admissible sets of functions, and statistical learning theoryCritique of deep learning, data inefficiency, and open challenges in AI

In this episode of Lex Fridman Podcast, featuring Lex Fridman and Vladimir Vapnik, Vladimir Vapnik: Statistical Learning | Lex Fridman Podcast #5 explores vladimir Vapnik on learning, intelligence, and the limits of deep learning Vladimir Vapnik discusses the philosophical and mathematical foundations of statistical learning, contrasting instrumentalism (prediction) with realism (understanding "God's laws"). He argues that modern machine learning overemphasizes brute-force prediction and deep learning, while neglecting conditional probabilities, invariants, and the role of a "teacher" in providing powerful predicates. Vapnik introduces his view that there are two mechanisms of learning—strong and weak convergence—with weak convergence relying on high‑level predicates like “swims like a duck” that dramatically reduce data requirements. He sees the central open problem as understanding intelligence: how good teachers generate such predicates, and how to formalize that process to achieve learning with far fewer examples.

Vladimir Vapnik on learning, intelligence, and the limits of deep learning

Vladimir Vapnik discusses the philosophical and mathematical foundations of statistical learning, contrasting instrumentalism (prediction) with realism (understanding "God's laws"). He argues that modern machine learning overemphasizes brute-force prediction and deep learning, while neglecting conditional probabilities, invariants, and the role of a "teacher" in providing powerful predicates. Vapnik introduces his view that there are two mechanisms of learning—strong and weak convergence—with weak convergence relying on high‑level predicates like “swims like a duck” that dramatically reduce data requirements. He sees the central open problem as understanding intelligence: how good teachers generate such predicates, and how to formalize that process to achieve learning with far fewer examples.

Key Takeaways

Distinguish prediction from understanding in machine learning.

Vapnik argues that most current ML is instrumentalist—focused on finding rules that predict well—whereas deeper understanding requires modeling conditional probabilities and “how God plays dice,” not just learning classifiers.

Get the full analysis with uListen AI

Use mathematics to uncover structures humans cannot intuit.

He maintains that the strongest human intuition resides in well‑chosen axioms; once those are set, following equations often reveals simple, beautiful principles that intuition alone would likely miss.

Get the full analysis with uListen AI

Incorporate predicates and invariants to reduce data needs dramatically.

Vapnik’s weak convergence mechanism uses high-level predicates (e. ...

Get the full analysis with uListen AI

Focus on constructing good admissible sets of functions, not just fitting models.

Statistical learning theory assumes a given hypothesis space, but Vapnik emphasizes that the real hard problem is building an admissible set: small VC dimension yet rich enough to contain a good solution, guided by invariants from data.

Get the full analysis with uListen AI

Reconsider the role and necessity of deep learning architectures.

He criticizes deep learning as largely interpretive “fantasy,” noting that mathematics doesn’t need neurons per se and that representer theorems point to shallow networks as sufficient optima in many learning formulations.

Get the full analysis with uListen AI

Study teachers to understand intelligence.

Vapnik sees the core of intelligence in what great teachers do—producing insightful predicates like “play like a butterfly” that instantly reshape a learner’s behavior—yet notes that we have almost no formal theory of this process.

Get the full analysis with uListen AI

Formulate concrete challenges to probe intelligence and efficiency.

He proposes benchmarks such as matching deep learning’s digit recognition performance using 100× fewer examples by introducing well-chosen invariants, framing this as a litmus test for progress on the intelligence problem.

Get the full analysis with uListen AI

Notable Quotes

The goal of machine learning is to find the rule for classification. That is true, but it is an instrument for prediction. For understanding, I need conditional probability.

Vladimir Vapnik

The best human intuition, it is putting in axioms, and then it is technical where to see where the axioms take you.

Vladimir Vapnik

In weak convergence mechanism you can use predicates—that’s what ‘play like a butterfly’ is—and it will immediately affect your playing.

Vladimir Vapnik

The most difficult problem is to create the admissible set of functions… This was out of consideration.

Vladimir Vapnik

I think that this [deep learning] is fantasy. Everything which… like deep learning, like features… they are not really want of the problem.

Vladimir Vapnik

Questions Answered in This Episode

How could we formally model the process by which a great teacher invents powerful predicates like “play like a butterfly” or “swims like a duck”?

Vladimir Vapnik discusses the philosophical and mathematical foundations of statistical learning, contrasting instrumentalism (prediction) with realism (understanding "God's laws"). ...

Get the full analysis with uListen AI

What practical steps can current ML researchers take to integrate invariants and weak convergence mechanisms into mainstream learning algorithms?

Get the full analysis with uListen AI

Are there concrete examples where shallow, theoretically motivated models can match or outperform deep networks with far less data, and what do they teach us?

Get the full analysis with uListen AI

How might a theory of intelligence that goes beyond imitation (à la Turing) incorporate Vapnik’s ideas about predicates, teachers, and shared ‘ground truths’?

Get the full analysis with uListen AI

What kinds of empirical experiments could distinguish between problems that truly require massive data and those that mainly need better invariants and hypothesis spaces?

Get the full analysis with uListen AI

Transcript Preview

Lex Fridman

The following is a conversation with Vladimir Vapnik. He's the co-inventor of support vector machines, support vector clustering, VC theory, and many foundational ideas in statistical learning. He was born in the Soviet Union and worked at the Institute of Control Sciences in Moscow. Then, in the United States, he worked at AT&T, NEC labs, Facebook Research, and now is a professor at Columbia University. His work has been cited over 170,000 times. He has some very interesting ideas about artificial intelligence and the nature of learning, especially on limits of our current approaches and the open problems in the field. This conversation is part of MIT course on Artificial General Intelligence and the Artificial Intelligence podcast. If you enjoy it, please subscribe on YouTube or rate it on iTunes or your podcast provider of choice, or simply connect with me on Twitter or other social networks @lexfridman, spelled F-R-I-D. And now, here's my conversation with Vladimir Vapnik. Einstein famously said that God doesn't play dice.

Vladimir Vapnik

Yeah.

Lex Fridman

You have studied the world through the eyes of statistics, so let me ask you in terms of the nature of reality, fundamental nature of reality, does God play dice?

Vladimir Vapnik

We don't know some factors, and because we don't know some factors which could be important, it looks like God play dice, but we only should describe. In philosophy, they distinguish between two positions, positions of instrumentalism, where you're creating theory for prediction, and position of realism, where you're trying to understand what God did.

Lex Fridman

Can you describe instrumentalism and realism a little bit?

Vladimir Vapnik

For example, if you have some mechanical laws, what is that? Is it law which true always and everywhere or it is law which allow you to predict position of moving element? The, uh, what, what you believe, you believe that it is God's law, that God created the world which adhere to this physical law-

Lex Fridman

Yeah.

Vladimir Vapnik

... or it is just law for predictions?

Lex Fridman

And which one is instrumentalism?

Vladimir Vapnik

For predictions.

Lex Fridman

Just predict.

Vladimir Vapnik

If you believe that this is law of God-

Lex Fridman

Mm-hmm.

Vladimir Vapnik

... and it's always true everywhere, that means that you're realist.

Lex Fridman

So you-

Vladimir Vapnik

You're trying to re- to really, uh, understood, understand the God's thought.

Lex Fridman

So the way you see the world is, is an instrumentalist?

Vladimir Vapnik

You know-

Lex Fridman

Absolutely.

Vladimir Vapnik

... I'm working for some models, model of, uh, machine learning. So in this model, we can see, um, setting and we try to solve, resolve the setting to solve the problem. And you can do it in two different way. From the point of view of instrumentalist, and that's what everybody does now, because, uh, they say the goal of machine learning is to, uh, find the rule for classification.

Lex Fridman

Mm-hmm.

Vladimir Vapnik

That is true, but it is instrument for prediction. But I can say the goal of, uh, machine learning is to, to learn about conditional probability, so how God play deuce, and he... if he play, what is probability for one, what is probability for another given situation. But for prediction, I don't need this. I need the rule.

Install uListen to search the full transcript and get AI-powered insights

Get Full Transcript

Get more from every podcast

AI summaries, searchable transcripts, and fact-checking. Free forever.

Add to Chrome