Skip to content
Lex Fridman PodcastLex Fridman Podcast

Dileep George: Brain-Inspired AI | Lex Fridman Podcast #115

Dileep George is a researcher at the intersection of neuroscience and artificial intelligence, co-founder of Vicarious, formerly co-founder of Numenta. From the early work on Hierarchical temporal memory to Recursive Cortical Networks to today, Dileep's always sought to engineer intelligence that is closely inspired by the human brain. Support this channel by supporting our sponsors. Click links, get discount: - Babbel: https://babbel.com and use code LEX - MasterClass: https://masterclass.com/lex - Raycon: https://buyraycon.com/lex EPISODE LINKS: Dileep's Twitter: https://twitter.com/dileeplearning Vicarious Research: https://www.vicarious.com/science PODCAST INFO: Podcast website: https://lexfridman.com/podcast Apple Podcasts: https://apple.co/2lwqZIr Spotify: https://spoti.fi/2nEwCF8 RSS: https://lexfridman.com/feed/podcast/ Full episodes playlist: https://www.youtube.com/playlist?list=PLrAXtmErZgOdP_8GztsuKi9nrraNbKKp4 Clips playlist: https://www.youtube.com/playlist?list=PLrAXtmErZgOeciFP3CBCIEElOJeitOr41 OUTLINE: 0:00 - Introduction 4:50 - Building a model of the brain 17:11 - Visual cortex 27:50 - Probabilistic graphical models 31:35 - Encoding information in the brain 36:56 - Recursive Cortical Network 51:09 - Solving CAPTCHAs algorithmically 1:06:48 - Hype around brain-inspired AI 1:18:21 - How does the brain learn? 1:21:32 - Perception and cognition 1:25:43 - Open problems in brain-inspired AI 1:30:33 - GPT-3 1:40:41 - Memory 1:45:08 - Neuralink 1:51:32 - Consciousness 1:57:59 - Book recommendations 2:06:49 - Meaning of life CONNECT: - Subscribe to this YouTube channel - Twitter: https://twitter.com/lexfridman - LinkedIn: https://www.linkedin.com/in/lexfridman - Facebook: https://www.facebook.com/LexFridmanPage - Instagram: https://www.instagram.com/lexfridman - Medium: https://medium.com/@lexfridman - Support on Patreon: https://www.patreon.com/lexfridman

Lex FridmanhostDileep Georgeguest
Aug 14, 20202h 10mWatch on YouTube ↗

EVERY SPOKEN WORD

  1. 0:004:50

    Introduction

    1. LF

      The following is a conversation with Daleep George, a researcher at the intersection of neuroscience and artificial intelligence, co-founder of Vicarious with Scott Phoenix, and formerly co-founder of Numenta with Jeff Hawkins, who's been on this podcast, and Donna Dubinsky. From his early work on hierarchical temporal memory to recursive cortical networks to today, Daleep's always sought to engineer intelligence that is closely inspired by the human brain. As a side note, I think we understand very little about the fundamental principles underlying the function of the human brain, but the little we do know gives hints that may be more useful for engineering intelligence than any idea in mathematics, computer science, physics, and scientific fields outside of biology. And so the brain is a kind of existence proof that says it's possible, keep at it. I should also say that brain-inspired AI is often overhyped and used as fodder just as quantum computing for, uh, marketing speak. But I'm not afraid of exploring these sometimes overhyped areas since where there's smoke, there's sometimes fire. Quick summary of the ads. Three sponsors, Babbel, Raycon earbuds, and MasterClass. Please consider supporting this podcast by clicking the special links in the description to get the discount. It really is the best way to support this podcast. If you enjoy this thing, subscribe on YouTube, review it with five stars on Apple Podcasts, support on Patreon, or connect with me on Twitter @LexFridman. As usual, I'll do a few minutes of ads now and never any ads in the middle that can break the flow of the conversation. This show is sponsored by Babbel, an app and website that gets you speaking in a new language within weeks. Go to babbel.com and use code LEX to get three months free. They offer 14 languages, including Spanish, French, Italian, German, and yes, Russian. Daily lessons are 10 to 15 minutes, super easy, effective, designed by over 100 language experts. Let me read a few lines from the Russian poem, (Russian) by Alexander Blok that you'll start to understand if you sign up to Babbel. (Russian) . Now, I say that you'll only start to understand this poem because Russian starts with language and ends with vodka. Now, the latter part is definitely not endorsed or provided by Babbel and will probably lose me the sponsorship, but once you graduate from Babbel, you can enroll in my advanced course of late night Russian conversation over vodka. I have not yet developed an app for that. It's in progress. So get started by visiting babbel.com and use code LEX to get three months free. This show is sponsored by Raycon earbuds. Get them at buyraycon.com/lex. They've become my main method of listening to podcasts, audiobooks, and music when I run, do push-ups and pull-ups, or just living life. In fact, I often listen to brown noise with them when I'm thinking deeply about something. It helps me focus. They're super comfortable, pair easily, great sound, great bass, six hours of playtime. I've been putting in a lot of miles to get ready for a potential ultra-marathon and listening to audiobooks on World War II. The sound is rich and really comes in clear. So again, get them at buyraycon.com/lex. This show is sponsored by MasterClass. Sign up at masterclass.com/lex to get a discount and to support this podcast. When I first heard about MasterClass, I thought it was too good to be true. I still think it's too good to be true. For 180 bucks a year, you get an all-access pass to watch courses from, to list some of my favorites, Chris Hadfield on space exploration, Neil deGrasse Tyson on scientific thinking and communication, Will Wright, creator of SimCity and Sims, on game design. Every time I do this read (laughs) , I really want to play a city builder game. Carlos Santana on guitar, Garry Kasparov on chess, Daniel Negreanu on poker, and many more. Chris Hadfield explaining how rockets work and the experience of being launched into space alone is worth the money. By the way, you can watch it on basically any device. Once again, sign up at masterclass.com to get a discount and to support this podcast. And now, here's my conversation with Daleep George.

  2. 4:5017:11

    Building a model of the brain

    1. LF

      Do you think we need to understand a brain in order to build it?

    2. DG

      Yes, if you want to build a brain, we definitely need to understand how it works. So Blue Brain or Henry Markram's project, uh, is trying to build a brain without understanding it. Like, you know, just trying to, uh, put details of the brain from neuroscience experiments into a giant simulation, uh, by putting more and more neurons, more and more details. Uh, but that is not going to work, uh, because when it doesn't perform as, uh, what you expect it to do, then what do you do? Do you just keep adding more details? How do you debug it? So it's a... So unless you understand, unless you have a theory about how the system is supposed to work, how the pieces are supposed to fit together, what they're going to contribute, you can't, you can't build it.

    3. LF

      At the functional level. Understand. So can you actually linger on and describe the Blue Brain project? It's kinda fascinating, uh, principle and idea to try to simulate the brain. As we're talking about the human brain, right?

    4. DG

      Right. Human brains and rat brains or cat brains have lots in common. The cortex, uh, the neocortex structure is very similar. So initially they were trying to just simulate a cat brain. Uh, and, uh-

    5. LF

      To understand the nature of evil.

    6. DG

      (laughs) And understand the nature of evil. Uh- Yeah. ... or, uh, as it happens in most of these simulations, uh, you, you easily get one thing out, which is oscillations, you know? Yeah. If you, you sim- if you simulate a large number of neurons, they oscillate, uh, and, uh, you can adjust the parameters and say that, "Oh, oscillations match the rhythm that we see in the brain," et cetera. But, uh-

    7. LF

      Oh, I see. So like, uh, so the idea is, uh, is the simulation at the level of, uh, individual neurons?

    8. DG

      Yeah. So the Blue Brain Project, the original idea as proposed was, um, you, you put very detailed bio physical neurons, uh, bio physical models of neurons, and you interconnect them according to the statistics of connections that we have found from real neuroscience experiments. And then, uh, turn it on and, uh, see what happens. Uh, and, and these neural models are, you know, incredibly complicated in themselves, right, because these neurons are, uh, modeled using, uh, this, uh, idea called Hodgkin-Huxley models, which are about how signals propagate in a cable. And there are a- active dendrites, all those phenomena which those phenomena themselves we don't understand that well. Uh, and then, uh, we put in connectivity, which is part guesswork, part, you know, observed. And of course, if you do not have any theory about how it is supposed to work, uh, we, you, you know, we just have to take whatever comes out of it as, "Okay, this is something interesting." You know?

    9. LF

      But in your sense, like these models of the way signal travels along... or like with the axons and all the basic models, that's... they're too crude?

    10. DG

      Oh. Well, actually they are pretty detailed and pretty sophisticated, and they do replicate the neural dynamics. If you take a single neuron and, uh, you, you try to, uh, turn on the different channels, the calcium channels and, uh, uh, the different receptors, uh, and see what effect of, uh, turning on or off those channels are in the neurons' spike output, people have built pretty sophisticated models of that. And, and they are, I, I would say, um, you know, in the regime of correct.

    11. LF

      Well, see, the correctness... That's interesting because you've mentioned at several levels, uh, the correctness is measured by looking at some kind of aggregate statistics.

    12. DG

      It, it would be more, uh, the, the spiking dynamics of a single neuron.

    13. LF

      Spiking dynamics of a single neuron. Okay.

    14. DG

      Yeah. Uh, and, and yeah, these models, because they are, they are going to the level of mechanism, right?

    15. LF

      Mm-hmm.

    16. DG

      So they are basically looking at, uh, okay, what, what is the effect of turning on an ion channel? Uh, and, um, and you can, you can model that using electric circuits, you know? And then so th- their model... S- so it is not just a, uh, function fitting. It is... People are looking at the mechanism underlying it and, uh, putting that in terms of electric circuit, uh, theory, signal propagation theory and, and modeling that. And so those models are sophisticated, but getting a single neuron's model 99% right does not still tell you how to... You know, it would be the analog of getting a transistor model right and now trying to build a microprocessor. Uh, and if you, if you just, uh, observe... you know, if you did not understand how a microprocessor works, uh, but you say, "Oh, I have... I now can model one transistor well, and now I will just try to interconnect, uh, the transistors according to whatever I could, you know, guess from the, the experiments and try to simulate it," um, then it is very unlikely that you will produce a functioning microprocessor. Um, you want to... you know, when you want to, uh, produce a functioning microprocessor, you want to understand Boolean logic, how does... how do the, the gates work, all those things, and then, you know, understand how do those gates get implemented using transistors.

    17. LF

      Yeah. There's actually... I remember... This reminds me, there's a paper, maybe you're familiar with it, that I remember going through in a reading group that approaches a microprocessor from a perspective of a neuroscientist.

    18. DG

      Yeah.

    19. LF

      I think it, it basically... it, uh, uses all the tools that we have of neuroscience to try to understand like as if we just aliens showed up to study computers, uh-

    20. DG

      Yeah.

    21. LF

      ... and, and to see if, if those tools can be used to get any kind of sense of how the microprocessor works. And I think the final... the takeaway from th- at least this initial, uh, exploration is that (laughs) we're screwed. There's no way that the tools of neuroscience would be able to get us to anything, like not even Boolean logic. I mean, it's just a- a- any aspect of the architecture, of the, uh, function, of the processes involved, uh, the, the clocks, the, the timing, all that, you can't figure that out from the tools of neuroscience.

    22. DG

      Yeah. So I'm very familiar with this, this particular paper.

    23. LF

      Yeah.

    24. DG

      I think it was, uh, called, um, Can, uh, a Neuroscientist Understand a Microprocessor or-

    25. LF

      Yeah.

    26. DG

      ... something like that. Following the methodology in that paper, even ele- electrical engineer would not understand microprocessors. So I couldn't have so... (laughs)

    27. LF

      (laughs)

    28. DG

      (laughs) So I co- so I don't think it is that bad in the sense of saying, um, neuroscientists do find valuable things, uh, by observing the brain. Uh, they, they do find good insights. Um, but those insight cannot be put together just as a simulation. You have to, you have to investigate what are the computational underpinnings- pinnings of those findings? How do, do all of them fit together from an information processing perspective? You have to, you have to... Somebody has to, uh, painstakingly put those things together and build hypotheses. Um, so I don't want to diss all of neuroscience as saying, "Oh, they're not finding anything." No, that, you know, that, uh, that paper almost went to that level of, uh, uh, neuroscientists will never understand. Uh, no, that- that's not true. I think they do find lots of useful things, but it has to be put together in a, in a computational framework.

    29. LF

      Yeah. I mean, but, you know, just...... the AI systems will be listening to this podcast a hundred years from now and, and they will probably, uh, th- there's some non-zero probability they'll find your words laughable. It's like, "Oh, remember-"

    30. DG

      (laughs) .

  3. 17:1127:50

    Visual cortex

    1. DG

      cycle.

    2. LF

      So what aspect of the brain are useful in this whole endeavor? Which, by the way, I should say you're, you're both a neuroscientist and an AI person. Uh, I guess the dream is to both understand the brain and to build AGI systems. So you're, it's like an engineer's perspective of trying to understand the brain. So what aspects of the brain, uh, functionally speaking, like you said, do you find interesting?

    3. DG

      Yeah. Quite a lot of things. All right. So one is, um, you know, if you look at the visual cortex, uh, uh, and, and, you, you know, the, the, uh, visual cortex is, is a large part of the brain. Uh, I forgot the ex- exact fraction, but it is, it's a, it- a huge part of our brain area is, uh, occupied by just, just vision. Um, so vision, visual cortex is not just a feedforward cascade of neurons. Um, uh, there are a lot more feedback connections in the brain compared to the feedforward connections. And, and it is surprising to the level of detail neuroscientists have actually studied this. If you, if you go into neuroscience literature and poke around and ask, you know, have they studied what will be the effect of poking a neuron in, uh, level IT, uh, in level V1 and, uh, um, have they studied that? Uh, and you'll say, "Yes, they have studied that." (laughs) .

    4. LF

      (laughs) . So every pos- every possible combination (laughs) -

    5. DG

      Yeah. I mean, it's, it's, uh, it's not a random exploration at all. It's, uh, very hypothesis-driven, right?

    6. LF

      Yeah.

    7. DG

      They, they, they are very, uh, experimental neuroscientists are very, very systematic in how they probe the brain.

    8. LF

      Yeah.

    9. DG

      Uh, because experiments are very costly to conduct. They take a lot of preparation, they, they need a lot of control. So they, they are very hypothesis-driven-

    10. LF

      Yeah.

    11. DG

      ... in how they probe the brain. And, um, often what I find is that when we have a question in, um, in AI, uh, about have, has anybody probed, uh, probed how lateral connections in the brain works? And when you go and read the literature, yes, people have probed it and people have probed it very systematically. And, and they have hypotheses about how those lateral connections are-... supposedly contributing to visual processing. Uh, but, of course, they haven't built very, very functional detail models of it.

    12. LF

      By the way, how do they... In those studies, sorry to interrupt, uh, do they, do they stimulate, like, a neuron in one particular area of the visual cortex and then see how the travel, uh, the signal travels kind of thing?

    13. DG

      Fascinate, very, very fascinating experiments, right? You know, so I can, I can give you one example I was impressed with. Um, this is, um... So before going to that, let me, like, let me give you a, uh, you know, a, uh, overview of how the, the layers in the cortex are organized, right?

    14. LF

      Sure.

    15. DG

      Uh, visual cortex is organized into roughly four hierarchical levels. Okay. So, uh, V1, V2, V4, IT and in V1-

    16. LF

      What happened to V3?

    17. DG

      Uh, well, yeah, there is another pathway. Uh-

    18. LF

      Okay.

    19. DG

      Okay. So there is a... This is-

    20. LF

      (laughs)

    21. DG

      This, I'm, I'm talking about just object recognition pathway.

    22. LF

      All right, cool.

    23. DG

      Okay. Uh, and then, um, in V1 itself, um, i- uh, so it's, there is a very detailed micro circuit in V1 itself. There is, there is organization within a level itself. Um, the cortical sheet is organized into, uh, you know, multiple layers and there are columnar structure. And, and this, this l- layer wise and columnar structure is repeated in V1, V2, V4, uh, uh, IT, all of them, right? Uh, and, and the connections between these layers within a level w- you know, in V1 itself, there are six layers, roughly. And the connections between them, there is a particular structure to them. Uh, and, um, now... So one example of, uh, an experiment, uh, uh, people did is, when I, when you present a stimulus, uh, which is, um, let's say requires, um, separating the foreground from the background of an object. So it is a, it's a textured triangle on a textured background. Uh, and, um, you can check, does the surface settle first or does the contour settle first?

    24. LF

      Settle?

    25. DG

      Settle in the sense that the... So w- when you fi- finally form the percept of the, of the, uh, triangle-

    26. LF

      Mm-hmm.

    27. DG

      ... you understand where the contours of the triangle are, and you also know where the inside of the triangle is, right? That's when you form the final percept. Um, now you can ask, what is the dynamics of forming that final percept?

    28. LF

      Mm-hmm.

    29. DG

      Um, do the, uh, d- do the neurons, um, first find the edges and converge on where the edges are-

    30. LF

      Mm-hmm.

  4. 27:5031:35

    Probabilistic graphical models

    1. LF

    2. DG

      So this is a classic example in, uh, uh, graphical models, probabilistic models. Um, so if you-

    3. LF

      Uh, what are those?

    4. DG

      Uh, okay. Um... (laughs)

    5. LF

      (laughs) Oh, I think it's useful to mention because we'll talk about them more.

    6. DG

      Yeah. Yeah. So neural networks are one class of machine learning models. Um, uh, you know, you have distributed set of, uh, nodes, which are called the neurons. A- you know, each one is doing a dot product. And you can- you can approximate any function using this, uh, multilevel, uh, network of neurons. So that's, uh, uh, a class of models which are used for- useful for function approximation. There is another class of models in machine learning, uh, called probabilistic graphical models. And you can think of them as each node in that model is variable, which is- which is talking about something. You know, it- it can be a variable representing, is- is an edge present in the input or not? Uh, and at the top of the, uh, uh, network, a node can be, uh, representing, is there an object present in the, uh, world or not? And- and then... So it can... It is- it is another way of encoding knowledge.

    7. LF

      Mm-hmm.

    8. DG

      And, uh, um, and then you... Once you encode the knowledge, you can, uh, do inference in the right way, you know. How... What is the best way to, uh, y- you know, explain some set of evidence using this model that you encoded? You know.

    9. LF

      Mm-hmm.

    10. DG

      So when you encode the model, you are encoding the relationship between these different variables. How is the edge connected to my, uh, the model of the object? How is the surface connected to the model of the object? Um, and then, um, of course, this is a very distributed, complicated model. And inference is how do you explain a piece of evidence when- when a set of stimulus comes in? If somebody tells me there is a 50% probability that there is an edge here in this part of the model, how does that affect my belief on whether I should think that there should be a... Is the square present in the image?

    11. LF

      Mm-hmm.

    12. DG

      So- so this is the process of inference. So one example of inference is having this explaining away effect between multiple causes. So, uh, graphical models can be used to represent causality in the world. Um, so let's say, um, y- you know, uh, your, uh, alarm, uh, the, uh, at home can be, uh, triggered by a, uh, burglar getting into your house, uh, or it can be triggered by an earthquake.

    13. LF

      Yeah.

    14. DG

      Both- both can be causes of the alarm going off. So now, you- you are dri- y- you know, y- you're in your office, you heard burglar alarm going off. You are heading, uh, home, uh, thinking that there's a burglar got in.

    15. LF

      Mm-hmm.

    16. DG

      But while driving home, if you hear on the radio that there was an earthquake in the vicinity, now your hype- you know, uh, strength of evidence for, uh, a burglar getting into your house is diminished.

    17. LF

      Mm-hmm.

    18. DG

      Because now that- that piece of evidence is explained by the earthquake in- being present. So if you- if you think about these two causes explaining a lower level, uh, variable, which is alarm, now w- w- what we are seeing is that increasing the evidence for some cause ex- you know, there is evidence coming in from below for alarm being present, and initially it was flowing to a burglar being present. But now since somebody, uh... Some- this evi- there is side evidence for this other cause, it explains away this evidence, and it- evidence will now flow to the other cause. Th- this is, you know, two competing causal, uh, things trying to explain the same evidence.

    19. LF

      And- and the brain has a similar kind of mechanism-

    20. DG

      Yes.

    21. LF

      ... for, uh, for doing so.

    22. DG

      Yes.

    23. LF

      That's kind of interesting. I mean, and that-How's

  5. 31:3536:56

    Encoding information in the brain

    1. LF

      that all encoded (sighs) in the brain? Like, where is the storage of information? Are we talking, just maybe to get it, uh, a little bit more specific, is it in the hardware of the actual connections? Is it in the chemical communication? Is it electrical communication? Do we, do we know?

    2. DG

      So, so this is, you know, a, a paper that we are bringing out soon-

    3. LF

      Which one is this?

    4. DG

      Um, this is the cortical microcircuits paper that-

    5. LF

      Got it.

    6. DG

      ... I sent you a draft of. Of course, this is, uh, a lot of it is still hypothesis. One hypothesis that a, you can think of a cortical column as encoding a, a concept. A concept, you know, think of it as a, uh, a, um, conc- an example of a concept is, um, is an edge present or not? Or is, is an object present or not? Okay. So it can, you can think of it as a binary variable. A binary random variable. The presence of an edge or not, or the presence of an object or not. So, each cortical column can be thought of as representing that one concept, one variable. And then the connections between these cortical columns are basically encoding the relationship between these random variables.

    7. LF

      Mm-hmm.

    8. DG

      And then there are connections within the cortical column. There are... Each cortical column im- is implemented using multiple layers of neurons-

    9. LF

      Mm-hmm.

    10. DG

      ... with very, very, very rich, um, structure there, you know. There are thousands of neurons in a cortical column. Uh-

    11. LF

      But, but that structure is similar across the different cortical columns.

    12. DG

      Correct.

    13. LF

      Yeah.

    14. DG

      Correct. And also, these cortical columns collect, connect to a substructure called thalamus in the, uh, in... So all, all cortical columns pass through this substructure. So, our hypothesis is that, yeah, the connections between the cortical columns implement this, uh, you know, that's where the knowledge is stored about, you know, how these different connups- concepts connect to each other. And then the, the neurons inside this cortical column and in thalamus in combination implement this, um, actual computations needed for inference, which includes explaining a way and competing between the different, uh, hypotheses. Um, and it is all very... So what is amazing is that, um, neuroscientists have actually done ex- experiments to the tune of showing these things. Uh, they might not be putting it in the overall inference framework, but they will show things like if I poke this higher level neuron, uh, it will inhibit, through this complicated loop through the thalamus, it will inhibit this other column.

    15. LF

      Mm-hmm.

    16. DG

      Uh, so they will, they will do such (laughs) experiments.

    17. LF

      But do they use terminology of concepts, for example? So, so you're-

    18. DG

      No.

    19. LF

      I mean, (sighs) uh, is it, uh, is it something where it's easy to anthropomorphize and think about concepts like, uh, you start moving into logic-based kind of reasoning systems. So, um, are we to think of concepts in that kind of way? Or is it s-, uh, is it a lot messier, a lot more gray area? You know, even, even more gray, even more messy than, uh, the artificial neural network kinds-

    20. DG

      Um-

    21. LF

      ... kinds of abstractions?

    22. DG

      Easiest way to think of it as a variable, right? It's a binary variable.

    23. LF

      Mm-hmm.

    24. DG

      Which is showing the presence or absence of something.

    25. LF

      S- but I guess what I'm asking is, is that something, uh, that... Are we supposed to think of something that's human interpretable of that something?

    26. DG

      It doesn't need to be. It doesn't need to be human interpretable. There is no need for it to be human interpretable. Uh, but it's, it's almost like, um, you, you will be able to find some interpretation of it, uh, because it is connected to the other things-

    27. LF

      Yeah. So it's-

    28. DG

      ... that you know about.

    29. LF

      And, and the, the point is it's useful somehow.

    30. DG

      Yeah.

  6. 36:5651:09

    Recursive Cortical Network

    1. LF

      uh, w- what can we say? What is the paper that, uh, you're working on, uh, propose about the ideas around these cortical microcircuits?

    2. DG

      So this is a fully functional model for the microcircuits of the visual cortex.

    3. LF

      So the, the paper focuses... and your idea and our discussions now is focusing on vision.

    4. DG

      Yeah.

    5. LF

      The, uh, visual cortex. Okay.

    6. DG

      Yeah. I-

    7. LF

      This is a model, this is a full model. This is-

    8. DG

      Th- this-

    9. LF

      This is how vision works.

    10. DG

      Uh, well, this is, this is a, yeah, model of-

    11. LF

      A hypothesis.

    12. DG

      Yeah.

    13. LF

      A hypothesis.

    14. DG

      So, okay, so let me, let me step back, uh, a bit. Um, so we looked at neuroscience for insights on how to build a vision model.Right. And, and, and we synthesized all those insights into a computational model. This is called the recursive cortical network model that we, we used for breaking captchas and, uh, and we are using the same model for robotic picking and, uh, uh, tracking of objects.

    15. LF

      And that, again, is a vision system.

    16. DG

      That's a vision system.

    17. LF

      Com- computer vision system.

    18. DG

      That's a computer vision system.

    19. LF

      Takes in images and outputs what?

    20. DG

      On one side, it outputs the class of the image, uh, and also segments the image. Uh, and you can also ask it further queries, "Where is the edge of the object? Where is the interior of the object?" So-

    21. LF

      Got it.

    22. DG

      ... so it's a, it's a model that you build to answer multiple questions. So you're not trying to build a model for just classification or just recommendation et cetera. It's a, it's a, it's a joint model that can do multiple things. Um, and, um, so, so that's the model that we built using insights from neuroscience. And some of those insights are, what is the role of feedback connections? What is the role of lateral connections? Uh, so all those things went into the model. The m- the model actually uses feedback connections.

    23. LF

      All these ideas from neu- from neuroscience.

    24. DG

      Yeah.

    25. LF

      Uh, so what, what, what the heck is a r- recursive cortical network? Like what, what are the architecture approaches, interesting aspects here, which is essentially a brain-inspired approach to computer vision?

    26. DG

      Yeah. So there are multiple layers to this question. Again, go from the very, very top and then zoom in. Okay?

    27. LF

      Mm-hmm.

    28. DG

      So one important thing, constraint that went into the model is that you should not think vision, think of vision as something in isolation. We should not think perception as something, as a pre-processor for cognition.

    29. LF

      Mm-hmm.

    30. DG

      Perception and cognition are interconnected. And so you should not think of one problem in separation from the other problem. Um, and so that means if you finally want to have a system that understand concepts, uh, about the world and can learn a, you know, very conceptual model of the world, and can reason and connect to language, all of those things, you need to, you need to have, think all the way through and make sure that your perception system is compatible with your cognition system and language system and all of them. And one aspect of that is top-down controllability. Um-

  7. 51:091:06:48

    Solving CAPTCHAs algorithmically

    1. DG

      you know, so one of the first applications, uh, that we showed in the paper was to track, uh, text-based CAPTCHAs.

    2. LF

      What are CAPTCHAs, by the way? I mean... (laughs)

    3. DG

      Uh... Yeah. (laughs)

    4. LF

      By the way, one of the most awesome, like, the people don't use this term anymore as human computation, I think. Uh, I love this term. The guy who created CAPTCHAs, I think, came up with this term.

    5. DG

      Yeah.

    6. LF

      I love it. Anyway, uh-

    7. DG

      Yeah. Yeah.

    8. LF

      Uh, wh- what, what are CAPTCHAs?

    9. DG

      So CAPTCHAs are those strings that you fill in, uh, when you're, you know, when- if you're open- opening a new account in Google, they show you a picture. Uh, you know, usually, it used to be set of garbled letters, uh, that you have to kind of, uh, figure out what- what- what is that string of characters and type it. And the reason CAPTCHAs exist is because, you know, um, Google or Twitter do not want automatic creation of accounts. You can use a computer to create millions of accounts, uh, and, uh, use that for, you know, nefarious purposes. Uh, so you want to make sure that, to the extent possible, the interaction that, uh, their system is having is with a human. So it's a- it's called a human interaction proof. A CAPTCHA is a human interaction proof.

    10. LF

      Yeah.

    11. DG

      Um, so- so this is- CAPTCHAs are, by design, things that are easy for humans to solve, but hard for computers.

    12. LF

      Hard for robots, yeah.

    13. DG

      Yeah. Um, so- and text-based CAPTCHAs were- was the one which is prevalent until around 2014. Because at that time, text-based CAPTCHAs were hard for computers to crack. Even now, they are actually, in the sense of an arbitrary text-based CAPTCHA will be unsolvable even now. But with the techniques that we have developed, it can be, you know, you can quickly develop a mechanism that solves all the CAPTCHA, uh-

    14. LF

      They- they've probably gotten a lot harder too. The people-

    15. DG

      Correct.

    16. LF

      (laughs) They've been getting cleverer and cleverer at generating these text CAPTCHAs.

    17. DG

      Correct. Correct.

    18. LF

      Yeah.

    19. DG

      Right.

    20. LF

      So, okay, so that was one of the things you've tested it on, is these kinds of CAPTCHAs-

    21. DG

      Yeah.

    22. LF

      ... in 2014, '15.

    23. DG

      Correct.

    24. LF

      That kind of stuff.

    25. DG

      Right. Right.

    26. LF

      So what, uh ... W- what I mean, why- by the way, why CAPTCHAs? Why?

    27. DG

      Yeah. Yeah. Even now, I would say CAPTCHA is a very- very good challenge problem, uh, if you want to understand how human perception works and if you want to build, uh, systems that work like the human brain. Uh, and I wouldn't say CAPTCHA is a solved problem. We have cracked the fundamental defense of CAPTCHAs, but it is not solved in the way that humans solve it. Um, so I can give you an example. I can, um, take a five-year-old child who has just learned characters, uh, and, uh, show them any new CAPTCHA that we create.

    28. LF

      Mm-hmm.

    29. DG

      They will be able to solve it. Uh, I can show you pretty much any new CAPTCHA, uh, from any new website. You'll be able to solve it without getting any training examples from that particular style of CAPTCHA.

    30. LF

      You're assuming I'm human, yeah.

  8. 1:06:481:18:21

    Hype around brain-inspired AI

    1. DG

      Uh-

    2. LF

      Solving now. Well, let me ask you this kind of touchy question. I have to- I- I've spoken with, uh, your friend, colleague, Jeff Hawkins too. I mean, he's, uh... I ha- I have to kind of ask, there is a bit of... whenever you have brain-inspired stuff-

    3. DG

      Yeah.

    4. LF

      ... and you make big claims-

    5. DG

      Yeah.

    6. LF

      ... uh, big sexy claims-

    7. DG

      Yeah.

    8. LF

      ... there's a, you know, uh, there's critics. I mean, machine learning subreddit.

    9. DG

      (laughs)

    10. LF

      Don't get me started on those people. Uh, they're- they're har- I mean, criticism is good, but they're a bit- uh, they're a bit over the top. Um, there is quite a bit of sort of skepticism and criticism, you know, does this work really as good as it promises to be?

    11. DG

      Yeah.

    12. LF

      What... Do you have thoughts on that kind of skepticism? Do you have comments on the kind of criticism you might have received, uh, about, you know, is this approach legit? Is this- is this a promising approach?

    13. DG

      Yeah.

    14. LF

      Or at least as promising as it seems to be s- you know, advertised as?

    15. DG

      Yeah. I- I can comment on it. Um, so you know, our- our RCN paper is published in Science, which I would argue is- is a very high quality journal, very hard to, uh, publish in. And us- you know, usually it is indicative of the- of the quality of the work. And, um, uh, I can- I can- I- I am very, very certain that the ideas that we brought together in that paper, uh, in terms of the importance of feedback connections, uh, recursive inference, lateral connections, uh, coming to best explanation of the scene as the problem to solve, trying to solve, um, recognition segmentation, uh, all jointly.... in a way that is compatible with higher level cognition, top-down attention, all those ideas that we brought together into something, you know, coherent and workable in the, uh, in the world, and solving a challenging, tackling a challenging problem. I think that will, that will stay and that, that contribution, I stand by, right?

    16. LF

      Yeah.

    17. DG

      Now, uh, I can, I can sh- uh, tell you a story, uh, which is funny in the, in the context of this, right? Um, so if you read the abstract of the paper and, like, you know, the argument we are putting in, we- you know, we are putting in, "Look, current deep learning systems take a lot of training data. Uh, they don't use these insights. And here is our new model, which is not a deep neural network, it's a graphical model. It does inference." This is wha- how the paper is, right? Now, once the paper was accepted and everything, um, it went to the press department in, in Science, you know, to play as science office. We, we didn't do any press release when it was published.

    18. LF

      Yeah.

    19. DG

      It was... it went to the press department. What did the, what was the press release that they wrote up? "A new deep learning model (laughs) so-

    20. LF

      Solves CAPTCHAs.

    21. DG

      Solves CAPTCHAs. And, uh, so, so you can see where was, you know, what, what was being hyped, uh, in that, uh, thing, right? So, so it's like, um, there is the, there is a dynamic in the, uh, in the community of, you know, so, uh, um, that's especially happens when there are lots of new people coming into the field and they get attracted to one thing, and some people are trying to think different, uh, compared to that. So there is, there is some, uh... I, I think skepticism in science is important and it is, eh, eh, um, you know, very much, uh, required, but it's also, it's not, uh, skepticism usually, it's mostly bandwagon effect that is happening, rather than inform-

    22. LF

      Well, well, but that's not even that. I mean, I'll tell you what they react to, which is like, uh, I'm sensitive too as well. If you look, if you look at just companies, OpenAI, DeepMind...

    23. DG

      Yeah.

    24. LF

      Um, Vicarious. I mean, they just, uh, there's a little bit of a race to the top and hype, right?

    25. DG

      Right.

    26. LF

      It's, it's like it doesn't pay off to be humble.

    27. DG

      (laughs) Right. (laughs)

    28. LF

      We... so like, uh... and, and the press is just, uh, irresponsible often. They, they just... I mean, don't get me started on the state of journalism today. Like, it seems like the people who write articles about these things, they literally have not even spent an hour on the Wikipedia article about what is neural networks. Like-

    29. DG

      Yeah.

    30. LF

      ... they haven't, like, invested just even the language, the laziness. It's like, uh, robots beat humans. Like, they, they write this kind of stuff-

  9. 1:18:211:21:32

    How does the brain learn?

    1. LF

      don't really understand the full d- uh, the full, uh, v- biophysics or whatever of how the brain learns.

    2. DG

      E- exactly. Exactly. So-

    3. LF

      But l- let me ask, and I'm sorry to interrupt. Like, eh, d- what's up- what's your sense, what's our best understanding of how the brain learns?

    4. DG

      So things like backpropagation, credit assignment. So, so many of these algorithms-

    5. LF

      Yeah.

    6. DG

      ... have- learning algorithms have things in common, right? It is- uh, yeah, backpropagation is one. We have credit assignment. There is another algorithm called expectation maximization, which is, you know, a- another b- weight adjustment algorithm.

    7. LF

      But is th- your sense the brain does something like this?

    8. DG

      Has to. There is no way around it, in the sense of saying that you do have to adjust the- the connections.

    9. LF

      You- so, yeah, and you're saying credit assignment, you have to reward the connections that were useful in making a correct prediction, and not... Yeah, I guess, what el- but, yeah, it doesn't have to be (laughs) differentiable. I mean... (laughs)

    10. DG

      Yeah, it doesn't have to be differentiable.

    11. LF

      Yeah.

    12. DG

      Yeah. But you have to have a, uh, you know, you have a model that you start with, you ve- you have data comes in, and you have to have a way of adjusting the model such that it better fits the data.

    13. LF

      Yeah.

    14. DG

      So that, that is all of learning. Right? And-

    15. LF

      Yep.

    16. DG

      ... and some of them can be using backprop to do that. Some of it can be using, uh, you know, very local, uh, g- graph changes to do that. Uh, there can b- you know, uh, many of these learning algorithms have similar update properties locally, i- in terms of what the neurons need to do locally.

    17. LF

      I wonder if small differences in learning algorithms can have huge differences in the actual effect. So the dynamics of, I mean, uh, sort of the rev- the reverse like spiking, like, the, uh, i- if credit assignment is like, uh, a lightning versus like, uh, a rainstorm or something. Like, uh, whether, whether there's a f- like a looping local type of situation with the credit assignment.

    18. DG

      Yeah.

    19. LF

      Uh, whether there is, uh, like regularization, like how, how, um, how it injects robustness into the whole thing.... like, whether it's chemical or electrical or mechanical-

    20. DG

      Yeah.

    21. LF

      ... uh, all those kinds of things.

    22. DG

      Yes. (laughs)

    23. LF

      Like that, I, I feel like it, it that, yeah, I feel like those differences could be essential, right?

    24. DG

      It could be. It's just that you don't know enough to, on the learning side, you don't know, uh, enough to say, "That is definitely not the way the brain does it."

    25. LF

      Right. Got it.

    26. DG

      Um-

    27. LF

      So you don't wanna be stuck to it.

    28. DG

      Right.

    29. LF

      So that, yeah, so you, you've been open-minded on that side of things.

    30. DG

      Correct. On the inference side, on the recognition side, I am much more, uh, amenable to being constrained, because I- it's much easier to do experiments because, you know, it's like, okay, here is a stimulus, you know, wha- how many steps did it get to take the answer?

  10. 1:21:321:25:43

    Perception and cognition

    1. DG

    2. LF

      So let, let's, let's go right into cortical microcircuits right back. So what, uh, what are these ideas beyond recursive cortical network that, uh, you're looking at now?

    3. DG

      So we have made a, uh, you know, pass through all, you know, multiple of the steps that we, uh, you know, as I, as I mentioned earlier, you know, we were looking at perception from the angle of cognition, right? It was not just perception for perception's sake. How do you, how do you connect it to cognition? Uh, how do you learn concepts and, uh, how do you learn abstract reasoning? Uh, similar to some of the things Francois, uh, uh, talked about, right? Um, so, um, so we have, uh, taken one pass through it basically saying, "What is the basic cognitive architecture that you need to have, which has a perceptual system, which has a system that learns dynamics of the world, and then has something like a routine program learning system on top of it, to learn concepts?" So we have, we have built one, the, you know, the version 0.1 of that system. Uh, this was another, uh, science robotics paper. Uh, it is, it's the title of that paper was, you know, something like Cognitive Programs. How do you build cognitive programs? Uh, I'm-

    4. LF

      And the application there was on m- uh, manipulation, robotic manipulation?

    5. DG

      It was, it was, um, so think of it like this. Suppose you, uh, wanted to tell, uh, a new person, uh, that you met, you don't know the language, uh, uh, that person uses. You want to communicate to that person, uh, to achieve some task.

    6. LF

      Mm-hmm.

    7. DG

      Right? So I want to say, "Hey, um, you need to pick up all the, the red cups from the kitchen counter and put it here."

    8. LF

      Mm-hmm.

    9. DG

      Right? Uh, how do you communicate that, right? You can show pictures. You can basically say, "Look, this is the starting state. Uh, the, the, the things are here, this is the ending state." And, and what does the person need to understand from that? The, the person need to understand what conceptually happened in those pictures from the input to the output, right?

    10. LF

      Yeah. Right.

    11. DG

      Um, so, um, so we are looking at pre-verbal conceptual understanding. Without language, how do you, how do you have a set of concepts that you can manipulate in your head? Uh, and from a s- you know, set of images of input and output, can you infer what is happening in those images?

    12. LF

      Got it. With concepts that are pre-language. Okay.

    13. DG

      Yeah.

    14. LF

      So wha- what's it mean to, for a concept to be pre-language? Like-

    15. DG

      Yeah.

    16. LF

      ... why, why is so, why (laughs) why is language, uh, so important here?

    17. DG

      So I, I want to make a distinction between concepts that are, are just learned from text-

    18. LF

      Ah.

    19. DG

      ... by, by just, just feeding brute force text. Uh, you can, you can start extracting things. Like, okay, a cow is likely to be on grass, uh, in a, in a-

    20. LF

      Got it.

    21. DG

      Uh, so those kinds of things you can extract purely from text. Um, uh, but that's kind of a simple association, uh, thing, rather than a concept as an abstraction of something that happens in the real world, uh, in a, in a, in a grounded way that I can, I can simulate it in my mind and connect it back to the real world.

Episode duration: 2:10:05

Install uListen for AI-powered chat & search across the full episode — Get Full Transcript

Transcript of episode tg_m_LxxRwM

Get more out of YouTube videos.

High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.

Add to Chrome