Lex Fridman PodcastJeffrey Shainline: Neuromorphic Computing and Optoelectronic Intelligence | Lex Fridman Podcast #225
EVERY SPOKEN WORD
150 min read · 30,051 words- 0:00 – 0:44
Introduction
- LFLex Fridman
The following is a conversation with Jeff Schoenlein, a scientist at NIST interested in optoelectronic intelligence. We have a deep technical dive into computing hardware that will make Jim Keller proud. I urge you to hop on to this roller coaster ride through neuromorphic computing and superconducting electronics, and hold on for dear life. Jeff is a great communicator of technical information, and so it was truly a pleasure to talk to him about some physics and engineering. To support this podcast, please check out our sponsors in the description. This is the Lex Fridman Podcast, and here is my conversation with Jeff Schoenlein.
- 0:44 – 20:02
How are processors made?
- LFLex Fridman
I got a chance to read a fascinating paper you, um, authored called Optoelectronic Intelligence. So maybe we can start by, uh, talking about this paper, and start with the basic questions. What is optoelectronic intelligence?
- JSJeffrey Shainline
Yeah, so in that paper the, the concept I was trying to describe is sort of an architecture for building brain-inspired computing that leverages light for communication in conjunction with electronic circuits for computation. In that particular paper, a lot of the work we're doing right now in our project at NIST is focused on superconducting electronics for computation. I'll go into why that is, but, uh, that might make a little more sense in context if we first describe what that is in contrast to, which is semiconducting electronics.
- LFLex Fridman
Mm-hmm.
- JSJeffrey Shainline
So is it worth taking a couple minutes to describe semiconducting electronics?
- LFLex Fridman
It might even be worthwhile to step back and, uh, talk about electricity and circuits and how circuits work-
- JSJeffrey Shainline
Right.
- LFLex Fridman
... uh, before we talk about superconductivity.
- JSJeffrey Shainline
Right. Okay.
- LFLex Fridman
How does a computer work, Jeff?
- JSJeffrey Shainline
Well, I, I won't go into everything-
- LFLex Fridman
(laughs)
- JSJeffrey Shainline
... that makes a computer work, but let's talk about the basic building blocks, a transistor, so... And even more basic than that, uh, a semiconductor material, silicon, say. So, uh, in, in silicon, silicon is a semiconductor, and what that means is at low temperature there are no free charges, no free electrons that can move around. So when you talk about electricity, you're talking about predominantly electrons moving to establish electrical currents, and they move under the influence of voltages. So you apply voltages, electrons move around. Those can be measured as currents, and you can represent information in that way. So semiconductors are special in the sense that they are really malleable. So if you have a, a semiconductor material, it, you can change the number of free electrons that can move around by putting different elements, different atoms in lattice sites. So what is a lattice site? Well, a semiconductor is a crystal which means all the atoms that comprise the material are at exact locations that are perfectly periodic in space. So if you start at any one atom and you go along the what are called the lattice vectors, you get to another atom and another atom and another atom, and for high-quality devices it's important that that is a, a perfect crystal with very few defects. But you can intent- intentionally replace a silicon atom with, say, a phosphorus atom, and then you can, you can change the number of free electrons that are in a region of space that has that excess of what are called dopants. So picture a device that has a left terminal and a right terminal, and if you apply a voltage between those two, you can cause electrical current to flow between them. Now we, uh, add a third terminal up on top there, and depending on the voltage between the left and right terminal and that third voltage, you can, you can change that current. So what's commonly done in digital electronic circuits is to leave a fixed voltage from left to right and then change that voltage that's applied at what's called the gate, the gate of the transistor. So, um, what you do is you, you make it to where there's an excess of electrons on the left, excess of electrons on the right, and very few electrons in the middle, and you do this by changing the concentration of different dopants in the lattice spatially. And then when you apply a voltage to that gate, you can either cause current to flow or turn it off, and so that's sort of your zero and one. You, if you apply voltage, current can flow. That current is representing a digital one, and, uh, from that, from that basic element, you can build up all the complexity of digital electronic circuits that have really had a profound influence on our society.
- LFLex Fridman
Now you're talking about electrons. Can you give a sense of what scale we're talking about when we're talking about in silicon, uh, being able to mass manufacture these kinds of, uh, gates?
- JSJeffrey Shainline
Yeah, so scale in a number of different senses. Well, at the scale of the silicon lattice, the, the distance between two atoms there is half a nanometer, so, um, people often like to compare these things to the, the width of a human hair. I think it's some six orders of magnitude smaller than the width of a human hair, uh, som- something on that order. So remarkably small. We're talking about individual atoms here, and electrons are of that length scale when they're in that environment. But there's another sense that scale matters in digital electronics. This is perhaps the more important sense, although they're related. Scale refers to a number of things. It refers to the size of that transistor. So, for example, I said you have a left contact, a right contact, and some space between 'em where the, the gate electrode sits. That, that's called the, the channel width, uh, or the channel length, and, um, what has enabled what we think of as Moore's law or the continued...... increased performance in silicon microelectronic circuits is the ability to make that size, that feature size, ever smaller, ever smaller at a, at a, uh, r- really remarkable pace. I mean, that h- that feature size has decreased, uh, consistently every couple of years for the l- since the 1960s. And that was, that was what Moore predicted in the 1960s. He thought it would continue for at least two more decades, and it's been much longer than that. And so, um, that is why we've been able to fit evermore devices, evermore transistors, evermore computational power on essentially the same size of chip. So, a user sits back and does essentially nothing. You're running the same computer program, but those devices are getting smaller, so they get faster, they get more energy-efficient. And all of our computing performance just continues to improve. And we don't have to think too hard about what we're, what we're doing as, say, uh, a software designer or someone like that. I- I absolutely don't mean to say that there's no innovation in software, the, or the user side of things. Of course there is. But from, from the hardware perspective, we just have been given this gift of continued performance improvement through this scaling that is ever-smaller feature sizes with very similar, um, say, power consumption. That power consumption is, has not continued to scale in the most recent decades. But, um, nevertheless, we had a really good run there for a while, and now we're down to gates that are seven nanometers, which is state-of-the-art right now, uh, maybe GlobalFoundries is trying to push it even lower than that. I- I can't keep up with where the predictions are that it's gonna end. But seven-nanometer, uh, seven-nanometer transistor has just, just a few tens of atoms along the length of the conduction pathway. So, a naive (laughs) semiconductor device physicist would think you can't go much further than that without some kind of revolution in the way we think about the physics of our devices.
- LFLex Fridman
Is there something to be said about the mass manufacture of these devices?
- JSJeffrey Shainline
Right, right. So, that's another thing. So, how have we been able to make those transistors smaller and smaller? Well, companies like Intel, GlobalFoundries, they invest a lot of money in the lithography. So, how- how are these, uh, chips actually made? Well, one of the most important steps is this, what's called ion implantation. So, you have, you start with sort of a pristine silicon crystal, and then using photolithography, which is a technique where you can pattern different shapes using light, you can define which regions of space you're going to implant with different, uh, different species of ions that are going to change the local electrical properties right there. So, by using ever-shorther wavelengths of light and different kinds of optical techniques and different kinds of lithographic techniques, things that go far beyond, uh, my knowledge base, you can just simply shrink that feature size down. And you say you're at seven nanometers. Well, the wavelength of light that's being used is over 100 nanometers. That's already deep in the UV, so, um, how, how are those minute features patterned? Well, there's- there's an extraordinary amount of innovation that has gone into that. But nevertheless, it's stayed very consistent in this ever-shrinking feature size, and now the question is, can you make it smaller? And even if you do, do you still continue to get performance improvements? But- but that's another kind of scaling, where these- these companies have been able to... S- so, okay, you- you picture a chip that has a processor on it. Well, that chip is not made as a chip. It's made as a, on a wafer.
- LFLex Fridman
Mm-hmm.
- JSJeffrey Shainline
And, um, using photolithography, you basically print the same pattern ac- on different dies all across the wafer, multiple layers, tens, probably, probably h- uh, a hundred-some layers in a mature foundry process. And y- and you do this on ever-bigger wafers, too. That's another aspect of scaling that's occurred in the last several decades. So, now you have this 300-millimeter wafer. It's like as big as a pizza, and it has maybe 1,000 processors on it, and then you dice that up using a saw. And now you can sell these things f- so cheap, because the- the manufacturing process was so streamlined. I think a technology as revolutionary as silicon microelectronics has to have that kind of, uh, manufacturing scalability, which I will just emphasize, I believe is enabled by physics. It's not, I mean, the, of course, there's human ingenuity that goes into it. But at least from my, my side, where- where I sit, it sure looks like the physics of our universe allows us to, to produce that, and we've- we've discovered how more so than we've invented it, although, of course, we have invented it. Humans have invented it. But it was, it's almost as if it was there waiting for us to, to discover it.
- LFLex Fridman
You mean the entirety of it, or are you specifically talking about the techniques of, uh, photolithography and, like, the optics involved? You mean, the- the entirety of the scaling down to the seven nanometers, that you're able to have electrons not interfere with each other in such a way that y- you could still have gates. Like, that's enabled... To achieve that scale, spatial and temporal, seems to be very special, and is enabled by the physics of our world.
- JSJeffrey Shainline
All of the things you just said. So, starting with the- the silicon material itself, silicon is a- a unique semiconductor. It has essentially ideal properties for making a specific kind of transistor that's extraordinarily useful. So, I mentioned that silicon has, uh, th- well, when you make a transistor, you have this gate contact that sits o- on top of the conduction channel, and depending on the voltage you apply there, you pull more carriers into the conduction channel or push them away so it becomes more or less conductive. In order to have that work without just sucking those carriers right into that contact, you need a- a very thin insulator. And- and part of scaling has been to gradually decrease the thickness of that, of that gate insulator so they can use a- a roughly similar voltage and still have the same current-voltage characteristics. So, the material that's used to do that, or I should say was initially used to do that, was, uh, silicon dioxide, which just naturally grows on the silicon surface. So, you expose silicon to the atmosphere that we breathe, and, uh, well, if you're manufacturing, you're gonna...... purify these gases. But nevertheless, that, that, what's called a native oxide will grow there. There are essentially no other materials on the entire periodic table that have as good of a gate insulator as, as that silicon dioxide. And that, that has to do with nothing but the physics of the interaction between silicon and oxygen. And if it wasn't that way, transistors could not, they, they could not perform in nearly the d- the degree of capability that they have. And that, that has to do with the way that the, the oxide grows, the reduced density of defects there. Its, its insulation, meaning essentially its energy gaps, you can apply a very large voltage there without having current leak through it. So, that's physics right there. Um, there, there are other things too. Silicon is a semiconductor in, in an elemental sense. You, you only need silicon atoms. A lot of other semiconductors, you need two different kinds of atoms, like a compound from group three and a compound from group five. That opens you up to lots of defects that can occur, where one atom's not sitting quite at the lattice site it is, and it's switched with another one. That degrades performance. Um, but then also on the side that you mentioned with the, the manufacturing, we have access to light sources that can produce these very short wavelengths of light. Um, how does photolithography occur? Well, you actually put this polymer on top of your wafer, and you expose it to light, and then you use a aqueous chemical processing to dissolve away the regions that were exposed to light and leave the regions that were not. And we are blessed with these polymers that have the right property where they can, um, cause scission events, where the polymer splits where a photon hits. I mean, you know, maybe, maybe that's not too surprising, but I don't know. It all, it all comes together to have this really complex, uh, manufacturable ecosystem where very sophisticated technologies can be devised and, uh, it- it works quite well.
- LFLex Fridman
And, and amazingly, like you said, with a wavelength at like 100 nanometers or something like that, you're still able to achieve, on this polymer, precision of whatever, whatever we said, seven nanometers.
- JSJeffrey Shainline
Yeah.
- LFLex Fridman
I think I've heard like four nanometers being talked about, something like that.
- JSJeffrey Shainline
Yes.
- LFLex Fridman
All right. If we could just pause on this, and we'll return to s- superconductivity. But in this whole journey, from a history perspective, what, what do you think is the most beautiful, at the intersection of engineering and physics, to you, in this whole process that we talked about with silicon and photolithography, things that people were able to achieve in order to, uh, push, uh, Moore's Law forward? Is it the early days, the, the invention of the transistor itself? Is it, uh, some particular cool little thing that, um, maybe not many people know about? Like, what do you think is most beautiful in this e- in this whole process, journey?
- JSJeffrey Shainline
The most beautiful is a little difficult to answer.
- LFLex Fridman
(laughs)
- JSJeffrey Shainline
Let me, let me try and sidestep it a little bit and just say, what strikes me about looking at the, the history of silicon microelectronics is that, uh, so when, when quantum mechanics was developed, people quickly began applying it to semiconductors. And it was broadly understood that these are fascinating systems, and people cared about 'em for their basic physics, but also their utility as devices. And then the transistor was invented in the late '40s, um, in a (laughs) relatively crude experimental setup, where you just crammed a metal electrode into the semiconductor. And, and that was, that was ingenious. These people were able to, um, make it work, you know? Uh, but so what, what I want to get to that, that really strikes me is that in those early days, there, there were a number of different semiconductors that were being considered. They had different properties, different strengths, different weaknesses. Most people thought germanium was the, the way to go. It, it, it had some, some nice properties, uh, related to things a- about how the electrons move inside the lattice. But other people thought that compound semiconductors, with group three and group five, also had really, really extraordinary, um, properties that might be conducive to, to making the best devices. So there were different groups exploring each of these, and that's great. That's how science works. You have to cast a broad net. But then w- I, what I, what I find striking is why, why is it that silicon won? Because it's not that, it's not that germanium is a useless material and it's not present in technology or compound semiconductors. They're both doing k- doing exciting and important things, slightly more niche applications, whereas silicon is the semiconductor material for microelectronics, which is the platform for digital computing, which has transformed our world.
- LFLex Fridman
Mm-hmm.
- JSJeffrey Shainline
Why did silicon win? It's because of a remarkable assemblage of qualities that no one of them was the clear winner, but it, it made these sort of compromises between a number of different influences. It had that really excellent, um, gate oxide that allowed it to, uh, that allowed us to make MOSFETs, these high performance transistors, so quickly and cheaply and easily without having to do a lot of materials development. The, the band gap of silicon, um, is actually ... So th- in a semiconductor, there's, there's an important parameter which is called the band gap, which tells you, uh, if you, there are, there are sort of electrons that fill up to one level in, in the energy diagram, and then there's a gap where electrons aren't allowed to have an energy in a certain range, and then there's another energy level above that. And that, that difference between the lower sort of filled level and the unoccupied level, that tells you how much voltage you have to apply in order to induce a current to flow. So with germanium, that's about 0.75 electron volts. Uh, that means you have to apply 0.75 volts to, to get a current moving.And it turns out that if you compare that to the- the- the thermal excitations that are induced just by the temperature of our environment, that gap's not quite big enough. You start to use it to perform computations, it gets a little hot and you get all these accidental carriers that are excited into the- the conduction band, and it- it causes errors in your computation. Silicon's bandgap is just a little higher, 1.1 electron volts, but you have an exponential dependence on the- the nu- the number of carriers that are present that can induce those errors, uh, it decays exponentially with that voltage. So just that- that slight extra energy in that bandgap really puts it in an ideal position to be operated in the- in the conditions of our- of our ambient environment.
- 20:02 – 22:31
Are engineers or physicists more important
- JSJeffrey Shainline
exactly right.
- LFLex Fridman
So this also- this is gonna be the most controversial part of our conversation where you're gonna make some enemies. So let me ask, 'cause we've been talking about physics and engineering, wh- which group of people is smarter and more important for this one? (laughs) Let me ask i- the question in a better way. Some of the big innovations, some of the beautiful things that we've been talking about, how much of it is physics? How much of it is engineering? My dad is a physicist, and he talks down to all the amazing engineering that we're doing in, um, the artificial intelligence and the computer science and the robotics and all that space. So, we argue about this all the time. So what do you think? Who gets more credit?
- JSJeffrey Shainline
I- I-
- LFLex Fridman
Who gets more credit?
- JSJeffrey Shainline
I'm genuinely not trying to just be politically correct here.
- LFLex Fridman
(laughs)
- JSJeffrey Shainline
I don't see how you would have any of the what we consider sort of the great accomplishments of society without both. It-
- LFLex Fridman
Okay.
- JSJeffrey Shainline
You absolutely need both of those things. Physics tends to play a key role earlier in the development, and then engineering optimization, these things take over, and, uh, I mean, the invention of the transistor or actually even before that, the- the understanding of semiconductor physics that allowed the invention of the transistor, that's all physics. So if you didn't have that physics, you're- you don't even get to get on the- on the- on the field. But once you have understood and demonstrated that this is in principle possible, Moore's Law is engineering. That-
- LFLex Fridman
Right.
- JSJeffrey Shainline
Wh- why we have, uh, computers more powerful than- than old supercomputers in ou- in each of our phones is- that's all engineering. And I- I think I would be quite foolish to say that, (laughs) that's a f- I mean, that- that's not valuable, that-
- LFLex Fridman
Yeah.
- JSJeffrey Shainline
... that's not a great contribution.
- LFLex Fridman
Uh, it's a beautiful dance. Would you put, like, silicon, uh, the understanding of the material properties in the space of engineering? Like, how does that whole process work? To understand that it has all these nice properties or even the development of, uh, uh, photolithography, is- is that basically- would you put that in the category of engineering?
- JSJeffrey Shainline
No. I would say that it is basic physics. It is applied physics. It's material science. It's, um, x-ray crystallography. It's polymer chemistry. It's- it's everything.
- LFLex Fridman
Okay.
- JSJeffrey Shainline
I mean-
- LFLex Fridman
So chemistry even is thrown in there?
- JSJeffrey Shainline
Absolutely. Yes.
- LFLex Fridman
Okay. The whole thing.
- JSJeffrey Shainline
Yes. Absolutely.
- LFLex Fridman
Just no biology.
- JSJeffrey Shainline
(laughs)
- LFLex Fridman
(laughs)
- JSJeffrey Shainline
Oh, except we can get to biology, yeah-
- LFLex Fridman
Well, the biology is in the humans that are engineering the systems.
- JSJeffrey Shainline
Absolutely. Right.
- LFLex Fridman
It's all integrated deeply.
- 22:31 – 38:18
Super-conductivity
- LFLex Fridman
Okay, so let's return. You mentioned this, uh, word superconductivity. So what does that have to do with what we're talking about?
- JSJeffrey Shainline
Right. Okay. So in a semiconductor, as I, uh, tried to describe a second ago, you can sort of, uh, in- induce currents by applying voltages, and those have sort of typical properties that you would expect from some kind of a conductor. Those electrons, they don't just flow perfectly without dissipation. If an electron collides with an imperfection in the lattice or another electron, it's gonna slow down. It's gonna lose its momentum. So you have to keep applying that voltage in order to- to keep the current flowing. In a superconductor, something different happens. If you get a current to start flowing, it will continue to flow indefinitely. There's- there's no dissipation. So that's crazy. How does that happen? Well, it happens at low temperature, and this is crucial. It has to- it has to be a- a- a quite low temperature, and what- what I'm talking about there, I... for es- essentially all of our conversation, I'm gonna be talking about conventional superconductors, um, sometimes called low-TC superconductors, low critical temperature superconductors. And so those materials have to be in- at a temperature around- say, around four Kelvin. I mean, their critical temperature might be 10 Kelvin, something like that, but you want to operate them at around four Kelvin, four degrees above absolute zero. And what happens at that temperature, at- at very low temperatures in certain materials, is that the- the noise of atoms moving around, the lattice vibrating, electrons colliding with each other, that becomes sufficiently low that the electrons can settle into this very special state. It's sometimes referred to as a macroscopic quantum state because if I had a- a piece of superconducting material here, let's say niobium is a very typical, um, superconductor, if I- if I had a block of niobium here, and we cooled it below its critical temperature-All of the electrons in that, in that superconducting state would be in one coherent quantum state. They would-
- LFLex Fridman
Oh.
- JSJeffrey Shainline
...the, the wave function of that state is described in terms of all of the particles simultaneously. But it extends across macroscopic dimensions. The size of a, whatever material, the size of whatever block of that material I have sitting here. And the way that, the way this occurs is that, you know, we, we ... let's try to be a little bit light on the technical details, but essentially, the electrons coordinate with each other. They, they are able to, in this macroscopic quantum state, they're able to sort of ... one can quickly take the place of the other. You can't tell electrons apart. They're, they're what's known as identical particles, so if this electron runs into a, a defect that would otherwise cause it to scatter, it can just sort of, um, almost miraculously avoid that defect, because it's not really in that location. It's part of a macroscopic quantum state, and the entire quantum state was not scattered by that defect. So, you can get a, a current that flows without dissipation, and that's called a supercurrent.
- LFLex Fridman
Mm-hmm.
- JSJeffrey Shainline
That's, uh, sort of just very much scratching the surface of, of superconductivity. It ... there, there's very deep and rich physics there, which is probably not the main subject we need to go into right now. But it turns out that when you have this material, you can, you can do usual things, like make wires out of it, so you can get current to flow in a, in a straight line on a chip, but you can also make other devices that perform different kinds of operations. Some of them are kind of logic operations, like you'd, like you'd get in a transistor. The most common or most, um, I would say diverse in its utility, uh, component is a Josephson junction. It's not analogous to a transistor in the sense that if you apply a voltage here, it changes how much current flows from left to right. But it is analogous in sort of a, a sense of ... it's the, it's the go-to component that a, that a circuit engineer is going to use to start to build up more complexity.
- LFLex Fridman
So, these are, uh, these junctions serve as gates?
- JSJeffrey Shainline
They can, they can serve as gates. They can ... so, I'm not sure how, how s- um, concerned to be with semantics, but let me just briefly say what a Josephson junction is-
- LFLex Fridman
Yeah, that would be great.
- JSJeffrey Shainline
...and we can talk about different ways that they can be used. Basically, if you have a, a superconducting wire, and then a, a small gap of, uh, a different material that's not superconducting, an insulator or normal metal, and then another superconducting wire on the other side, that's a Josephson junction. So, it's sometimes referred to as a superconducting weak link. So, you have this superconducting state on one side, and on the other side, and that ... the superconducting wave function actually tunnels across that gap. And when you, when you create such a physical entity, it has very unusual, um, current voltage characteristics.
- LFLex Fridman
With a ... in, in that gap. Like, like weird stuff happens.
- JSJeffrey Shainline
Through the entire circuit.
- LFLex Fridman
Yeah.
- JSJeffrey Shainline
So, you can imagine, suppose you had a loop set up that had one of those weak links in, in the loop. Current would flow in that loop, independent ... even if you hadn't applied a voltage to it, and that's called the Josephson effect. So, the fact that there's this phase difference in the quantum wave function from one side of the tunneling barrier to the other induces current to flow.
- LFLex Fridman
So, how does you change state in a ... ?
- JSJeffrey Shainline
Right, exactly. So, how do you change state? Now, picture, if I have a, uh, current bias coming down this line on my circuit, and then there's a Josephson junction right in, in the middle of it, and now I'd, I make another wire that goes around the Josephson junction, so I have a loop here, a superconducting loop. I can add current to that loop by exceeding the critical current of that Josephson junction. So, like any superconducting material, it can carry this supercurrent that I've described, this current that can propagate without dissipation, up to a certain level. And if you try and pass more current than that through the material, it's going to become, uh, a resistive material, a normal, normal material. So, in the, in the Josephson junction, the same thing happens. I can bias it above its critical current, and then what it's gonna do, it's going to add, uh, a quantized amount of current into that loop. And what I mean by quantized is, it's going to come in discrete packets with a well-defined value of current.
- LFLex Fridman
Mm-hmm.
- JSJeffrey Shainline
So, in the vernacular of, of some people working in this community, you would say you pop a fluxon into the loop. So, a fluxon-
- LFLex Fridman
(laughs) You pop a fluxon into the loop?
- JSJeffrey Shainline
Yeah. So, a fluxon-
- LFLex Fridman
Sounds like skateboarder talk. I love it.
- JSJeffrey Shainline
(laughs)
- LFLex Fridman
Okay. Co- uh, sorry. Go ahead. (laughs)
- JSJeffrey Shainline
A fluxon is one of these quantized, uh, sort of, um, amounts of current that you can add to a loop. And, and this is a cartoon picture, but I think it's-
- LFLex Fridman
Mm-hmm.
- JSJeffrey Shainline
... sufficient for our purposes.
- LFLex Fridman
So, which, uh, maybe is useful to say, what is the speed at which these discrete packets of current travel? Because we'll be talking about light a little bit. It seems like the speed is important.
- JSJeffrey Shainline
The speed is important. That's an excellent question. Sometimes I wonder where you ... how you became so astute.
- LFLex Fridman
(laughs)
- JSJeffrey Shainline
But, um, so-
- 38:18 – 42:55
Computation
- LFLex Fridman
you used the word computation. When you talk about computation, how do you think about it? Do you think purely, uh, on just, um, the g- the switching, or do you think something a little bit larger scale, a circuit taken together performing the basic arithmetic, uh, operations that are then required to do the kind of computation that makes up a computer? 'Cause when we talk about the speed of computation, does it boil down to the basic switching or is there some bigger picture that you're thinking about?
- JSJeffrey Shainline
Well, all right, so maybe we should disambiguate. There are a variety of different kinds of computation. I don't pretend to be an expert in the theory of computation or anything like that. I guess it's important to, to differentiate, though, between digital logic, which represents information as a series of bits, binary digits, which, you know, uh, you can think of them as zeros and ones or whatever. Usually, they correspond to a physical system that has two very well separated states. And then other kinds of computation, like we'll get into more the way your brain works, which it is, I, I think, indisputably processing information. But where the computation begins and ends is not anywhere near as well defined. It, it doesn't depend on these two levels, here's a zero, here's a one. It's, there's a lot of gray area that's usually referred to as analog computing. Um, also in, in conventional digital computers or, um, digital computers in, in general, you have a concept of what's called arithmetic depth, which is jargon that basically means how many sequential operations are performed to turn an input into an output. And those kinds of computations in, in digital systems are, are highly serial, meaning that data streams, they don't branch off too far to the side. You, you do, you have to pull some information over there and access memory from here and stuff like that. But by and large, the, the computation proceeds in a, in a serial manner. It's not that way in the brain. In the brain, you're always drawing information from different places. It's much more network-based computing. N- Neurons don't wait for their turn. They fire when they're ready to fire, and so it's, it's asynchronous. So one of the other things about a, a digital system is you're performing these operations on a clock, and that's a, that's a crucial aspect of it. Get rid of a clock in a digital system, nothing makes sense anymore. The brain has no clock. It, it builds its own timescales based on its internal activity. So...
- LFLex Fridman
So, you can think of the brain as kind of, um, l- like this, like, network computation where it's actually really trivial, simple computers, uh, just a huge number of them, and they're networked, uh, together.
- JSJeffrey Shainline
I would say it is complex, sophisticated little processors, and there's a huge number of them.
- LFLex Fridman
Sure, sure. Sorry.
- JSJeffrey Shainline
Neurons are not, are not-
- LFLex Fridman
No, no offense. I don't mean to offend neurons.
- JSJeffrey Shainline
Sure, no, no.
- LFLex Fridman
They're very complicated and beautiful, and, yeah. But, w- we often oversimplify them. Yes, they're actually ext- Like, there's computation happening within a neuron, right?
- JSJeffrey Shainline
Right. So, I, I would, uh, say, to think of a, a transistor as the building block of a digital computer is accurate. You use a few transistors to make your logic gates. You build up more. You build up processors from logic gates and things like that. So you can think of a transistor as a fundamental building block, or you can think of, as we get into more, uh, highly parallelized architectures, you can think of a processor as a fundamental building block. To make the analogy to the neuro side of things, a neuron is not a transistor. A neuron is a, is a processor.
- LFLex Fridman
Yeah.
- JSJeffrey Shainline
It has synapses. Even synapses are not transistors, but they are more, um, l- they're lower on the information processing hierarchy in a sense. They do a, a bulk of the computation, but neurons are entire, um, processors in and of themselves that can take in many different kinds of inputs on many different spatial and temporal scales and produce l- many different kinds of outputs so that they can perform different computations in different contexts.
- LFLex Fridman
So, this is where enters this distinction between computation and communication. So, you can think of neuron as performing computation, and the inter- the networking, the interconnectivity of neurons is communication between neurons. And you see this with very large server systems. I've been... I mentioned offline, we were talking to Jim Keller, who dreams to build giant computers that, uh, you know... The, the bottleneck there is often the communication between the different-
- JSJeffrey Shainline
Yes.
- LFLex Fridman
... pieces of computing. So,
- 42:55 – 46:36
Computation vs communication
- LFLex Fridman
in this paper that we, uh, mentioned, Optoelectronic Intelligence, you say, uh, "Electrons excel at computation."... while light, eh, is excellent for communication. Maybe you can linger and say, in this context, what he mean by computation and communication, what's, what are electrons, what is light, and why do they excel at those two tasks.
- JSJeffrey Shainline
Yeah. Just to, to first speak to computation versus communication, I would say computation is essentially taking in some information, performing operations on that information, and producing new, hopefully more useful, information. So, for example, um, imagine you have a, a picture in front of you and there is a, a key in it, and that's what you're looking for, for whatever reason. You wanna, you wanna find the key. We all-
- LFLex Fridman
Mm-hmm.
- JSJeffrey Shainline
... wanna find the key. So-
- LFLex Fridman
Yeah.
- JSJeffrey Shainline
... um, the input is that, that entire picture, and the output might be the coordinates where the key is. So, you've reduced the total amount of information you have, but you found the useful information-
- LFLex Fridman
Mm-hmm.
- JSJeffrey Shainline
... for you in that present moment. That's the useful information.
- LFLex Fridman
And you think about this computation a- as, like, controlled synchronous sequential?
- JSJeffrey Shainline
Not necessarily. It could be. That could be how w- your system is performing the computation, or it could be asynchronous. It, there are lots of ways to find the key.
- LFLex Fridman
Okay.
- JSJeffrey Shainline
It depends, it depends on the nature of the data, it depends on, um... That's a very simplified example, a picture with a key in it. What about if you're in the world and you're trying to decide the best way to, um, live your life, you know? That... (laughs)
- LFLex Fridman
It might be interactive, it might be, there might be some recurrence or something-
- JSJeffrey Shainline
Yeah, sounds-
- LFLex Fridman
... weird. Asynchrony, I got it. So, but, there's an input, there's an output, and you do some stuff in the middle that-
- JSJeffrey Shainline
Yeah.
- LFLex Fridman
... actually goes from the input to the output.
- JSJeffrey Shainline
You've taken in information and output different information.
- LFLex Fridman
Yeah.
- JSJeffrey Shainline
Hopefully reducing the total amount of information and extracting what's useful.
- LFLex Fridman
Yeah.
- JSJeffrey Shainline
Communication is then getting that information from the location in which it's stored, because information is physical, as Landauer de- emphasized. And so, it, it is more, in one place, and you need to get that information to another place so that something else can use it for whatever computation it's working on. Maybe it's part of the same network and you're all trying to solve the same problem, but neuron A over here just deduced something based on its inputs and it's now sending that information across the network to another location. So, that would be the active communication.
- LFLex Fridman
Can you linger on Landau on saying information is physical?
- JSJeffrey Shainline
Rolf Landauer, not to be confused with Lev Landau. Yeah, and he, he, eh, made huge contributions to our, our understanding of the reversibility of information and, and this concept that energy has to be dissipated in computing when the computation is irreversible. But if you can manage to make it reversible, then you, you don't need to expend energy. But if you, um, if you do expend energy to perform a computation, there's sort of a minimal amount that you have to do, and it's KT log 2.
- LFLex Fridman
And it's also how related to the second law of thermodynamics and that the universe is an information pros- and that we're living in a simulation.
- JSJeffrey Shainline
(laughs)
- LFLex Fridman
So, okay, sorry. Sorry for that tangent. So, informa- so that's the, the defining the, the, the distinction between computation and communication.
- JSJeffrey Shainline
Let me say one more thing-
- LFLex Fridman
Yes.
- JSJeffrey Shainline
... just to clarify. Communication ideally does not change the information. It moves it from one place to another-
- 46:36 – 57:19
Electrons for computation and light for communication
- LFLex Fridman
versus light distinction, and why are electrons, uh, good at computation and light good at communication?
- JSJeffrey Shainline
Yes. This is, um... There's a lot that goes into it, I guess, but just try to speak to the simplest part of it. Electrons interact strongly with one another. They're charged particles, so if I pile a bunch of 'em over here, they're feeling a, a certain amount of force and they wanna, they wanna move somewhere else, they're, they're strongly interactive. You can also get them to sit still. You can... An electron has a mass, so you can, you can cause it to be spatially localized. So, for computation, that's useful, because now I can make these little devices that put a bunch of electrons over here and then I change the, the state of a gate like I've been describing, put a different voltage o- on this gate, and now I move the electrons over here. Now they're sitting somewhere else. I have a physical mechanism with which I can represent information. It's spatially localized and I have knobs that I can adjust to change where those electrons are or what they're doing. Light, by contrast, photons of light, uh, which are the discreet packets of energy that were identified by Einstein, they do not interact with each other, um, especially at, at low light levels. If you're in a medium and you have a, a high, a bright high light level, you, you can get them to interact with each other through the interaction with that medium that they're in.
- LFLex Fridman
Mm-hmm.
- JSJeffrey Shainline
But that's, that's a little bit more exotic. And for the purposes of this conversation, we can assume that photons don't interact with each other. So, if you have a bunch of them all propagating in the same direction, they don't interfere with each other.
- LFLex Fridman
Mm-hmm.
- JSJeffrey Shainline
If I wanna send... If I, if I have a, a communication channel and I put one more photon on it, it doesn't screw up what those other ones, it doesn't change what those other ones were doing at all.
- LFLex Fridman
Mm-hmm.
- JSJeffrey Shainline
So, that's really useful for communication because that means you can sort of allow a lot of these photons to flow, um, with- without disruption of each other, and they can, they can branch really easily and things like that. But it's not good for computation, because it's very hard for this packet of light to change what this packet of light is doing.
- LFLex Fridman
Hm.
- JSJeffrey Shainline
They, they pass right through each other, so in computation, you want to change information, and if photons don't interact with each other, it's difficult to get them to change the information represented by the others.
- LFLex Fridman
So, that, that's the fundamental difference. Is, is there also something about the way they travel through different materials? I- i- or, or is that just, um, a particular engineering...
- JSJeffrey Shainline
No, it's not. That's deep physics, I think. So, this gets back to electrons interact with each other, and photons don't. So say, say I'm trying to get a, a packet of information from me to you, and we have a wire going between us. In order for me to send electrons across that wire, I first have to raise the voltage on my end of the wire, and that means putting a bunch of charges on it, and then that, that charge packet has to propagate along the wire. And it has to get all the way over to you. There's, that wire is gonna have something that's called capacitance, which basically tells you how much charge you need to put on the wire in order to, to raise the voltage on it. And the capacitance is gonna be proportional to the length of the wire. So the longer the, the length of the wire is, the more charge I have to put on it, and the energy required to charge up that line and move those electrons to you is also proportional to the capacitance and goes as the voltage squared. So-
- LFLex Fridman
Mm-hmm.
- JSJeffrey Shainline
... you get this huge penalty-
- LFLex Fridman
Yeah.
- JSJeffrey Shainline
... if you, if you wanna send electrons across a wire over appreciable distances.
- LFLex Fridman
So distance is, is an important thing here when you're doing communication.
- JSJeffrey Shainline
Distance is an important thing. So is the number of connections I'm trying to make.
- LFLex Fridman
Right.
- JSJeffrey Shainline
Me to you, okay, one, that's not so bad. If I want to now send it to 10,000 other friends, then I, then all of those wires are adding tons of extra capacitance. Now, not only does it take forever to put the charge on that wire and raise the voltage on all those lines, but it takes a ton of power. And the number 10,000 is not randomly chosen. That's roughly how many connections each neuron in your brain makes. So it... A neuron in your brain needs to send 10,000 messages every time it has something to say. You can't do that if you're trying to drive electrons from here to 10,000 different places. The brain does it in a slightly different way, which we can discuss.
- LFLex Fridman
How can light achieve the 10,000 connections, and why is it, um, why is it better in terms of, like, the energy use, uh, required to use light for the communication of the 10,000 connections?
- JSJeffrey Shainline
Right, right. So now, instead of trying to send electrons from me to you, I'm trying to send photons. So I can make what's called a wave guide, which is just, uh, a simple piece of, uh, material. It could be glass, like an optical fiber, or silicon, uh, on a, on a chip. And I just have to, I just have to inject photons into that wave guide. And independent of how long it is, independent of how many different connections I'm making, it doesn't change the, the voltage or anything like that that I have to raise up on the, on the wire. So if I have one more connection, if I add additional connections, I need to add more light to the wave guide because those photons need to split and go to different-
- LFLex Fridman
Mm-hmm.
- JSJeffrey Shainline
... paths. That makes sense. But I don't have a capacitive penalty, that sometimes these are called wiring parasitics. There are no parasitics associated with light in that same sense. So-
- LFLex Fridman
Well, just a... This might be a dumb question, but how do, how do I catch a photon on the other end? Uh, what's... Is it material? Is this the polymer stuff you were talking about for the, for a different application for, uh, photolithography? Like, how do you catch a photon?
- JSJeffrey Shainline
There's a lot of ways to catch a photon. It's not a dumb question. It's a, it's a deep and important question that basically defines a lot of the work that goes on in our group at NIST.
- LFLex Fridman
Mm-hmm.
- JSJeffrey Shainline
Uh, one of my group leaders, Sae Woo Nahm, has built his career around these superconducting single-photon detectors. So if you're going to try to sort of reach a lower limit and detect just one particle of light, superconductors come back into our conversation. And just picture a simple device where you have current flowing through a superconducting wire-
- LFLex Fridman
Mm-hmm.
- JSJeffrey Shainline
... and, um-
- 57:19 – 1:22:11
Neuromorphic computing
- LFLex Fridman
question, but the, the, le- let's zoom in at this very particular question of, uh, computation, uh, on a processor and communication between processors. So what does this system look like that you're envisioning? Uh, one of the places you're envisioning it is in the paper on opto-electronic intelligence. So what are we talking about? Are we talking about something that starts to look a lot like the human brain, or does it still look a lot like a computer? What are the size of this thing? Does it go inside a smartphone, or as you said, does it go inside something that's more like a house? Like, uh, wha- what should we be imagining? What are you thinking about when you're thinking about these fundamental, uh, systems?
- JSJeffrey Shainline
Let me introduce the word neuromorphic. There's this concept of neuromorphic computing, where what that broadly refers to is, um, computing based on the information processing principles of the brain.
- LFLex Fridman
Mm-hmm.
- JSJeffrey Shainline
And as digital computing seems to be pushing towards some fundamental performance limits, people are considering architectural advances, drawing inspiration from the brain, more distributed parallel network kind of architectures and stuff. And so there's this continuum of- of neuromorphic from things that are p- pretty similar to digital computers, but maybe there are more cores and the way they send messages is a little bit more like the way brain n- neurons send spikes. But for the most part, it's still digital electronics. And then, you know, you have some things in between where maybe you're- you're using transistors, but now you're starting to use them instead o- of in a digital way in an analog way. And so you're trying to get those circuits to behave more like neurons.
- LFLex Fridman
Mm-hmm.
- JSJeffrey Shainline
And then that's a little bit, uh, quite, quite a bit more o- on the neuromorphic side of things. You're trying to get your circuits, although they're still based on silicon, you're trying to get them to perform operations that are highly analogous to the operations in the brain. That's where a great deal of work is in neuromorphic computing. People like Iacomo Windiveri and Gerd Kahlenberg, S- Jennifer Hassler, countless others. It's a, it's a rich and exciting field, uh, going back to Carver Mead in the late 1980s. And then all the way on the other extreme of the continuum is where you say, "I'll give up anything related to transistors or semiconductors or anything like that. I'm not, I'm not starting with the assumption that I'm gonna use any kind of conventional computing hardware. And instead what I wanna do is try and understand what makes the brain powerful at the kind of information processing it does. And I wanna think from first principles about what hardware is best going to enable us to capture those information processing principles in an artificial system." And that's where I live. That's where, that's where I'm doing my exploration these days.
- LFLex Fridman
So, uh, what are the first principles of brain-like computation communication?
- JSJeffrey Shainline
Right. Yeah, this is, this is so important and I'm glad we booked 14 hours for this because, uh-
- LFLex Fridman
I only have 13. I'm sorry.
- JSJeffrey Shainline
(laughs)
- LFLex Fridman
(laughs)
- JSJeffrey Shainline
Okay, so th- the brain is notoriously complicated.
- LFLex Fridman
Yeah.
- JSJeffrey Shainline
And I think that's a- an important part of why it, why it can do what it does, but okay. Let me, let me try to break it down. Uh, starting with the devices, neurons, w- as I, as I said before, they're, they're sophisticated devices in and of themselves and synapses are too. They, they can, um, change their state based on the activity, so they, they adapt over time. That's crucial to the way the brain works. They don't just adapt on one timescale. They can adapt on m- myriad timescales from the, the spacing between pulses, the spacing between spikes that come from neurons all the way to the age of the organism. Um-Also relevant, perhaps, I think the most important thing that's guided my thinking is the, the network structure of the brain. So-
- LFLex Fridman
Which can also be adjusted-
- JSJeffrey Shainline
Yeah. Absolu-
- LFLex Fridman
... on different scales.
- JSJeffrey Shainline
Absolutely, yeah. So- so you're- you're making new cont- you're changing the strength of contacts, you're changing the- the spatial distribution of them. Although, spatial distribution doesn't change that much once you're a- a mature organism, but that network structure is- is really crucial. So, let me dwell on that for a second. Um, you can't talk about the brain without emphasizing that most of the neurons in the- the neocortex or prefrontal cortex, the part of the brain that we think is most responsible for high level reasoning and things like that, those neurons make thousands of connections. So, you have this network that is highly interconnected, and m- I- I think it's safe to say that one of the primary reasons that they make so many different connections is that allows information to be communicated very rapidly from any spot in the network to any other spot in the network. So, that's a, that's a sort of spatial aspect of it. You can quantify this, uh, in terms of concepts that are related to fractals and scale and variance, which I think is a- is a very beautiful concept. So, what I mean by that is, kind of no matter what spatial scale you're looking at in the brain, within certain bounds, you see the same general statistical pattern. So if- if I draw a box around some region of my cortex, most of the connections that those neurons within that box make are gonna be within the box to each other in their local neighborhood, and that's sort of called clustering, loosely speaking. But a non-negligible fraction is gonna go outside of that box, and then if I draw a bigger box, the pattern is gonna be exactly the same. So, you have this scale and variance, and you also have a- a non-vanishing probability of a neuron making con- connection very far away. So, suppose you- you wanna plot the probability of a neuron making a- a connection as a function of distance. If that were an exponential function, it would go E to the minus radius over some characteristic radius, and it would- it would drop off. Up to some certain radius, the probability would be reasonable, clo- close to one, and then a- a beyond that characteristic length, R- R0, it would- it would drop off sharply. And so that would mean that the neurons in your brain are- are really localized-
- LFLex Fridman
Mm-hmm.
- JSJeffrey Shainline
... and that's not what we observe. In- instead what you see is that the probability of making a longer distance connection, it does drop off, but it drops off as a power law. So the probability-
- LFLex Fridman
Mm-hmm.
- JSJeffrey Shainline
... that you're gonna have a connection at some radius R goes as R to the minus sum power, and that's more, that's what we see with- with forces in nature, like the-
- LFLex Fridman
Mm-hmm.
- JSJeffrey Shainline
... electromagnetic force between two particles, or gravity, goes as one over the radius squared. So-
- LFLex Fridman
You can see this in fractals. I love that there's a- a, like a fractal dynamics of the brain, that if you zoom out, you draw the box and you increase that box by certain step sizes, you're gonna see the same statistics.
- JSJeffrey Shainline
I think that's probably v- very important to the way the brain processes information. It's not just in the spatial domain. It's also in the temporal domain.
- LFLex Fridman
Temporal.
- JSJeffrey Shainline
And what I mean by that is-
- LFLex Fridman
Pew!
- JSJeffrey Shainline
(laughs)
- 1:22:11 – 1:25:28
What is NIST?
- JSJeffrey Shainline
- LFLex Fridman
Can I, can I quickly ask, what is NIST and where is this amazing group of people located?
- JSJeffrey Shainline
NIST is the National Institute of Standards and Technology. They, uh... Th- the larger facility is out in Gaithersburg, Maryland. Our team is located in Boulder, Colorado. Um, we, NIST, is a, is a, a federal agency under the Department of Commerce. We do a lot with... By "we," I mean other people at NIST, but they'll do a lot with standards, you know, um, making sure that we understand the system of units, international system of units, uh, precision measurements. There's a lot going on in, uh, electrical engineering, material science.
- LFLex Fridman
And it's historic. I mea- I mean, it's like... It's one of those. It's like MIT or something like that. It has a reputation over many decades of just being this really, um, uh, a place where there's a lot of brilliant people have done a lot of amazing things. But in terms of the people in your team, in this team of people involved in the concept we're talking about now, I'm just curious, what kind of disciplines are we talking about? What... Is it-
- JSJeffrey Shainline
Mostly physicists and electrical engineers. Some material scientists, um, but I would say, yeah, I think physicists and electrical engineers. My background is in photonics, the use of light for technology, so coming from there, I, I tend to have found colleagues that are more from that background. Although, uh, Adam McKone, more of a superconducting electronics background. We need a diversity of folks. This project is s- sort of cross-discipline. I would love to be working more with neuroscientists and things, um, but w- we haven't, we haven't reached that scale yet, but yeah.
- LFLex Fridman
But you- you're, you're focused on the hardware side, which requires all, uh, all the disciplines that you mentioned?
- JSJeffrey Shainline
Yes.
- LFLex Fridman
And then, of course, you know, science may be a source of inspiration-
- JSJeffrey Shainline
Yes.
- LFLex Fridman
... for some of the, the, the, the long-term vision.
- JSJeffrey Shainline
I would actually call it more than inspiration. I would call it sort of, um, a roadmap, you know? I, we're, we're not trying to, to build exactly the brain, but I don't think it's enough to just say, "Oh, neurons kinda work like that. Let's kinda do that thing."
- LFLex Fridman
Right.
- JSJeffrey Shainline
I- I mean, we're, we're, we're very much following the concepts that the cognitive scientists have laid out for us, which I believe is a, is a really robust roadmap. I mean, just o- on a little bit of a tangent, it's often stated that we just don't understand the brain, and so it's really hard to replicate it because we j- we just don't know what's going on there.
- LFLex Fridman
Yeah.
- JSJeffrey Shainline
And I... Maybe five or seven years ago, I would, I would have said that, but as I got more interested in the subject, I ha- read more of the neuroscience literature, and I was just taken by the exact opposite sense. I can't believe how much they know about this. I can't believe how mathematically rigorous and, um, sort of theoretically complete a lot of the concepts are. That's not to say we understand consciousness or we understand the self or anything like that. But why is the brain... What is the brain doing and why is it doing those things? We have a... Neuroscientists have a lot of answers to those questions, so there's a lot... If you're a hardware designer that just wants to get going, whoa, it's pretty clear which direction to go in, I, I think.
Episode duration: 2:56:42
Install uListen for AI-powered chat & search across the full episode — Get Full Transcript
Transcript of episode EwueqdgIvq4
Get more out of YouTube videos.
High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.
Add to Chrome