Skip to content
The Twenty Minute VCThe Twenty Minute VC

Arvind Narayanan: AI Scaling Myths, The Core Bottlenecks in AI Today & The Future of Models | E1195

Arvind Narayanan is a professor of Computer Science at Princeton and the director of the Center for Information Technology Policy. He is a co-author of the book AI Snake Oil and a big proponent of the AI scaling myths around the importance of just adding more compute. He is also the lead author of a textbook on the computer science of cryptocurrencies which has been used in over 150 courses around the world, and an accompanying Coursera course that has had over 700,000 learners. ----------------------------------------------- Timestamps: (00:00) Intro (01:18) AI Hype vs. Bitcoin Hype: Similarities & Differences (03:49) The Misalignment Between Compute & Performance (08:10) Synthetic Data (09:30) Creating Effective Agents Despite Incomplete Data (12:00) Why Is the AI Industry Shifting Toward Smaller Models (16:31) The Growing Gap Between AI Models & Compute Capabilities (19:44) Predictions on the Timeline for AGI (27:00) Policy Proposals for U.S. and European AI Regulation (29:29) AI & Deepfakes: The Risk of Discrediting Real News (35:59) Revolutionising Healthcare with AI in Your Pocket (40:29) Is AI Job Replacement Fear Overhyped or Real? (41:46) AI's Potential as a Weapon (46:19) Quick-Fire Round ----------------------------------------------- In Today’s Episode with Arvind Narayanan We Discuss: 1. Compute, Data, Algorithms: What is the Bottleneck: Why does Arvind disagree with the commonly held notion that more compute will result in an equal and continuous level of model performance improvement? Will we continue to see players move into the compute layer in the need to internalise the margin? What does that mean for Nvidia? Why does Arvind not believe that data is the bottleneck? How does Arvind analyse the future of synthetic data? Where is it useful? Where is it not? 2. The Future of Models: Does Arvind agree that this is the fastest commoditization of a technology he has seen? How does Arvind analyse the future of the model landscape? Will we see a world of few very large models or a world of many unbundled and verticalised models? Where does Arvind believe the most value will accrue in the model layer? Is it possible for smaller companies or university research institutions to even play in the model space given the intense cash needed to fund model development? 3. Education, Healthcare and Misinformation: When AI Goes Wrong: What are the single biggest dangers that AI poses to society today? To what extent does Arvind believe misinformation through generative AI is going to be a massive problem in democracies and misinformation? How does Arvind analyse AI impacting the future of education? What does he believe everyone gets wrong about AI and education? Does Arvind agree that AI will be able to put a doctor in everyone’s pocket? Where does he believe this theory is weak and falls down? ----------------------------------------------- Subscribe on Spotify: https://open.spotify.com/show/3j2KMcZTtgTNBKwtZBMHvl?si=85bc9196860e4466 Subscribe on Apple Podcasts: https://podcasts.apple.com/us/podcast/the-twenty-minute-vc-20vc-venture-capital-startup/id958230465 Follow Harry Stebbings on Twitter: https://twitter.com/HarryStebbings Follow Arvind Narayanan on Twitter: https://twitter.com/random_walker Follow 20VC on Instagram: https://www.instagram.com/20vchq Follow 20VC on TikTok: https://www.tiktok.com/@20vc_tok Visit our Website: https://www.20vc.com Subscribe to our Newsletter: https://www.thetwentyminutevc.com/contact ----------------------------------------------- #20vc #harrystebbings #arvindnarayanan #princetonuniversity #ai #venturecapital #samaltman #alexwang #openai #computerscience #technology

Arvind NarayananguestHarry Stebbingshost
Aug 28, 202450mWatch on YouTube ↗

EVERY SPOKEN WORD

  1. 0:001:18

    Intro

    1. AN

      We're not gonna have too many more cycles, possibly zero more cycles of a model that's almost an order of magnitude bigger in terms of the number of parameters than what came before, and thereby more powerful. And I think a reason for that is data becoming a bottleneck. These models are already trained on essentially all of the data that companies can get their hands on. While data is becoming a bottleneck, I think more compute still helps, but maybe not as much as it used to.

    2. HS

      Ready to go? (instrumental music plays) Arvind, I am so excited for this, dude. I was telling you just now, I am one of your biggest fans on the Substack newsletter. I can't wait for the book. So thank you so much for joining me today.

    3. AN

      Thank you. I really appreciate that and super excited for this conversation.

    4. HS

      Now, I wanna get pretty much straight into it, but for those that don't read the Substack, which they should do, can you just provide a 60-second intro, some context on why you're so well-versed to speak on the topics that we are today?

    5. AN

      Uh, so I'm a professor of computer science, and I would say I do kind of three things. Uh, one is technical AI research, and another is understanding the societal effects of AI, and the third is advising policymakers.

  2. 1:183:49

    AI Hype vs. Bitcoin Hype: Similarities & Differences

    1. AN

    2. HS

      I'd just love to start, before we dive in deep on infrastructure, how does the AI hype today compare to Bitcoin hype? How is it the same, and how is it different?

    3. AN

      So, I've spent years of my time on this. I really believed that decentralization, uh, could have tremendous societal impacts. And that was the angle that really mattered to me, right? How is this going to make society better? It was not the money angle. But by around 2018, I had started to get really disillusioned, and that was because of, uh, a couple of main things. One is, uh, in a lot of cases where I thought cr- where I had thought crypto or blockchain was going to be the solution, I realized that that was not the case. So for instance, uh, you know, while there is potential for crypto to help the world's unbanked, uh, the tech is not the real bottleneck there. And the other part of it was just the philosophical aspects of this community. Uh, you know, I- I do believe that many of our institutions are in need of reform or maybe decentralization, whatever it is, and that includes academia, by the way. So many reforms so badly needed. And in an ideal world, we would have this, you know, hard but important conversation about, how do you fix our institutions? But instead, these students have been sold on blockchain, and they wanna replace these institutions with a script. And, uh, that just didn't seem like the right approach to me. So both from a technical perspective and from a philosophical perspective, I really soured on it. While there are harms, uh, around AI, I think it has been a net positive for society. I can't say the same thing about Bitcoin.

    4. HS

      Are we in an AI hype cycle right now?

    5. AN

      (laughs) I think that's possible. I think there is a lot of talk about whether, uh, AI is a bubble right now. And I, you know, I won't make predictions about what's gonna happen, but I do think that AI companies, generative AI companies specifically, made some serious mistakes in the last year or two about how they went about things.

    6. HS

      What mistakes did they make, Arvind?

    7. AN

      When ChatGPT was released, uh, people found, you know, a thousand new applications for it, right, that OpenAI might not have anticipated. And that was great, but I think developers, AI developers took the wrong lesson from this. They thought that AI is so powerful and so special that you can just put these models out there and people will figure out what to do with them. They didn't think about actually building products, uh, you know, making things that people want, finding product-market fit, and all of those things that are so basic in tech, but somehow, AI companies deluded themselves into thinking that the normal rules

  3. 3:498:10

    The Misalignment Between Compute & Performance

    1. AN

      don't apply here.

    2. HS

      I do wanna ask, uh, we can start with kind of the hardest question of all, but it's the most important. And you've written about this, and I loved your piece. The kind of core question that everyone's asking right now is, does more compute equal an increased level of performance, or have we reached a point where it is misaligned and more compute will not create that significant spike in performance? Kevin Scott at Microsoft says, "Absolutely. We have a lot more room to run." Why are you skeptical, and have we gotten to a stage of diminishing returns on compute?

    3. AN

      So if we look at what's happened historically, the way in which, uh, compute has improved model performance is with companies building bigger models, right? So I- I- in my view, at least, the biggest thing that changed between GPT-3.5 and GPT-4 was the size of the model. And it was, you know, also trained with, uh, more data presumably, although they haven't made the details of that public, and more compute and so forth. So I think that's running out. I think we're not going to be... we're not gonna have too many more cycles, possibly zero more cycles, of a model that's, you know, um, uh, almost an order of magnitude bigger in terms of the number of parameters, uh, than what came before, and thereby more powerful. And I think a reason for that is data becoming a bottleneck. These models are already trained on essentially all of the data that companies can get their hands on. While data is becoming a bottleneck, I think more compute still helps, but maybe not as much as it used to. And the reason for that is that, uh, perhaps ironically, more compute allows one to build smaller models with the same capability level, and that's actually the trend we've been seeing over the last year or so, as you know. You know, the models today have gotten somewhat smaller and cheaper than when GPT-4 initially came out but with the same capability level. So I think that's probably gonna continue. Are we going to see a GPT-5 that's as big a leap over GPT-4 as GPT-4 was over GPT-3? I'm frankly skeptical.

    4. HS

      Can we just take them one by one there? There was a lot of great things that I just wanna unpack.

    5. AN

      (laughs)

    6. HS

      You said there about kind of, you know, potentially the shortage of data being the bottleneck to performance. A lot of people say, "Well, there's a lot of data that we haven't mined yet," which...And the obvious example that many have suggested is kind of YouTube, which has obviously, I think, 150 billion hours of video. Um, and then secondarily to that, synthetic data, the creation of artificial data that hasn't, that isn't exis- in existence yet. To what extent are those effective pushbacks?

    7. AN

      So there aren't, uh, there are a lot of sources that haven't been mined yet, but when we start to look at the volume of that data, how many tokens is that, I think, uh, the picture is a little bit different. 150 billion hours of video sounds, you know, really impressive, uh, but when you put that video through a speech recognizer and actually extract the text tokens out of it and de-duplicate it and so forth, it's actually not that much. It's an order of magnitude smaller than, uh, what some of the largest models today have already been trained with. Now, training on video itself, instead of text extracted from the video, uh, I think that could be, uh, that could lead to some new capabilities, but not in the same fundamental way that we've had before where you have the emergence of new capabilities, uh, right? Models being able to do things, uh, that, uh, people just weren't anticipating. So, like, the kind of shock that the AI community had when I think, back in the day I think it was GPT-2, was trained primarily on English text and they had actually tried to filter out text in other languages to keep it clean, but a tiny amount of text from other languages had gotten into it, and it turned out that that was enough for the model, uh, to pick up a reasonable level of competence for conversing in various other languages. So these are the kinds of emergent capabilities that really spooked people, uh, that has led to both a lot of hype and a lot of fears about what bigger and bigger models are going to be able to do. But I think that has pretty much run out because we're, you know, we're training on capabilities that humans have expressed, like translating between languages, and have already put out there in the form of text. Uh, so if you make the data set, uh, you know, a little bit more diverse with YouTube video, I don't think that's fundamentally gonna change. Multimodal capabilities, yes, there's a lot of room there, but new emergent text capabilities, I'm not sure.

  4. 8:109:30

    Synthetic Data

    1. AN

    2. HS

      What about synthetic data? What about the creation of new data that doesn't exist yet?

    3. AN

      So, uh, there's two ways to look at this, right? So one is the way in which synthetic data is being used today, which is not to increase the volume of training data, but it's actually to overcome limitations in the quality of the training data that we do have. Uh, so for instance, if in a particular language there is too little data, you can try to augment that or you can try to, um, uh, let's say, uh, y- have a model, uh, you know, solve a bunch of mathematical equations, throw that into the training data, uh, and so for the next training run, that's going to be part of the pre-training and so the model will get better at doing that. Uh, and the other way to look at synthetic data is, okay, you take one trillion tokens, you train a model on it, and then you output 10 trillion tokens so you get to the next bigger model, all right, and then you use that to output 100 trillion tokens. Uh, you know, I'll, uh, I'll bet that's just not going to happen. That's just a snake eating its own tail. And what we've learned in the last two years is that the quality of data matters a lot more than the quantity of data. So if you're using synthetic data, uh, to, uh, try to augment the, the quantity, I think it's just coming at the expense of quality. You're, you're not learning new things from the data. You're only learning things that are already there.

  5. 9:3012:00

    Creating Effective Agents Despite Incomplete Data

    1. AN

    2. HS

      One thing that, o- while we're on kind of utility value of data, when we look at effectiveness of agents, you know, I've had Alex Wang, uh, you know, Scale.AI on the show, and he said the hardest thing about, like, building effective agents is most of the work that one does in an organization you don't actually codify down in data. You know, like, you remember when you were at school and it says, like, "Show your thinking," or, "Show your work"? You don't do that in an organization. You draw on the whiteboard, you map it out, and then you put down what you think in the document. The whiteboard is often not correlated in a data source. To what extent do we have the data of showing your work for models, agents to actually do in a modern enterprise?

    3. AN

      Yeah, I think that's really spot-on. I think, uh, one way in which people's intuitions have been, uh, kind of misguided by the rapid improvements in LLMs is that all of this has been, you know, in the paradigm of learning from data on the web that's already there, and once that runs out, you have to switch to new kinds of learning. Uh, it's the analog of, uh, you know, riding a bike that's just kind of tacit knowledge. It's not something, uh, that's been written down. So a lot of what happens in organizations is the cognitive equivalent of, uh, I think what happens in the physical skill of riding a bike. And I think for models to learn, uh, a lot of these diverse kinds of tasks that they're not gonna pick up from the web, you have to have the cycle of actually using the AI system in your organization and for it to learn from that back-and-forth experience instead of just passively ingesting.

    4. HS

      To what extent do you think enterprises today are willing to let passive AI products into their enterprises to observe, to learn, to test, and is there really that willingness, do you think?

    5. AN

      I think it's, it's, it's gotta be more than passive observation. It's got, you have to, uh, actually deploy AI for, uh, uh, to be able to, uh, get to certain types of learning. Uh, and I think that's gonna be very slow, and I think the, uh, a good analogy is self-driving cars, of which we had prototypes, you know, two or three decades ago, but for, for these things to actually be deployed, you have to roll it out on slightly larger and larger scales while you collect data, while you make sure you get to the next nine of reliability, you know, four nines of reliability to five nines of reliability. So it's that very slow rollout process. It's a very slow feedback loop, and I think that's gonna happen with a lot of AI deployment in organizations

  6. 12:0016:31

    Why Is the AI Industry Shifting Toward Smaller Models

    1. AN

      as well.

    2. HS

      You said about smaller models. Help me just understand again. I'm sorry. I, I, uh, the show's very successful, Arvin, because I think I ask the questions that everyone asks but they're too afraid to actually admit they don't know the answers to. Um-Why are we seeing this trend towards smaller models, and why do we think that is the most likely outcome in the model landscape to have a world of many smaller models?

    3. AN

      My view is that, uh, in a lot of cases, the adoption of these models is not bottlenecked by capability. If these models were actually deployed today, uh, to do all the tasks that they are capable of, it would truly be a striking economic transformation. The bottlenecks are things other than capability. And one of the big ones is cost. And cost, of course, is roughly proportional to the size of the model, and that's putting a lot of downward pressure on model size. Once you get a model small enough that you can run it on-device (clears throat) , excuse me, that of course opens up a lot of new possibilities, uh, both in terms of, uh, privacy. You know, people are much more comfortable with on-device models, especially if it's something that's going to be listening to their phone conversations or looking at their desktop screenshots, which are exactly the kinds of AI assistants that companies are, uh, building and pushing. Uh, and, uh, just, you know, from the perspective of cost, you don't have to dedicate servers to run that model. So, I think those are a lot of the reasons why, uh, companies are furiously working on making models smaller without a big hit in capability.

    4. HS

      Will Moore's Law not mean cost goes down dramatically in actually a relatively short three-to-five-year period?

    5. AN

      You're right. Cost is going down dramatically. I think, in certain applications, uh, cost is going to become much less of a barrier, but not across the board.

    6. HS

      Where does it become a barrier, and where does it not?

    7. AN

      There's this interesting concept called Jevons paradox, and this was first in the context of, uh, uh, coal in England in the 18th century. I think when coal mining got cheaper, there was more demand for coal, and so the amounts invested into coal mining actually increased. And I predict that we're gonna see the same thing with models. When models, uh, get cheaper, they're put into a lot more things, and so, uh, the total amount that companies are spending on inference is actually going to increase. On a task like a- uh, uh, in an application like a chatbot, let's say, you know, it's text in, text out, no big deal. I think costs are going to come down. Even if someone is chatting with, uh, a chatbot all day, it's probably not going to get t- too expensive. On the other hand, if you want to scan all of someone's emails, for instance, right? If a model gets cheaper, you know, you're just gonna have it running always on in the background, and then from emails, you're gonna get to all their documents, right? And some of those attachments might be many megabytes long. Uh, and so there, even with Moore's Law, I think cost is going to be significant in the medium term. And then, you get to applications like writing code, where what we're seeing is that it's actually very beneficial to let the model do the same task tens of times, thousands of times, sometimes literally millions of times, and pick the best answer. So, in those cases, it doesn't matter how much cost goes down. You're gonna t- just proportionally increase the number of retries so that you can get a better quality of output.

    8. HS

      So, we have smaller models, and, but they're effective, Asby said that, because of cost, and they're popular because of cost. What does that do to the requirements in terms of compute?

    9. AN

      So, there is training compute, which is when the developer is building the model, and then there is inference compute, when the model is being deployed and the user is using it to do something. And it might seem like, uh, really the, uh, training cost is the one we should worry about since, you know, it's trained on all of the text on the internet or whatever. But it turns out that over the lifetime of a model when you have billions of people using it, the inference cost actually adds up. And for many of the popular models, that's the cost that dominates. Let's talk about, uh, each of those two costs. Uh, with respect to training cost, if you want to build a smaller model at the same level of capability or without compromising capability too much, you have to actually train it for longer. So, that increases training costs. But that's maybe okay because you have a smaller model. You can push it to the consumer device, or, you know, your- e- e- even if it's running on the cloud, your server costs are- are lower, so your training cost increases, your inference cost decreases. But because it's the inference cost that dominates, the total cost is probably gonna come down.

    10. HS

      So, total cost comes down.

    11. AN

      If you have the same workload and you have a smaller model doing it, then the total cost is gonna come down.

  7. 16:3119:44

    The Growing Gap Between AI Models & Compute Capabilities

    1. AN

    2. HS

      When we think about the alignment in compute and models, uh, y- we had David Kahn from Sequoia on the show, and he said that you would never train a frontier model on the te- same data center twice, meaning that essentially there is now a misalignment in the development speed of models, and they- that is much faster than the development speed of new hardware and compute. How do you think about that? We s- we are releasing new models so fast that computers-

    3. AN

      Yeah.

    4. HS

      ... unable to keep up with them, and as a result, you won't want to train your new model on old H100 hardware that is 18 months old. You need continuously the newest hardware for every single new frontier model.

    5. AN

      I think we are still in a period where, uh, you know, these models have not yet quite become commoditized. There's, uh, there's obviously a lot of progress, and there's a lot of demand on hardware as well. Hardware cycles are also improving rapidly. But, you know, there is the saying that every exponential is a sigmoid in disguise. So, a sigmoid curve is one that looks like an exponential at the beginning, so imagine the S letter shape, uh, but then after a while, it has to taper off like every exponential has to taper off. So, I think that's gonna happen both with models as well as with these hardware cycles. You know, I can't predict how long that's gonna take, but we are, I think, gonna get to a world where models do get commoditized.

    6. HS

      Speaking of that commoditization, th- the thing that I'm interested by there is kind of the benchmarking or the determination that they are suddenly commoditized or kind of equal performance. You said before LLM evaluation is a minefield. Help me understand. Why is LLM evaluation a minefield?

    7. AN

      A big part of it is this issue of vibes, right? So, you evaluate LMs on these benchmarks, but then, uh, you know, it- it seems to perform really well on the benchmarks, but then the vibes are off. In other words, you start using it, and somehow it doesn't feel adequate. It makes a lot of mistakes in ways that are not captured in the benchmark.And the reason for that is simply that when there is so much pressure to do well on these benchmarks, developers are intentionally or unintentionally optimizing these models in ways that look good on the benchmarks but, uh, don't look good in real world evaluation. So w- when GPT-4 came out and OpenAI claimed that it passed the bar exam and the medical licensing exam, uh, people were very excited/slash, uh, scared about what this means for doctors and lawyers, and the answer turned out to be approximately nothing, right? Because it's not like a, a, a lawyer's job is to answer bar exam questions all day. Uh, these benchmarks that models are being tested on don't really capture what we would use them for in the real world. So that's one reason why LLM evaluation is a minefield, and there's also just a, a very, uh, simple factor of contamination. Maybe the model has already trained on the answers that it's being evaluated on in the benchmark and so if you ask it new questions, it's gonna struggle, uh, and there are various other pitfalls. So I think, um, you know, we shouldn't put too much stock into benchmarks. We should look at people who are actually trying to use these in professional contexts, whether it's lawyers or, you, you know, really anybody else, and we should go based on their experience of using these AI assistants.

  8. 19:4427:00

    Predictions on the Timeline for AGI

    1. AN

    2. HS

      We, we mentioned that, you know, some of the early use cases in terms of passing the bar and, um, some real kind of wild applications in terms of how models are applied. W- I do just want us to kind of move a layer deeper to the companies building the products and the leaders leading those companies. You've got Zuck and Demis who are saying that AGI is further out than we think, and then you have Sam Altman and you have, uh, Dario and Elon in some cases saying it's sooner than we think. What are your reflections and analysis on company leader predictions on AI... on AGI?

    3. AN

      So let's, uh, talk for a second about what AGI is. Different people mean different things by it and so often talk past each other. Um, the definition that we consider most relevant is AI that is capable of automating most economically valuable tasks. So it's a very pragmatic definition. It doesn't care about, you know, whether it's conscious, does it think like a person. Those questions are, uh, frankly not that interesting to us but also harder to predict or reason about. Uh, and, uh, so by this definition, you know, of automating most economically valuable tasks, if we did have AGI, that would truly be, uh, a profound thing in our society. Okay, so now for the CEO predictions. I think one thing that's helpful to keep in mind is that there have been these predictions of imminent AGI since the earliest days of AI, for more than a half century. Alan Turing, when the first, uh, computers were built or about to be built, people thought, "You know, the two main things we need for AI are hardware and software. We've done the hard part, the hardware. Now th- there's just one thing left, the easy part, the software." Uh, (laughs) but of course now we know how hard that is. So I think historically what we've seen is, it's kind of like climbing a mountain. Wherever you are, it looks like there's just kind of one step to go but when you climb up a little bit further, the complexity reveals itself and so we've seen that over and over and over again. Now it's like, "Oh, you know, we just need to make these bigger and bigger models," so you have some silly projections based on that. But soon the limitations of that b- start becoming apparent and now the next layer of complexity re- reveals itself. So that's my view. I, I, I wouldn't put too much stock into these overconfident predictions from CEOs.

    4. HS

      Is it possible to have a dual strategy of chasing AGI and superintelligence, as OpenAI very clearly are, and creating valuable models... oh, sorry, valuable products at the same time that can be used in everyday use? Or is that balance actually mutually exclusive?

    5. AN

      Certainly think the balance is possible. Uh, to some extent, every big company does this. A lot of, uh, their research, uh, that happens in the research labs.

    6. HS

      If you think about your priority, your priority at OpenAI is, say, achieving superintelligence and AGI. Their best researchers, their best developers, the core of their budgets will go to that. When you have dual priorities, one takes the priority, and so there is that conflict.

    7. AN

      That's fair, and I think, you know, i- it would take discipline from management, uh, to be able to pull it off in a way that one part of the company doesn't distract another too much. And we've seen this happen with OpenAI, which is the folks focused on superintelligence didn't feel very welcome at the company and there has been an exodus of very prominent people and Anthropic has picked up a lot of them. So it seems like we're seeing a split emerging where OpenAI is more focused on products and Anthropic is more focused on, uh, superintelligence. Uh, but while I can see the practical reasons why that is happening, I don't think it's impossible, uh, to have disciplined management that, uh, focuses on both objectives.

    8. HS

      What did you mean when you said to me that AI companies should pivot from creating gods to building products?

    9. AN

      In the past, uh, you know, they didn't have this balance. They, um, were so enamored by this prospect of, uh, creating AGI that they didn't think there was a need to build products at all. And, you know, the craziest example for me is when OpenAI put out ChatGPT, there was no mobile app for six months, uh, and the Android app took even longer than that. And there was this... you know, there was this assumption that ChatGPT was just going to be this kind of, uh, uh, really demo to show off the capabilities of the models. OpenAI was, you know, in the business of building these models and, uh, third party developers would take the API and put it into products, but really, AGI was coming so quickly that, that, you know, even the notion of productization seemed obsolete. This was, you know... I, I, I'm not trying to put words in anyone's mouth, but this was kind of a coherent but, in my view, incorrect philosophy that I think a lot of AI developers had. Um-... and I think that has, uh, uh, changed quite a bit now, and I think that's a good thing. Uh, so if they had to pick one, I think they should pick building products. But it certainly doesn't make sense for a company to be just an AGI company and not try to build products, not try to build something that people want, and just assuming that, uh, AI is gonna be so general that it's just gonna, uh, you know, do everything that people want and, and, and that the company doesn't actually need to make products.

    10. HS

      Do you think it's even possible for companies to compete in any level of AGI pursuit? When you look at the players and the cash that they're willing to spend, you know, Zuck has committed $50 billion over the next three years. When you look at how much OpenAI has raised over the last three years and they carry on that run rate, it's something crazy, like that'd still be $38 billion short of a Zuck spend over a three-year period. Can you create AGI-like products or godlike products unless you are Google, Amazon, Apple or Facebook?

    11. AN

      You know, we've been in this kind of historically, um, interesting period where a lot of progress has come from building bigger and bigger models. That need not continue in the future. It might, or what might happen is that the models themselves get commoditized, and a lot of the interesting development happens in a layer above the models. We're starting to see a lot of that happen now with AI agents. And if that's the case, great ideas could come from anywhere, right? It could come from a two-person startup. It could come from an academic lab. Uh, and my hope is that we will transition to that kind of, uh, mode of progress in AI development, uh, relatively soon.

    12. HS

      With the commoditization of those models and the appreciation that value can be built on top of them, does that not go back to what I said, though, which is really there is three to four core models which are financed by cash cow cloud businesses, f- f- you know, the obvious. There's Amazon and there's Google, and then for Facebook, there's obviously Instagram and News Feed, and there are three large model providers which sit as the foundational model there, and then every bit of values built on top of them.

    13. AN

      Uh, that might happen. I think that's a, a very serious possibility, and I think, uh, this is actually one area where regulators should be paying attention. You know, what does this mean for market concentration, antitrust, and so forth? And I've been gratified that, uh, these are topics that, um, uh, at least in my experience, US regulators are considering, and I believe in the UK, the CMA, the Competition and Markets Authority, as well, and certainly in the EU. Uh, so yeah, in many jurisdictions, now that I think about it, this is something that regulators have

  9. 27:0029:29

    Policy Proposals for U.S. and European AI Regulation

    1. AN

      been worried about.

    2. HS

      If you were suggesting, as, uh, you said at the beginning about kind of your work on policy, you have US regulators and European regulators. What would you put forward as the most proactive and effective policy for US and European regulation around AI and models?

    3. AN

      So, in a sense, AI regulation is a misnomer. Let me give you an example from just this morning. The FTC, uh, has been worried about, uh... The Federal Trade Commission in the US, uh, uh, you know, which is, um, an antitrust and consumer protection authority, has been worried about, uh, people writing fake reviews for their products, and this has, of course, been a problem for many years. It's become a lot easier to do that with AI. So now someone who thinks about this in terms of AI regulation might say, "Oh, you know, regulators have to ensure that AI companies don't allow their products to be used for generating fake reviews," and I think this is a losing proposition. Like how would an AI model know whether something is a fake review or a real, real review, right? It just depends on who's, uh, writing the review. But instead, you know, that's not the approach that the FTC took. They recognized correctly that it's a problem whether AI is generating the fake review or people are, so what they actually banned is fake reviews, right? And so what is often thought of as AI regulation is better understood as regulating certain harmful activities whether or not AI is used as a tool for doing those harmful activities. So I think, you know, 80% of what gets called AI regulation is better seen this way.

    4. HS

      When I had Ethan Mollick on from Wharton, he was like, "You know, the best thing to do actually is like a allow and watch a policy." He had, he had a much more academic approach to it in terms of naming than that with, you know, wonderful principal from some, you know, uh, ancient learning professor. Um, but he said essentially we should let everything flourish and then regulate from there rather than proactively regulate ahead of time not knowing outcomes. Does that ring true to you?

    5. AN

      I broadly agree with that. I will add a couple of, uh, uh, additions to that. One is there are many kinds of harms which we already know about and are quite serious. So the use of, uh, AI to make, uh, non-consensual deepfakes, for instance, deepfake nudes, and this has affected, you know, thousands, perhaps hundreds of thousands of people, primarily women, around the world, and, uh, governments, uh, are taking action now finally, so

  10. 29:2935:59

    AI & Deepfakes: The Risk of Discrediting Real News

    1. AN

      that's a good thing.

    2. HS

      Just on the verification side there and kind of what you said about the deepfakes, I think it was Sayash said on Twitter recently with a, a great highlighting that the biggest danger of AI to him was actually not that we would believe fake news. It was that we would start to distrust and not believe real news.

    3. AN

      I agree. So we call this the liar's dividend. Uh, so people have been worried, for instance, about bots creating misinformation with AI and, uh, influence in elections and that sort of thing. Uh, we're very, very skeptical that that's going to be a real danger.

    4. HS

      How are you not skeptical that that's a real danger? We're a media company. We have amazing media people who use AI every day. We could create some terrible things with AI today that people would believe.

    5. AN

      But you could have created those things without AI. I don't think this is-

    6. HS

      No, we could-

    7. AN

      ... an AI problem.

    8. HS

      We could, we could not have created Trump fakes with his voice declaring war on China. I could do a fake show with Trump today and release it and pretend that it's real and have him declare war on China.

    9. AN

      Yeah, I think that's fair, but I think the reason that might, uh, fool a lot of people, uh, is because it came from a legitimate media company. So, I think the ability to do this, you know, emphasizes some of the things that have always been important but have now become more important, like source credibility.

    10. HS

      Do we not see then in that world that actually a lot more value accrues to significant mainstream media outlets who are verified and have brand validity already? Because people know the difference.

    11. AN

      That is... Yeah. That's right, yeah.

    12. HS

      If it's from York Times, okay.

    13. AN

      That's right. So that, that actually is our prediction. People, we predict, are going to be forced to rely much more on getting their news from trusted sources.

    14. HS

      Does that worry you?

    15. AN

      (sighs) (laughs)

    16. HS

      Like, I, I, I, I understand, but sadly I don't think people are always as smart as we give them credit for. And when you look at the spread of misinformation, and when you look at the willingness to accept misinformation from large swathes of the population, a tweet with an AI-generated picture with whatever it could be in there can create such societal damage. Uh, uh, th- this is really worrying.

    17. AN

      Uh, so misinformation is a problem. I- in a way, I think misinformation is more of, uh, uh, a symptom than a cause. Uh, I think, you know, misinformation slots into, uh, and, and affirms people's existing beliefs as opposed to, uh, changing their beliefs. And I think the impact on AI here, again, has been tremendously exaggerated. Uh, sure, you know, you can create, uh, a Trump deepfake like you were talking about. But when you look at the misinformation that's actually out there, it's things that are as crude as video game footage, because again, it's telling people what they want to believe w- in a situation where they're not very skeptical.

    18. HS

      You said there about kind of confirming existing beliefs. Does that distinction matter though? Because actually, you could have someone who is naturally... You know, we've had riots in the UK in recent... in the last few weeks. And actually, you could have AI-generated images with many more migrants or many more rioters than there actually are, with the incitement that you should join because this is happening. And the confirmation material, which is that AI-generated material, leads to action to take place. So, it doesn't actually matter. The point is, it incites action.

    19. AN

      For sure, yeah. But again, I wanna, I wanna push on, you know, is this really an AI problem? Um, these are, you know, deep problems in our society. So creating an image that, you know, that looks like there were a lot more people there than there were, um, yeah, it's become easier to, to do that with AI today. But you could have paid someone $100 to do that with Photoshop, you know, even before AI. And it's, it's a problem we've had. It's a problem we have been dealing with, often not very successfully. My worry is that if we treat this as a technology problem and try to intervene on the technology, we're gonna miss what the real issues are and, uh, the hard things that we need to be doing to tackle those issues, which are, you know, which relate to issues of trust in society. And to the extent it's a technology problem, it's more of a social media problem really than an AI problem, because the hard part of misinformation is not generating it. It's distributing it to people and persuading them, uh, and social media is often the medium for that. Uh, and so I think there should be, uh, more responsibility placed on social media companies. And my worry is that, you know, treating this as an AI problem is distracting from all of those more important interventions.

    20. HS

      So are social media companies, AI the dis- uh, AKA the distribution platforms, are they the arbiters of justice on what is a malicious AI image versus what isn't?

    21. AN

      Yeah, I think the primary control, uh, is being exercised today by social media companies.

    22. HS

      I feel like I worry more about this than you on the content misinformation side, and so I'm intrigued on the concerns that you have. What would you say is a more pressing concern for you?

    23. AN

      So when we were talking about deepfakes, I'm much less worried about misinformation deepfakes and more worried about, uh, deepfake nudes that I was talking about, right? So those are things that can destroy a person's life. It's been shocking to me how little attention this got from the press and from policymakers until it happened to Taylor Swift a few months ago, uh, and then it got a lot of attention. So there were deepfake nudes of Taylor Swift posted on Twitter/X, uh, and after that, you know, policymakers started paying attention. But it has been happening for many years now, even before the latest wave of generative AI tools. So that's a type of misuse, you know, that, that is very clear. And then there are other kinds of misuses that are not necessarily dangerous in the same way but impose a lot of costs on society. So when students are using AI to, uh, do their, uh, homework, for instance, now, you know, high school teachers and college teachers everywhere have to, uh, revamp how they're teaching in order to account for the fact that students are doing this and there's no way really to catch AI-generated, uh, text or, or homework answers. And so these are costs upon society. I'm not saying that the availability of AI makes education worse. I don't think that's necessarily the case. But, uh, you know, it forces a lot of costs upon the education system, and ideally, AI companies should be bearing some of

  11. 35:5940:29

    Revolutionising Healthcare with AI in Your Pocket

    1. AN

      that cost.

    2. HS

      One thing I think we're just so far off on, Arland, is, uh, the medical, like, um, doctors. Everyone says you're gonna have a GP in your pocket with AI. Are you high? Like, GPs feel your elbow. They look at X-rays. They, um, uh, look inside your ear and see very specific things. They look up your nose. Uh, w- you're not gonna shove your smartphone up your nostril. (laughs) So, uh, you know, it can't feel your arm. Can you help me understand why I'm wrong?... and why AI will revolutionize medical with a GP in everyone's pocket?

    3. AN

      (laughs) Sure. Uh, so I don't think you're wrong. Uh, I think the reason there is a lot of talk about this is, uh, it goes back to something we've observed over and over, which is that when there are problems with an institution, like the medical system, right? Like, the wait times are too long or it's too costly, or, uh, in a lot of countries, you know, people don't even have access. You know, in developing countries there might be entire villages with, uh, no physician. Uh, then this kind of technological bandaid becomes very appealing. So, I think that's what's going on here. I think the responsible way to use AI in medicine is for it to be integrated into the medical system. And actually, the, the medical system has been a very enthusiastic adopter of technology, including AI. So you can consider, you know, CAT scans, for instance, to be a form of AI to be able to reconstruct, um, you know, what's going on, uh, uh, inside a person based on, uh, certain imaging. And now, w- with generative AI as well, there's a lot of interest from the medical system in figuring out, you know, can this be useful for a diagnosis, uh, or for more mundane things like summarizing medical notes and so forth. So, I think that work is really important. I think that should continue. Uh, it still does leave us with a harder question of, you know, uh, here in America, you know, if it takes me three weeks to get a GP appointment, it's very tempting to ask, uh, ChatGPT a question about my symptoms. So, what do we do about that? You know, is that... Can that actually be helpful with appropriate guardrails, or should that be discouraged? I don't know the answer to that.

    4. HS

      (laughs) I'm glad that I'm not alone in my skepticism there. 'Cause applied to education, again, everyone says, "It's amazing. You have a tutor in your pocket." Yeah, I get you, but we do also have your videos that we can watch at home. Like, y- a tutor has personal relationships. It's one-to-one, where I want to impress you, Arvind, and I have that personal desire to fulfill, uh, you know, abilities, potentials that he doesn't have. How do you think AI impacts the future of education, one-on-one tuition, and that upleveling of students?

    5. AN

      I think there's, uh, you know, different populations of students. I think, you know, uh, there's a, uh, small subset of learners who are very self-motivated, will learn very well even if there's no, uh, you know, physical tutor, uh, whether it's at the, uh, uh, the, uh, primary school level or it's at the college level or at the expert level. I think those... There are those kinds of learners at, uh, at all different levels. And then there's the vast majority of learners for whom the social aspect of learning is really the most critical thing, and if you take that away, um, they're just not gonna be able to learn very well. And I think this is often forgotten, especially because in the AI developer community, there are a lot of these, uh, self-taught learners. I- I- I'm among them, right? I j- just paid zero attention throughout school and college, and everything that I know, literally, is stuff that I taught myself. So I grew up in India. The education system wasn't very great there. Uh, our geography teacher thought that India was in the southern hemisphere. True story (laughs) , right? Right, so a- again, I, I literally mean it when I say everything that I know, I taught myself. Um, and so, you know, you have a lot of AI developers who are thinking of themselves as the typical learner, and they're not. And I think for someone like me, AI is, on a daily basis, uh, an incredible, uh, tool for, for learning. I use, uh, you know, uh, uh, generative AI tools for learning, and it's very helpful because it's, it's a new way of learning compared to a book or really anything else. Of course, I haven't stopped using books for learning, but I can't ask a book a question, or, you know, I can't summarize my understanding of my topic to a book and ask it if I'm right. These are things I can do with AI. But I am very skeptical that these new kinds of learning are going to get to a point anytime soon where they're going to become the default way in which people

  12. 40:2941:46

    Is AI Job Replacement Fear Overhyped or Real?

    1. AN

      learn.

    2. HS

      Do you think people dramatically overestimate the fear of job replacement? We always see job replacement with any new technology, and then it tends to create a lot more jobs than, uh, previously were. Uh, do you think that is the case here, or do you think job replacement fears are justified?

    3. AN

      I think for now, they are very much overblown. My favorite example is a thing you said of technology creating jobs is, uh, bank tellers. Uh, when ATMs became a thing, uh, you know, it would have been reasonable to assume that, uh, bank tellers were just gonna go away. But in fact, the number of tellers increased, and the reason for that is that it became much cheaper for banks to open regional branches. And once they did open those regional branches, they didn't need humans for some of the things that you couldn't do with an ATM. You know, the more abstract way of saying that is, uh, as economists would put it, jobs are bundles of tasks, and AI automates tasks, not jobs. So if there are, you know, 20 different tasks that comprise a job, uh, the odds that AI is gonna be able to automate all 20 of them are pretty low. And so there are some occupations, certainly, that have already been affected a lot by AI, like translation or stock photography, but, you know, for, for most jobs out there, I don't think we're anywhere close to

  13. 41:4646:19

    AI's Potential as a Weapon

    1. AN

      that.

    2. HS

      Can I ask... Another one that does worry me is actually defense. You know, we s- had Alex Wang from Scale on, I mentioned earlier. He said that AI has the potential to be a bigger weapon than nuclear weapons. How do you think about that? And if that is the case, should we really have open models?

    3. AN

      I think, you know, it's a, it's a good question to ask. I think it's a bit of a, uh, a category error there. I mean, a nuclear weapon is an actual weapon. AI is not a weapon. AI is something that, you know, might enable, uh, adversaries to do certain things more effectively to, um, you know, f- for example, find, uh, vulnerabilities, cybersecurity vulnerabilities in critical infrastructure, right? So that's one way in which, uh, AI could be used on the, quote unquote, "battlefield." That being the case, I think it would be a big mistake to view it analogously to a weapon and to argue that it should be closed up.... for a couple of reasons. First of all, that's not going to work at all. Uh, so I think we have, uh, you know, closed those state-of-the-art AI models that can already run on people's personal devices, and I think that trend is only going to accelerate. We talked earlier about Moore's law, and it still continues to apply, uh, to these models. And even if one country decides that models should be closed, the odds of getting every country to enact that kind of, uh- uh, rule are, you know, just vanishingly small. So, if our approach to safety with AI is going to be premised on ensuring that, quote-unquote, "bad guys" don't get access to it, we've already lost because it's only a matter of time before it becomes impossible to do that. And instead, I think we should radically embrace the opposite, which is to figure out how we're gonna use AI for safety in a world where AI is very widely available because it is going to be widely available. And when we look at how we've done that in the past, uh, it's actually a very reassuring story. When we go back to the cybersecurity example, for, you know, 10 or 20 years, uh, the software development community has been using automated tools, some of which you could call AI, to improve cybersecurity because software developers can use them to find bugs and fix bugs in software before they put them out there, before hackers even have a chance to take a crack at them. My hope is that the same thing is gonna happen with AI. We're gonna be able to, uh- uh, you know, acknowledge the fact that it's gonna be widely available and to shape its use for defense more than offense.

    4. HS

      Arvin, what did you believe about, kind of, the developments we've seen in AR-... AI over the last two years that you now no longer believe?

    5. AN

      (laughs) So, I think, like a lot of people, I was fooled by how quickly after GPT-3.5, GPT-4 came out. It was just, you know, uh, three months or so, but it had been in training for 18 months. That was only revealed later. So, it gave a lot of people, including me, uh, an inflated idea of how quickly AI has- ha- uh, was progressing. And what we've seen in the nearly year and a half since GPT-4 came out is that we haven't really had models that have, uh, surpassed it in a meaningful way. Uh, and this is not based on benchmarks. Again, I think benchmarks are not that useful. It's more based on vibes. When you get people using these things, what do they say? I don't think models have, you know, really qualitatively improved on- on GPT-4 and I don't think things are moving as quickly as I did 12 months ago.

    6. HS

      And the reasons for that lack of progression? Sorry.

    7. AN

      Making models bigger and bigger doesn't seem to be working anymore. I think new developments have to come from different scientific ideas. Maybe it's agents, maybe it's something else.

    8. HS

      What do you think society's biggest misconception of AI is today?

    9. AN

      I think our intuitions are too powerfully shaped by sci-fi portrayals of AI, and I think that's really a big problem. Uh, you know, this idea that AI can become self-aware. When we look at the way that AI is architected today, that kind of fear has no basis in reality. Uh, maybe one day in the future, you know, people are going to build, uh, AI systems where that becomes, uh, at least somewhat possible, and we should, you know, we should have, uh, visibility, transparency, monitoring regulation around these systems to make sure that developers don't, but that would be a choice. That's a choice that society can make, that governments and companies can make. It's not that despite our best efforts, AI is going to become conscious and have agency and do things that are harmful to humanity. That whole line of fear, I think, is, uh, completely

  14. 46:1950:20

    Quick-Fire Round

    1. AN

      unfounded.

    2. HS

      Arvin, I'd love to do a quick-fire round. I- I could talk to you all day, but I'd love to give you a quick-fire round. So, I say a short statement, you give me your immediate thoughts. Does that sound okay?

    3. AN

      (laughs) Uh, let's do it.

    4. HS

      Why are AI leaderboards no longer useful?

    5. AN

      Because the gap between benchmarks and the real world is big and it's only growing bigger. As AI becomes more useful, it's harder to figure out how useful it is based on these artificial environments.

    6. HS

      If you were CEO of OpenAI for a day, what would you do?

    7. AN

      (laughs) Uh, I would resign. I don't think I would be a good CEO. But if, uh, there were one thing I could change about OpenAI, I think the need for the public to know what is going on with AI development overrides the, uh, you know, commercial interests of any company. So, I think there needs to be a lot more transparency.

    8. HS

      What is your vision for the future of agents?

    9. AN

      Uh, so, my hope is that the kind of thing we saw in the movie Her, uh, not the sci-fi aspects of it, but the more kind of mundane aspects of it where you give your device a command and it interprets it in a pretty nuanced way and does what you want it to do, right? Like, book flight tickets, for instance. Uh, or really build an app based on what you want it to look like. So, these are things that are potentially automatable, don't have, like, massively dubious societal consequences. Those are the things that I hope can happen.

    10. HS

      Will companies increasingly move into the chip and compute layer and compete with NVIDIA, or do you think it will be a continuous NVIDIA monopoly, all of them buying from NVIDIA?

    11. AN

      I do find it interesting that NVIDIA itself has been trying to migrate really, really hard out of hardware into- into becoming a services company.

    12. HS

      And everyone tries to migrate into their business.

    13. AN

      Right, right.

    14. HS

      Why is tech policy frustrating 90% of the time?

    15. AN

      A lot of technologists, uh, you know, kind of have a disdain for policy. They see policymakers as, you know, well, morons, to put it bluntly. But I don't think that's the case. I think there are a lot of legitimate. reasons why policy is very slow and doesn't often go in the way that a tech expert might want it to, um, and that's the 90% frustration. And the reason it- w- I say it's only 90% is that the other 10% is really worth it. We really need policy, and despite how frustrating it is, we need a lot of tech experts in policy.

    16. HS

      Which nation will be an AI leader and which will be an AI laggard?

    17. AN

      I think when it comes to generative AI, um, I- you know, I think, uh, I really like, uh, the- the US approach where, uh, there has been, um, an understanding of the need to regulate, but without being so premature that it, uh, stifles, uh, the development of the technology. I haven't looked that closely into policy in other countries, but I'm optimistic about the future of AI development in the US.

    18. HS

      Penultimate one. Yann LeCun or Geoff Hinton, which side are you on?

    19. AN

      I have to say, I really like, uh, Yann LeCun's perspectives, uh, on various things, including his view that LLMs are, quote-unquote, "off-ramp to superintelligence." That, you know, in other words, we need a lot more scientific breakthroughs as well as tamping down the fears of, uh, super advanced AI.

    20. HS

      What question are you never asked that you should be asked?

    21. AN

      You know, it's- it's weird for me to be saying this, but I have to say, think of the children. I'm never asked this because... And what I mean by that is that AI- the role of AI in kids' lives, kids who are born today, for instance, is going to be so profound, and it's something that technologists should be thinking about, every parent should be thinking about, policymakers should be thinking about because it can be profoundly good or profoundly bad or anything in between. And both as a technologist and as a parent, I think about that a lot.

    22. HS

      Listen, Arvin, as I said, I- I've loved your writing. I can't wait for the book. Thank you so much for putting up with my, uh, deviating questions, but I've so enjoyed having you on the show.

    23. AN

      This has been really, really fun. I apologize for rambling occasionally, but I hope that it's, uh... Yeah, I'm re- really looking forward to hearing it when it's, uh- when it's out there.

Episode duration: 50:20

Install uListen for AI-powered chat & search across the full episode — Get Full Transcript

Transcript of episode 8CvjVAyB4O4

Get more out of YouTube videos.

High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.

Add to Chrome