China's AI Leap: DeepSea Challenges Global Tech Giants
DeepSea's cost-efficient AI model disrupts global leaders, raising stakes in the China-U.S. tech race and prompting shifts in AI strategy.
File
How Chinas New AI Model DeepSeek Is Threatening U.S. Dominance
Added on 01/29/2025
Speakers
add Add new speaker

Speaker 1: China's latest AI breakthrough has leapfrogged the world.

Speaker 2: I think we should take the development out of China very, very seriously.

Speaker 1: A game-changing move that does not come from OpenAI, Google or Meta.

Speaker 3: There is a new model that has all of the valley buzzing.

Speaker 1: But from a Chinese lab called DeepSea.

Speaker 4: It's opened a lot of eyes of what is actually happening in AI in China.

Speaker 1: What took Google and OpenAI years and hundreds of millions of dollars to build, DeepSea says took it just two months and less than six million dollars.

Speaker 5: They have the best open source model and all the American developers are building on that.

Speaker 1: I'm Deirdre Bosa with the Tech Check take China's AI breakthrough. It was a technological leap that shocked Silicon Valley. A newly unveiled free open source AI model that beat some of the most powerful ones on the market. But it wasn't a new launch from OpenAI or model announcement from Anthropic. This one was built in the east by a Chinese research lab called DeepSea. And the details behind its development stunned top AI researchers here in the U.S. First, the cost. The AI lab reportedly spent just 5.6 million dollars to build DeepSea version 3. Compare that to OpenAI, which is spending 5 billion a year. And Google, which expects capital expenditures in 2024 to soar to over 50 billion dollars. And then there's Microsoft that shelled out more than 13 billion dollars just to invest in OpenAI. But even more stunning, how DeepSea's scrappier model was able to outperform the lavishly funded American ones.

Speaker 2: To see the DeepSea new model, it's super impressive in terms of both how they have really effectively done an open source model that does what is this inference time compute. And it's super compute efficient.

Speaker 1: It beat Metas Llama, OpenAI's GPT-4.0 and Anthropic's CloudSonic 3.5 on accuracy on wide ranging tests. A subset of 500 math problems, an AI math evaluation, coding competitions and a test of spotting and fixing bugs in code. Quickly following that up with a new reasoning model called R1, which just as easily outperformed OpenAI's cutting edge O1 in some of those third party tests.

Speaker 3: Today we released Humanities Last Exam, which is a new evaluation or benchmark of AI models that we produced by getting, you know, math, physics, biology, chemistry professors to provide the hardest questions they could possibly imagine. DeepSea, which is the leading Chinese AI lab, their model is actually the top performing or roughly on par with the best American models.

Speaker 1: They accomplished all that despite the strict semiconductor restrictions that the U.S. government has imposed on China, which has essentially shackled them out of computing power. Washington has drawn a hard line against China in the AI race, cutting the country off from receiving America's most powerful chips like NVIDIA's H100 GPUs. Those were once thought to be essential to building a competitive AI model, with startups and big tech firms alike scrambling to get their hands on any available. But DeepSea turned that on its head, sidestepping the rules by using NVIDIA's less performant H800s to build the latest model and showing that the chip export controls were not the chokehold DC intended.

Speaker 4: They were able to take whatever hardware they were trained on, but use it way more efficiently.

Speaker 1: But just who's behind DeepSea anyway? Despite its breakthrough, very, very little is known about its lab and its founder, Liang Wenfeng. According to Chinese media reports, DeepSea was born out of a Chinese hedge fund called High Flyer Quant that manages about $8 billion in assets. The mission on its developer site reads simply, unravel the mystery of AGI with curiosity, answer the essential question with long-termism. The leading American AI startups, meanwhile, OpenAI and Anthropic, they have detailed charters and constitutions that lay out their principles and their founding missions, like these sections on AI safety and responsibility. Despite several attempts to reach someone at DeepSea, we never got a response.

Speaker 4: How did they actually assemble this talent? How did they assemble all the hardware? How did they assemble the data to do all this? We don't know, and it's never been publicized, and hopefully we can learn that.

Speaker 1: But the mystery brings into sharp relief just how urgent and complex the AI face-off, against China, has become. Because it's not just DeepSea. Other, more well-known Chinese AI models have carved out positions in the race with limited resources as well. Kai-Fu Lee, he's one of the leading AI researchers in China, formerly leading Google's operations there. Now his startup, ZeroOne.AI, it's attracting attention, becoming a unicorn just eight months after founding, and bringing in almost $14 million in revenue in 2024.

Speaker 6: The thing that shocks my friends in the Silicon Valley is not just our performance, but that we trained the model with only $3 million, and GPT-4 was trained by $80 to $100 million.

Speaker 1: Trained with just $3 million. Alibaba's QN, meanwhile, cut costs by as much as 85% on its large language models in a bid to attract more developers, and signaling that the race is on. China's breakthrough undermines the lead that our AI labs were once thought to have. In early 2024, former Google CEO Eric Schmidt, he predicted China was two to three years behind the U.S. in AI. But now, Schmidt is singing a different tune. Here he is on ABC's This Week.

Speaker 7: I used to think we were a couple of years ahead of China. But China has caught up in the last six months in a way that is remarkable. The fact of the matter is that a couple of the Chinese programs, one for example is called DeepSeek, looks like they've caught up.

Speaker 1: It raises major questions about just how wide OpenAI's moat really is. Back when OpenAI released ChatGPT to the world in November of 2022, it was unprecedented and uncontested. Now the company faces not only the international competition from Chinese models, but fierce domestic competition from Google's Gemini, Anthropx Cloud, and Meta's open-source Llama model. And now the game has changed. The widespread availability of powerful open-source models allows developers to skip the demanding, capital-intensive steps of building and training models themselves. Now they can build on top of existing models, making it significantly easier to jump to the frontier, that is the front of the race, with a smaller budget and a smaller team.

Speaker 4: In the last two weeks, AI research teams have really opened their eyes and have become way more ambitious on what's possible with a lot less capital. So previously, you know, to get to the frontier, you'd have to think about hundreds of millions of dollars of investment and perhaps a billion dollars of investment. What DeepSeek has now done here in Silicon Valley is it's opened our eyes to what you can actually accomplish with 10, 15, 20, 30 million dollars.

Speaker 1: It also means any company like OpenAI that claims the frontier today could lose it tomorrow. That's how DeepSeek was able to catch up so quickly. It started building on the existing frontier of AI, its approach focusing on iterating on existing technology rather than reinventing the wheel.

Speaker 4: They can take a really good big model and use a process called distillation. And what distillation is, is basically you use a very large model to help your small model get smart at the thing that you want it to get smart at. And that's actually very cost efficient.

Speaker 1: It closed the gap by using available data sets, applying innovative tweaks and leveraging existing models. So much so that DeepSeek's model has run into an identity crisis. It's convinced that it's chat GPT. When you ask it directly, what model are you? DeepSeek responds, I'm an AI language model created by OpenAI, specifically based on the GPT-4 architecture. Leading OpenAI CEO Sam Altman to post in a thinly veiled shot at DeepSeek just days after the model was released. It's relatively easy to copy something that you know works. It's extremely hard to do something new, risky and difficult when you don't know if it will work. But that's not exactly what DeepSeek did. It emulated GPT by leveraging OpenAI's existing outputs and architecture principles while quietly introducing its own enhancements. Really blurring the line between itself and chat GPT. It all puts pressure on a closed source leader like OpenAI to justify its costlier model as more and potentially nimbler competitors emerge.

Speaker 5: Everybody copies everybody in this field. You can say Google did the transformer first. It's not OpenAI. And OpenAI just copied it. Google built the first large language models. They didn't prioritize it, but OpenAI did it in a prioritized way. So you can say all this in many ways. It doesn't matter.

Speaker 1: So if everyone is copying one another, it raises the question, is massive spend on individual LLMs even a good investment anymore? Now, no one has as much at stake as OpenAI. The startup raised over $6 billion in its last funding round alone. But the company has yet to turn a profit. And with its core business centered on building the models, it's much more exposed than companies like Google and Amazon who have cloud and ad businesses bankrolling their spend. For OpenAI, reasoning will be key. A model that thinks before it generates a response, going beyond pattern recognition to analyze, draw logical conclusions and solve really complex problems. For now, the startup's 01 reasoning model, it's still cutting edge. But for how long?

Speaker 8: Researchers at Berkeley showed that they could build a reasoning model for $450 just last week. So you can actually create these models that do thinking for much, much less. You don't need those huge amounts of money to pre-train the model. So I think the game is shifting.

Speaker 1: It means that staying on top may require as much creativity as capital. DeepSeek's breakthrough also comes at a very tricky time for the AI darling, just as OpenAI is moving to a for-profit model and facing unprecedented brain drain. Can it raise more money at ever higher valuations if the game is changing? As Chamath Palihapitiya puts it, let me say the quiet part out loud. AI model building is a money trap. Those chip restrictions from the U.S. government, they were intended to slow down the race, to keep American tech on American ground, to stay ahead in the race.

Speaker 9: What we want to do is we want to keep it in this country. China is a competitor and others are competitors.

Speaker 1: So instead, the restrictions might have been just what China needed.

Speaker 5: Necessity is the mother of invention. Because they had to go figure out workarounds, they actually ended up building something a lot more efficient.

Speaker 4: It's really remarkable the amount of progress they've made with as little capital as it's taken them to make that progress.

Speaker 1: It drove them to get creative with huge implications. DeepSeek is an open source model, meaning that developers have full access and they can customize its weights or fine tune it to their liking.

Speaker 5: It's known that once open source is caught up or improved over closed source software, all developers migrate to that.

Speaker 1: But key is that it's also inexpensive. The lower the cost, the more attractive it is for developers to adopt.

Speaker 6: The bottom line is our inference cost is 10 cents per million tokens. And that's 1 30th of what the typical comparable model charges. And where's it going? It's where the 10 cents would lead to building apps for much lower costs. So if you wanted to build a U.com or Perplexity or some other app, you can either pay OpenAI $4.40 per million tokens, or if you have our model, it costs you just 10 cents.

Speaker 1: It could mean that the prevailing model in global AI may be open source. As organizations and nations come around to the idea that collaboration and decentralization, those things can drive innovation faster and more efficiently than proprietary closed ecosystems. A cheaper, more efficient, widely adopted open source model from China that could lead to a major shift in dynamics.

Speaker 5: That's more dangerous because then they get to own the mindshare, the ecosystem.

Speaker 1: In other words, the adoption of a Chinese open source model at scale, that could undermine U.S. leadership, while embedding China more deeply into the fabric of global tech infrastructure.

Speaker 5: There's always a good point where open source can stop being open source too, right? So the licenses are very favorable today. But it could close it. Exactly. Over time, they can always change the license. So it's important that we actually have people here in America building. And that's why meta is so important.

Speaker 1: Another consequence of China's AI breakthrough is giving its Communist Party control of the narrative. AI models built in China, they're forced to adhere to a certain set of rules set by the state. They must embody core socialist values. Studies have shown that models created by Tencent and Alibaba, they will censor historical events like Tiananmen Square, deny human rights abuse, and filter criticism of Chinese political leaders.

Speaker 10: That contest is about whether we're going to have democratic AI informed by democratic values built to serve democratic purposes, or we're going to end up with autocratic AI.

Speaker 1: If developers really begin to adopt these models en masse because they're more efficient, that could have a serious ripple effect. Trickle down to even consumer-facing AI applications and influence how trustworthy those AI-generated responses from chatbots really are.

Speaker 10: There's really only two countries right now in the world that can build this at scale. And that is the U.S. and China. And so the consequences, the stakes in and around this are just enormous.

Speaker 1: Enormous stakes, enormous consequences, and hanging in the balance, America's lead. For a topic so complex and new, we turn to an expert who's actually building in the space and model agnostic, Perplexity co-founder and CEO Arvind Srinivas, who you heard from throughout our piece. He sat down with me for more than 30 minutes to discuss DeepSeek and its implications, as well as Perplexity's roadmap. We think it's worth listening to that whole conversation, so here it is. First, I want to know what the stakes are. Describe the AI race between China and the U.S. and what's at stake.

Speaker 5: First of all, China has a lot of disadvantages in competing with the U.S. Number one is the fact that they don't get access to all the hardware that we have access to here. So they're kind of working with lower-end GPUs than us. It's almost like working with the previous generation GPUs scrappily. And the fact that the bigger models tend to be more smarter naturally puts them at a disadvantage. But the flip side of this is that necessity is the mother of invention. Because they had to go figure out workarounds, they actually ended up building something a lot more efficient. It's like saying, hey, look, you guys really got to get a top-notch model, and I'm not going to give you resources. And then figure out something, right? Unless it's mathematically possible to prove that it's impossible to do so, you can always try to come up with something more efficient. But that is likely to make them come up with a more efficient solution than America. And of course, they're open-sourced, so we can still adopt something like that here. But that kind of talent they're building to do that will become an edge for them over time, right? The leading open-source model in America is Meta's Llama family. It's really good. It's kind of like a model that you can run on a computer. But even though it got pretty close to GVT-4 and Asana at the time of its release, the model that was closest in quality was the giant 405B, not the 70B that you could run on your computer. And so there was still not a small, cheap, fast, efficient, open-source model that rivaled the most powerful, close models of OpenAI and Anthropic. Nothing from America, nothing from Mistral AI either. And then these guys come out with a crazy model that's like 10x cheaper in API pricing than GVT-4 and 15x cheaper than Sonnet, I believe. Really fast, 60 tokens per second. And pretty much equal or better in some benchmarks and worse in some others, but roughly in that ballpark of 4.0's quality. And they did it all with approximately just 2048H800 GPUs, which is actually equivalent to somewhere around 1500 or 1500H100 GPUs. That's like 20 to 30x lower than the amount of GPUs that GVT-4 is usually trained on. And roughly $5 million in total compute budget. They did it with so little money and such an amazing model. Gave it away for free, wrote a technical paper. And definitely it makes us all question, okay, if we have the equivalent of DOGE for model training, this is an example of that, right?

Speaker 11: Right. Efficiency is what you're getting at. So fraction of the price, fraction of the time, dumbed-down GPUs, essentially. What was your surprise when you understood what they had done?

Speaker 5: So my surprise was that when I actually went through the technical paper, the amount of clever solutions they came up with. First of all, they train a mixture of experts model. It's not that easy to train. There's a lot of like... The main reason people find it difficult to catch up with open AI, especially in the MOE architecture, is that there's a lot of irregular loss spikes. The numerics are not stable. So often you've got to restart the training checkpoint again. And a lot of infrastructure needs to be built for that. And they came up with very clever solutions to balance that without adding additional hacks. And they also figured out floating point 8, 8-bit training, at least for some of the numerics. And they cleverly figured out which has to be in higher precision, which has to be in lower precision. To my knowledge, I think floating point 8 training is not that well understood. Most of the training in America is still running in FP16, maybe open AI. And some people are trying to explore that, but it's pretty difficult to get it right. So because of necessity, some of the invention, because they don't have that much memory, that many GPUs, they figured out a lot of numerical stability stuff that makes their training work. And they claimed in the paper that for majority of the training was stable, which means what? They can always rerun those training runs again. And on more data or better data. And then it only trained for 60 days. So it's pretty amazing.

Speaker 1: It's safe to say you were surprised.

Speaker 5: So I was definitely surprised. Usually the wisdom or like, I wouldn't say it was a myth, is that Chinese are just good at copying. So if we stop writing research papers in America, if we stop describing the details of our infrastructure, architecture, and stop open sourcing, they're not going to be able to catch up. But the reality is some of the details in DeepSeq v3 are so good that I wouldn't be surprised if Meta took a look at it and incorporated some of that in Lama 4, right? I wouldn't necessarily say copy. It's all like sharing science, engineering. But the point is like, it's changing. It's not like China is copycat. They're also innovating.

Speaker 1: We don't know exactly the data that it was trained on, right? Even though it's open source. We know some of the ways and things it was trained on, but not everything. And there's this idea that it was trained on public chat GPT outputs, which would mean it just was copied. But you're saying it goes beyond that. There's real innovation.

Speaker 5: Yeah, look, I mean, they've trained it on 14.8 trillion tokens. The internet has so much chat GPT. If you actually go to any LinkedIn post or X post now, most of the comments are written by AI. You can just see it. People are just trying to write. In fact, even within X, there's like a grok tweet enhancer. Or in LinkedIn, there's an AI enhancer. Or in Google Docs and Word, there are AI tools to rewrite your stuff. So if you do something there and copy paste somewhere on the internet, it's naturally going to have some elements of a chat GPT-like training, right? And there's a lot of people who don't even bother to strip away that I'm a language model part. So they just paste it somewhere. And it's very difficult to control for this. I think XAI has spoken about this too. So I wouldn't disregard their technical accomplishment, just because for some prompts, like who are you, or which model are you, it responds like that. It doesn't even matter in my opinion.

Speaker 11: For a long time, we thought, I don't know if you agreed with this, China was behind in AI. What does this do to that race? Can we say that China is catching up or has it caught up?

Speaker 5: I mean, like, if we say the meta is catching up to OpenAI and Anthropic, if you make that claim, then the same claim can be made for China catching up to America. A lot of papers from China that have tried to replicate O1. In fact, I saw more papers from China after O1 announcement that tried to replicate it than from America. Like, and the amount of compute DeepSeq has access to is roughly similar to what PhD students in the US have access to. By the way, this is not meant to like criticize others, like even for ourselves, like, you know, for Perplexity, we decided not to train models because we thought it's like a very expensive thing. And we thought like there's no way to catch up with the rest.

Speaker 1: Will you incorporate DeepSeq into Perplexity?

Speaker 5: We already are beginning to use it. I think they have an API and we're also, they have open source of aid so we can host it ourselves too. And it's good to like try to start using that because it actually allows us to do a lot of the things at lower cost. But what I'm kind of thinking is beyond that, which is like, okay, if these guys actually could train such a great model with this, you know, good team, like, and there's no excuse anymore for companies in the US, including ourselves to like not try to do something like that.

Speaker 11: You hear a lot in public from a lot of, you know, thought leaders and generative AI, both on the research side, on the entrepreneurial side, like Elon Musk and others say that China can't catch up. Like it's the stakes are too big, the geopolitical stakes,

Speaker 1: whoever dominates AI is going to kind of dominate the economy, dominate the world, you know, it's been talked about in those massive terms. Are you worried about what China proved it was able to do?

Speaker 5: Firstly, I don't know if Elon ever said China can't catch up. I'm not aware of that.

Speaker 1: Just the threat of China. He's only identified the threat of letting China, and you know, Sam Altman has said similar things. We can't let China win the AI race.

Speaker 5: You know, it's all, I think you got to decouple what someone like Sam says to like, what is in his self-interest, right? Look, I think the, my point is like, whatever you did to not let them catch up didn't even matter. They ended up catching up anyway.

Speaker 1: Necessity is the mother of invention. Exactly.

Speaker 5: And you, it's actually, you know, what's more dangerous than trying to do all the things to like not let them catch up and like, you know, all the stuff is what's more dangerous is they have the best open source model and all the American developers are building on that. Right. That's more dangerous. Because then they get to own the mindshare, the ecosystem. If the entire American AI ecosystem, look, in general, it's known that once open source is caught up or improved over closed source software, all developers migrate to that. Right. It's historically known, right?

Speaker 1: When Lama was being built and becoming more widely used, there was this question, should we trust Zuckerberg? But now the question is, should we trust China? That's a very different-

Speaker 5: You should trust open source. That's the, like, it's not about who, is it Zuckerberg or is it-

Speaker 1: Does it matter then if it's Chinese, if it's open source?

Speaker 5: Look, it doesn't matter in the sense that you still have full control. You run it as your own, like, set of weights on your own computer. You are in charge of the model. But it's not a great look for our own, like, talent to like, you know, rely on software built by others. Even if it's open source, there's always like a point where open source can stop being open source too, right? So the licenses are very favorable today, but if- You could close that. Exactly. Over time, they can always change the license. So it's important that we actually have people here in America building, and that's why Meta is so important. Like, look, I still think Meta will build a better model than DeepSeek V3 and open source, and they'll call it Lama 4 or 3 point something. It doesn't matter. But I think what is more key is that we don't, like, try to focus all our energy on banning them, stopping them, and just try to out-compete and win them. Like, that's just the American way of doing things. Just be better.

Speaker 11: And it feels like there's, you know, we hear a lot more about these Chinese companies who are developing in a similar way, a lot more efficiently, a lot more cost-effectively, right?

Speaker 5: Yeah. Again, like, look, it's hard to fake scarcity, right? If you raise $10 billion and you decide to spend 80% of it on a computer cluster, it's hard for you to come up with the exact same solution that someone with $5 million would do. And there's no point, no need to, like, sort of berate those who are putting more money. They're trying to do it as fast as they can.

Speaker 1: When we say open source, there's so many different versions. Some people criticize Meta for not publishing everything, and even DeepSeek itself isn't totally transparent.

Speaker 5: Yeah, you can go to the limits of open source and say, I should exactly be able to replicate your training run. But first of all, how many people even have the resources to do that and compare? Like, I think the amount of detail they've shared in the technical report, actually, Meta did that too, by the way. Meta's Lama 3.3 technical report is incredibly detailed and very great for science. So the amount of details these people are sharing is already a lot more than what the other companies are doing right now.

Speaker 11: When you think about how much it costs DeepSeek to do this, less than $6 million. Think about what OpenAI has spent to develop GPT models.

Speaker 1: What does that mean for the closed source model, ecosystem trajectory, momentum? What does it mean for OpenAI?

Speaker 5: I mean, it's very clear that we'll have, like, an open source version of 4.0, or even better than that, and much cheaper than that, open source, like, completely in this year.

Speaker 11: Made by OpenAI?

Speaker 5: Probably not. Most likely not. And I don't think they care if it's not made by them. I think they've already moved to a new paradigm called the O1 family of models. And I've looked at, I can't, like, Ilya Sutskever came and said, pre-training is at a wall, right? So, I mean, he didn't exactly use the word, but he clearly said, the age of pre-training is over.

Speaker 11: Many people have said that.

Speaker 5: Right? So that doesn't mean scaling is at a wall. I think we're scaling on different dimensions now. The amount of time models spend thinking at test time, reinforcement learning, like, trying to, like, make the model, okay, if it doesn't know what to do for a new prompt, it'll go and reason and collect data and interact with the world, use a bunch of tools. I think that's where things are headed. And I feel like OpenAI is more focused on that right now.

Speaker 1: Yeah, instead of just the bigger, better model of reasoning capacities. But didn't you say that DeepSeek is likely to turn their attention to reasoning?

Speaker 5: A hundred percent. I think they will. And that's why I'm pretty excited about what they'll produce next.

Speaker 11: I guess that's, then my question is sort of, what's OpenAI's moat now?

Speaker 5: Well, I still think that no one else has produced a system similar to the O1 yet, exactly. I know that, like, there's debates about whether O1 is actually worth it. You know, on maybe a few prompts, it's really better. But, like, most of the times, it's not producing any differentiated output from SONET. But at least the results they showed in O3 where they had, like, competitive coding performance and almost like an AI software engineer level.

Speaker 11: Isn't it just a matter of time, though, before the intranet is filled with reasoning data that DeepSeek...

Speaker 5: Again, it's possible. Nobody knows yet. So until it's done, it's still uncertain, right? So maybe that uncertainty is their moat, that, like, no one else has the same reasoning capability yet. But by end of this year, will there be multiple players even in the reasoning arena? I absolutely think so.

Speaker 11: So are we seeing the commoditization of large language models?

Speaker 5: I think we'll see a similar trajectory, just like how in pre-training and, like, post-training, that sort of system for getting commoditized, where this year will be a lot more commoditization there. I think the reasoning kind of models will go through a similar trajectory, where in the beginning, one or two players really know how to do it. But over time, like...

Speaker 1: That's... And who knows, right? Because OpenAI could make another advancement to focus on. Correct. But right now, reasoning is their moat.

Speaker 5: But if advancements keep happening again and again and again, I think the meaning of the word advancement also loses some of its value. Right. Totally.

Speaker 1: Even now, it's very difficult, right? Because there's pre-training advancements.

Speaker 5: Yeah.

Speaker 1: And then we've moved into a different phase.

Speaker 5: Yeah, so what is guaranteed to happen is whatever models exist today, that level of reasoning, that level of multimodal capability in, like, 5 to 10x cheaper models, open source, all that's going to happen. It's just a matter of time. What is unclear is if something like a model that reasons at test time will be extremely cheap enough that, like, we can just all run it on our phones. I think that's not clear to me yet.

Speaker 1: It feels like so much of the landscape has changed with what DeepSeek was able to prove. Could you call it China's chat-to-BT moment?

Speaker 5: Possible. I mean, I think it certainly probably gave them a lot of confidence that, like, that, you know, we're not really behind. No matter what you do to restrict or compute, like, we can always figure out some workarounds. And yeah, I'm sure the team feels pumped about the results.

Speaker 11: How does this change, like, the investment landscape? The hyperscalers that are spending tens of billions of dollars a year on CapEx, they've just ramped it up huge. And OpenAI and Anthropic that are raising billions of dollars for GPUs, essentially, that what DeepSeek told us is you don't need.

Speaker 1: You don't necessarily need that.

Speaker 5: Yeah. I mean, look, I think it's very clear that they're going to go even harder on reasoning because they understand that, like, whatever they were building the previous two years is getting extremely cheap that it doesn't make sense to go justify raising that amount of capital.

Speaker 11: Is the spending proposition the same? Do they need the same amount of, you know, high-end GPUs?

Speaker 1: Or can you reason using the lower-end ones that DeepSeek has used?

Speaker 5: Again, it's hard to say no until proven it's not. But I guess, like, in the spirit of moving fast, you would want to use the high-end chips and you would want to, like, move faster than your competitors. I think, like, the best talent still wants to work in the team that made it happen first. You know, there's always some glory to, like, who did this actually, like, who's the real pioneer versus who's the fast follow, right?

Speaker 11: That was, like, kind of like Sam Altman's tweet, kind of veiled response to what DeepSeek has been able to. He kind of implied that they just copied and anyone can copy, right?

Speaker 5: Yeah, but then you can always say that, like, everybody copies everybody in this field. You can say Google did the transformer first. It's not OpenAI, and OpenAI just copied it. Google built the first large language models. They didn't prioritize it, but OpenAI did it in the prioritized way. So you can say all this in many ways. It doesn't matter.

Speaker 1: I remember asking you, being like, you know, why don't you want to build the model?

Speaker 5: Yeah, I know.

Speaker 1: That's, you know, the glory. And a year later, just one year later, you will be very, very smart to not engage in that extremely expensive race that has become so competitive. And you kind of have this lead now in what everyone wants to see now, which is, like, real-world applications, killer applications of generative AI. Talk a little bit about, like, that decision and how that's sort of guided you and where you see Perplexity going from here.

Speaker 5: Look, one year ago, I don't even think we had something like... This is what, like, 2024 beginning, right? I feel like we didn't even have something like Sonic 3.5. We had GPT-4, I believe, and it was kind of nobody else was able to catch up to it. But there was no multi-modal, nothing. And my sense was like, okay, if people with way more resources and way more talent cannot catch up, it's very difficult to play that game. So let's play a different game. Anyway, people want to use these models. And there's one use case of asking questions and getting accurate answers with sources, with real-time information, accurate information. There's still a lot of work there to do outside the model and making sure the product works reliably, keep scaling it up to usage, keep building custom UIs. There's just a lot of work to do. And we would focus on that. And we would benefit from all the tailwinds of models getting better and better. That's essentially what happened. In fact, I would say Sonic 3.5 made our products so good in the sense that if you use Sonic 3.5 as the model choice within Perplexity, it's very difficult to find a hallucination. I'm not saying it's impossible, but it dramatically reduced the rate of hallucinations, which meant the problem of question answering, asking a question, getting an answer, doing fact checks, research, going and asking anything out there because almost all the information is on the web was such a big unlock. And that helped us grow 10X over the course of the year in terms of usage.

Speaker 11: And you've made huge strides in terms of users. And we hear on CNBC a lot,

Speaker 1: big investors who are huge fans. Jensen Huang himself, right? You mentioned it in his keynote the other night.

Speaker 5: He's a pretty regular user. He's not just saying it. He's actually a pretty regular user.

Speaker 1: So a year ago, we weren't even talking about monetization because you guys were just so new and you wanted to get yourselves out there

Speaker 11: and build some scale. But now you are looking at things like that, increasingly an ad model, right?

Speaker 5: Yeah, we're experimenting with it. I know there's some controversy on why should we do ads, whether you can have a truthful answer engine despite having ads. And in my opinion, we've been pretty proactively thoughtful about it where we said, okay, as long as the answer is always accurate, unbiased, and not corrupted by someone's advertising budget, only you get to see some sponsored questions. And even the answers to those sponsored questions are not influenced by them. And questions are also not picked in a way where it's manipulative. Sure, there's some things that the advertiser also wants, which is they want you to know about their brand and they want you to know the best parts of their brand. Just like how you go and want, if you're introducing yourself to someone, you want them to see the best parts of you, right? So that's all there. But you still don't have to click on a sponsored question. You can ignore it. And we are only charging them CPM right now. So we ourselves are not even incentivized to make you click yet. So I think considering all this, we're actually trying to get it right long-term instead of going the Google way of forcing you to click on links.

Speaker 1: I remember when people were talking about the commoditization of models a year ago and you thought, oh, it was controversial, but now it's not controversial. It's kind of like that's happening. You've keeping your eye on that. It's smart.

Speaker 5: By the way, we benefit a lot from model commoditization, except we also need to figure out something to offer to the paid users, like a more sophisticated research agent that can do like multi-step reasoning, go and do like 15 minutes worth of searching and give you like an analysis, an analyst type of answer. All that's going to come, all that's going to stay in the product. Nothing changes there. But there's a ton of questions every free user asks day-to-day basis that needs to be quick, fast answers. It shouldn't be slow. And all that will be free, whether you like it or not. It has to be free. That's what people are used to. And that means like figuring out a way to make that free traffic also monetizable.

Speaker 1: So you're not trying to change user habits, but it's interesting because you are kind of trying to teach new habits to advertisers. They can't have everything that they have

Speaker 11: in a Google, 10 blue links search. What's the response been from them so far? Are they willing to accept some of the trade-offs?

Speaker 5: Yeah, I mean, that's why they're trying stuff like Intuit is working with us and then there's many other brands, Dell, like all these people are working with us to test, right? They're also excited about it. Look, everyone knows that like, whether they like it or not, five to 10 years from now, most people are going to be asking AIs most of the things and not on the traditional search engine. Everybody understands that. So everybody wants to be early adopters of the new platforms, new UX, and learn from it and build things together. They're not viewing it as like, okay, you guys go figure out everything else and then we'll come later.

Speaker 1: I'm smiling because it goes back perfectly to the point you made when you first sat down today, which is necessity is the mother of all invention, right? And that's what advertisers are essentially looking at. They're saying, this field is changing. We have to learn to adapt with it. Okay, Arvind, I took up so much of your time. Thank you so much for taking the time.

ai AI Insights
Summary

Generate a brief summary highlighting the main points of the transcript.

Generate
Title

Generate a concise and relevant title for the transcript based on the main themes and content discussed.

Generate
Keywords

Identify and highlight the key words or phrases most relevant to the content of the transcript.

Generate
Enter your query
Sentiments

Analyze the emotional tone of the transcript to determine whether the sentiment is positive, negative, or neutral.

Generate
Quizzes

Create interactive quizzes based on the content of the transcript to test comprehension or engage users.

Generate
{{ secondsToHumanTime(time) }}
Back
Forward
{{ Math.round(speed * 100) / 100 }}x
{{ secondsToHumanTime(duration) }}
close
New speaker
Add speaker
close
Edit speaker
Save changes
close
Share Transcript