Four big reasons to worry about DeepSeek (and four reasons to calm down)
I. DeepSeek rising
Today, let’s talk about DeepSeek.
On Monday, the Nasdaq fell 3.1 percent as investors considered what the Chinese company’s high-performing, cheap-to-train, App Store-topping new models meant for the future of artificial intelligence and the tech industry overall. The leading maker of specialized AI chips, Nvidia, suffered most of all — its stock price dropped 17 percent, erasing $600 billion in market value.
As news of DeepSeek’s achievement spread over the weekend, it became a kind of Rorschach test. While everyone is impressed that DeepSeek built the best open-weights model available for a fraction of the money that its rivals did, opinions about its long-term significance are all over the map.
To many prominent voices in AI, DeepSeek seems to have confirmed what they already believed. To AI skeptics, who believe that AI costs are so high that they will never be recouped, DeepSeek’s success is evidence of Silicon Valley waste and hubris. To AI bulls, who think America needs to build artificial general intelligence before anyone else as a matter of national security, DeepSeek is a dire warning to move faster. And to AI safety researchers, who have long feared that framing AI as a race would increase the risk of out-of-control AI systems doing catastrophic harm, DeepSeek is the nightmare that they have been waiting for.
Whatever the truth is won’t be known for some time. Reading the coverage over the past few days, and talking with folks who work in the industry, I’m convinced that DeepSeek is a huge story deserving of our ongoing attention. At the same time, I’m not sure that the emergence of a powerful, low-cost Chinese AI model changes the dynamics of competition quite as much as some observers are saying.
With that in mind, here are some reasons worth worrying about DeepSeek — and some reasons to calm down.
II. Reasons to worry
No one really knows what DeepSeek’s long-term game is. As you may know by now, DeepSeek was created by a 10-year-old Chinese quantitative hedge fund named High-Flyer; Liang Wenfeng, DeepSeek’s CEO, is also a cofounder of the fund. High-Flyer developed AI algorithms for use in trading; in 2023, it started a lab to build AI tools unrelated to its core business.
Over the next year or so, it made a series of technical innovations in building large language models. Its stated mission, as posted on its X profile, is to “Unravel the mystery of AGI with curiosity.” The company has committed to open-sourcing its models, and has offered them to developers at very cheap prices.
For the moment, DeepSeek doesn’t seem to have a business model to match its ambitions. For most big US AI labs, the (yet unrealized) business model is to develop the best service and to sell it at a profit. To date, DeepSeek has positioned itself as a kind of altruistic giveaway.
That could change at any time. DeepSeek could introduce subscriptions, or place new restrictions on its developer APIs. Zvi Mowshowitz theorizes that the company could take user data and give it to the hedge fund for trading insights.
And at some point, the Chinese government will have something to say about one of its companies trying to give away powerful AI to anyone who wants it, including China’s adversaries.
In the meantime, though, we can only guess what DeepSeek’s ambitions are. And that worries me, because in some very real sense we don’t know what we’re dealing with here.
The big AI labs don’t seem to have much of a moat. Somewhat lost in the DeepSeek conversation so far is that the company’s impressive v3 and r1 models were built on top of American innovations. It was the US AI labs who developed the underlying architecture for large language models and the newer reasoning models; what DeepSeek has done is to cleverly optimize that architecture using old hardware and less computing power.
In the old days, by which I mean the time of GPT-3, it took OpenAI’s rivals months or longer to reverse-engineer its process and absorb its innovations. It might take a year for those techniques to filter down to the open-source models that are made available for free.
But DeepSeek shows that the open-source labs have gotten much better at reverse-engineering — and that any lead the US AI labs come up with can be quickly erased. This is a problem if your main business is selling models to developers: switching costs are low, and the cost savings they can achieve by using DeepSeek are huge.
For the AI labs, this is a business problem. But it could be a geopolitics problem, too: DeepSeek’s innovation shows that there will be no keeping AI out of anyone’s hands, for better and for worse. As Anthropic co-founder Jack Clark put it today: “DeepSeek means AI proliferation is guaranteed.”
DeepSeek’s success will lead more people to see AI as an arms race against China.
For some venture capitalists in particular, it has long been a goal to frame AI progress as a contest against China. This idea was central in “Situational Awareness,” Leopold Aschenbrenner’s viral essay from last year about AI progress, whose publication coincided with the former AI researcher announcing that he had started a new venture capital firm.
VCs love this framing for lots of reasons. It builds on a rational fear: that an authoritarian government will create superhuman intelligence before democracies do and use it against them. But it is also meant to serve as ammunition against regulations, which would slow both AI progress and returns to VCs’ portfolios; and to drum up interest in military tech, which generates further returns to those same portfolios.
In general I think we should be against the world’s richest people cheerleading us into armed conflict.
The more that people believe AI is an existential race against China, the less safely it will be built.
Say what you will about the failures of US AI labs — and there are many — but they have at least tried to outline methods for building powerful AI safely. DeepSeek, by contrast, has not said one word about AI safety. If it has a lone AI safety researcher, that’s news to me.
To accelerationists, this could be a reason for US companies to abandon their safety efforts — or at least, reduce future investments in them.
It’s important to remember that all of the most important AI safety problems remain unsolved. Should one of the corporate AI labs suddenly invent and release a superhuman intelligence, there is no way to ensure it is aligned with human values or desires, and no plan for what to do next. The Biden administration placed some gentle restrictions on US AI labs with an executive order, but Trump repealed it on day one.
As Mowshowitz writes: “These people really think that the best thing humanity can do is create things smarter than ourselves with as many capabilities as possible, make them freely available to whoever wants one, and see what happens, and assume that this will obviously end well and anyone who opposes this plan is a dastardly villain.”
III. Reasons for calm
Everyone basically already assumed that all of this was going to happen. By “all of this,” I mean that (1) open-source companies would reverse-engineer everything the big labs are doing and (2) that costs for AI training and inference would decline dramatically over time.
Ethan Mollick, a professor at the Wharton School of the University of Pennsylvania, noted over the weekend that “costs for GPT-4 level intelligence dropped by 1000x in the last 18 months.” For that reason, he writes, “A 95% price drop in reasoning models seems not to be something that will break the labs.”
Funnily enough, Google released its own super-cheap reasoning model just five days ago, built on similar techniques, and no one cared at all. (It’s not open source or free or Chinese. But still!)
Anyone who has sent the same query to ChatGPT, Claude, and Gemini on the same day has known for more than a year that you can get basically as good an answer from any of them. And anyone who has used Llama has known for more than a year that the open-weights version that arrives later is only slightly worse.
Right now a lot of investors are catching up to these basic facts at the same time, and stock prices are falling accordingly. But it’s not clear to me that much of it was really news to the AI labs and tech platforms.
Tech companies can still find good use for all those AI capital expenditures.
American tech companies plan to spend tens of billions of dollars building data centers to serve their AI needs this year. One question many folks were asking today was whether DeepSeek would make all of those investments moot. If you can build a best-in-class model with older hardware, what’s the point?
The point is to (1) train more powerful models and explore techniques that open-source developers haven’t ripped off yet; and (2) to serve the demand that those more powerful and capable models generate. While much of the AI discussion over the past six months has revolved around the bottleneck that a lack of chips has created for training new models, the real bottleneck is that no one has as much computing power as they want.
For the most part, the same servers and chips that fund model training can be used for inference. DeepSeek’s innovation means that the day you can run a state-of-the-art model on your laptop is much closer. But we’re not there yet.
The export controls are helping.
Some observers have said that DeepSeek’s progress shows that the Biden administration’s restrictions on chip exports have failed. I don’t think that’s right. As Jordan Schneider writes at ChinaTalk, these export controls are relatively new — and need more time to really have an effect.
A primary effect of the export controls is that China should have less computing power than the United States overall for some time. That means that even as Chinese companies like DeepSeek release more powerful models, China may not be able to deploy them as widely as it would like to.
That same computing power is also essential for inventing more powerful AI systems. As Miles Brundage, a former policy researcher at OpenAI, put it recently on a podcast with Schneider: “There are all sorts of ways of turning compute into better performance, and American companies are currently in a better position to do that because of their greater volume and quantity of chips.”
American AI labs are still in the lead.
DeepSeek’s innovations are real, and they go a long way toward making the AI systems we have today cheaper and more accessible. But to the extent that they represent conceptual breakthroughs, it’s only in optimizing technology that OpenAI and others invented first. There can be a ton of value in optimization, of course, and I’m sure every American company wishes they had come up with these first. (Meta’s teams are now reportedly in overdrive trying to reverse-engineer them.)
At the same time, it was only last week that OpenAI made available to Pro plan users a computer that can use itself. The whole US AI industry has shifted its focus to building AI agents and full-fledged virtual coworkers. To date, very little of that is visible. (And as I wrote last week, OpenAI’s Operator can be unbelievably frustrating to use.) Perhaps DeepSeek or another Chinese company could beat the United States to the punch with agents. But it seems likelier that they will simply wait for an American company to release a good one and try to copy it.
I can understand why some people look at the progress DeepSeek has made and assume that it is about to overtake every US lab. If the company is as careless about safety as it appears to be, there may someday be an actual reason to panic. For the moment, though, I think everyone would benefit from taking a few long deep breaths.
Elsewhere in DeepSeek:
- I learned a ton from Ben Thompson’s comprehensive DeepSeek FAQ, even if I disagree with most of the policy conclusions. (Stratechery)
- Here’s the V3 technical paper that caused such a stir.
- Trump is impressed with DeepSeek, calling it “good, because you don’t have to spend as much money.” (Stephanie Lai and Billy House / Bloomberg)
- DeepSeek limited registrations on Monday amid an alleged cyberattack. (Hayden Field / CNBC)
- For a nice, economical profile of High-Flyer and Deepseek, try this one. (Raffaele Huang / Wall Street Journal)
- Dan Primack suspects DeepSeek could represent an “extinction-level event” for VC firms that over-invested in AI companies that build foundational models. (Axios)
- DeepSeek won’t answer politically sensitive questions about China — unless you host if yourself, in which case it isn’t censored.
- 10 wild DeepSeek demos. (The Neuron)
Governing
- Trump is said to be in talks with multiple prospective TikTok buyers and will likely have a decision in the next 30 days, he said. (Nandita Bose, Dawn Chmielewski, Milana Vinn and Kanishka Singh / Reuters)
- Perplexity has reportedly submitted a revised bid to merge with TikTok that would give the US government up to 50 percent ownership of the merged entity. (Anthony Ha / TechCrunch)
- Perplexity’s revised bid reportedly proposes ByteDance to give TikTok US (without its recommendation algorithm) in exchange for its investors receiving equity in a new company called NewCo, with Perplexity doing the same for its investors. (Hayden Field / CNBC)
- The Trump administration is reportedly working on a plan that would have Oracle and a group of investors take control of TikTok’s global operations. Trump denied this, though. (Bobby Allyn / NPR)
- Some congressional members are confused by Trump’s approach of brokering a “joint venture” deal to save TikTok, with some questioning the legality. (Anthony Adragna and Christine Mui / Politico)
- A look inside TikTok’s game of chicken with the Biden administration to avoid a ban. (Drew Harwell / Washington Post)
- Stargate, the massive AI infrastructure project backed by Donald Trump, will reportedly exclusively serve OpenAI. (George Hammond, Tabby Kinder and Madhumita Murgia / Financial Times)
- Some of Trump’s allies and staff are reportedly “furious” that Elon Musk has publicly criticized Stargate, with one ally accusing Musk of abusing his closeness to Trump. (Dasha Burns and Holly Otterbein / Politico)
- DOGE’s takeover of the existing United States Digital Service signals the likely limits that Musk will have to work under. (Michael C. Bender, Madeleine Ngo and Theodore Schleifer / New York Times)
- Musk has reportedly started conversations with various blockchains about using the technology at DOGE. (Olga Kharif and Stephanie Lai / Bloomberg)
- An in-depth analysis of the various surveillance tools that are at the Trump administration’s disposal to crack down on immigration. (Adam Satariano, Paul Mozur, Aaron Krolik and David McCabe / New York Times)
- Mark Zuckerberg has “a lot more ass-kissing” to do if he wants to curry favor with Trump, a senior Trump administration official said. (Asawin Suebsaeng and Andrew Perez / Rolling Stone)
- Advertisers are concerned that Meta’s content moderation changes could put their own reputations at risk if ads are served near offensive content. (Suzanne Vranica and Patience Haggins / Wall Street Journal)
- Instagram and Facebook have blurred, blocked or removed posts and accounts of some abortion pill providers and hid them from search and recommendations in recent weeks. Some posts and accounts were restored. (Claire Cain Miller, Kate Conger and Mike Isaac / New York Times)
- Character.AI argued its platform is protected against liability by the First Amendment in a motion to dismiss a lawsuit from a parent who said their child died by suicide after developing an emotional attachment to a chatbot. (Kyle Wiggers / TechCrunch)
- Trump signed an executive order that would “promote the development and growth” of “legitimate dollar-backed stablecoins,” and barred work on a central bank digital currency. (Ryan Weeks, Olga Kharif and Suvashree Ghosh / Bloomberg)
- A look at how Trump’s memecoin, $Trump, divided the crypto industry, as some saw the move as a cash grab that could undermine the industry’s legitimacy. (David Yaffe-Bellany and Eric Lipton / New York Times)
- NFTs and memecoins should be viewed as “collectibles” and not securities or commodities, White House crypto czar David Sacks said. (Catherine McGrath / Fortune)
- Google agreed to crack down on businesses in the UK that use fake reviews to boost their ratings on the company’s reviews platform, the UK’s competition regulator said. (Jess Weatherbed / The Verge)
- X, Meta, Snap, Google, TikTok, LinkedIn and its parent Microsoft will be asked by EU officials about their plans to combat disinformation ahead of the next election. (Gian Volpicelli / Bloomberg)
- Israel’s military reportedly relied heavily on Microsoft’s cloud technology and AI systems, and struck at least $10 million in deals, during the most intense phase of its attack on Gaza. (Harry Davies and Yuval Abraham / The Guardian)
- News outlets owned by Indian billionaires Gautam Adani and Mukesh Ambani, along with outlets like the Indian Express and the Hindustan Times, accused OpenAI of copyright infringement. (Aditya Kalra and Arpan Chaturvedi / Reuters)
- X has refused to take down a video depicting a stabbing that was watched by a man who murdered three children in Australia, despite requests from Australian and British authorities. (Anna Gross, Jennifer Williams, Jim Pickard and Nic Fildes / Financial Times)
- Sam Altman’s Tools for Humanity must stop paying people for iris scans in exchange for cryptocurrency and digital IDs, Brazil’s data protection authority said. (Reuters / Economic Times)
- Top AI figures were divided on the safety issues that come with rapidly advancing AI technology at the World Economic Forum. (Stephen Morris / Financial Times)
Industry
- OpenAI’s board is reportedly struggling to determine the price of Microsoft’s stake in the company in its quest to convert to a for-profit company. (Cristina Criddle and George Hammond / Financial Times)
- ChatGPT’s Canvas, a feature that lets users share and collaborate a document or code, now works with the o1 model and can render HTML and React code. (Simon Willison’s Weblog)
- Banks are reportedly reaching out to investors and preparing to sell billions in debt borrowed by X. (Justin Baer, Alexander Saeedy and Alexa Corse / Wall Street Journal)
- The banks looking to sell X debt are reportedly offering potential investors a claim on X’s interest in xAI, based on the latest funding round at a valuation of about $50 billion. (Sirdhar Natarajan and Carmen Arroyo / Bloomberg)
- Meta is set to invest $60 billion to $65 billion in AI in 2025, which includes building an AI engineer to contribute code to research and design efforts, Mark Zuckerberg said. (Jaspreet Singh / Reuters)
- The 2025 spending signifies a 70 percent jump from last year, with the announcement coming just days after OpenAI’s massive Stargate investment. (Meghan Bobrowsky / Wall Street Journal)
- Meta AI will now use account information from the company’s apps to give personalized recommendations, the company said. (Kyle Wiggers / TechCrunch)
- Meta’s Ray-Ban smart glasses’ live translation feature works as intended for the most part, but struggles with slang and with multiple voices at once. (Victoria Song / The Verge)
- Whatever users and engagement were lost by the news of Meta’s fact-checking and DEI changes were gained back after news of the TikTok ban, data shows. (Pranav Dixit / Business Insider)
- Threads is now testing ads with some brands in the US and Japan. (Ivan Mehta / TechCrunch)
- Google is set to invest $120 million in AI education programs and has partnered with community colleges for job training programs that include AI education. (Kenrick Cai / Reuters)
- Apple has reportedly moved Kim Vorrath, vice president of program management who’s known for fixing struggling products, to its AI and machine learning division. (Mark Gurman / Bloomberg)
- Siri with AI is dumber than before, this piece argues, as the old Siri acknowledged when it didn’t know the answer and redirected instead of providing answers full of errors. (John Gruber / Daring Fireball)
- RedNote lost more than half its daily active users in the US – a high of 32.5 million – after Trump paused the TikTok ban. (Sarah Perez / TechCrunch)
- Billionaire Mukesh Ambani is reportedly building a data center in India that is set to become the world’s biggest by capacity, using Nvidia’s AI semiconductors. (Saritha Rai / Bloomberg)
Those good posts
For more good posts every day, follow Casey’s Instagram stories.
(Link)
(Link)
(Link)
Talk to us
Send us tips, comments, questions, and DeepSeek queries: casey@platformer.news. Read our ethics policy here.