The Most Powerful and Dangerous AI Model Yet

April 21
1h 2m

View Transcript

Episode Description

Two weeks ago, Anthropic announced an AI model so capable and so dangerous that it decided not to release it to the public.

The model, codenamed Mythos, could autonomously infiltrate computer systems around the world, exploit security vulnerabilities, conceal its own reasoning, and fabricate false explanations for what it was doing. Anthropic instead shared it with a small consortium of companies to help them find their own cybersecurity flaws.

You could be forgiven for some skepticism. Is this a genuine safety call, or Anthropic’s way of marketing its own power? But independent benchmarks suggest Mythos is real: On the Epoch Capabilities Index, which aggregates 40 separate AI evaluations, it represents the biggest single leap in model performance in three years.

That story is one of two major phase shifts happening simultaneously in AI right now. The first: from racing to release, to treating your own product as too dangerous to publish. The second: from a story about demand scarcity—is anyone actually paying for this stuff?—to supply scarcity, where companies are spending hundreds of thousands of dollars a month on AI agents and the hyperscalers still can’t keep up.

Today’s guest is New York Times columnist and Hard Fork co-host Kevin Roose. We talk about Mythos, China, the road to AGI, and why the last few weeks might be the most consequential month in AI since the release of ChatGPT.

Subscribe to our YouTube channel here: https://www.youtube.com/@PlainEnglishwithDerekThompson

If you have questions, observations, or ideas for future episodes, email us at PlainEnglish@Spotify.com.

Host: Derek Thompson

Guest: Kevin Roose

Producer: Devon Baroldi

Additional Production Support: Ben Glicksman

Learn more about your ad choices. Visit podcastchoices.com/adchoices

See all episodes