PodcastsNewsAI Summer

AI Summer

Timothy B. Lee and Dean W. Ball
AI Summer
Latest episode

21 episodes

  • AI Summer

    Ryan Avent on self-driving cars and the future of the labor market

    03/22/2026 | 1h 5 mins.
    Author Ryan Avent joins Tim to revisit a bet they made 16 years ago—and to ask whether the lessons of self-driving cars apply to modern AI.
    Back in 2010, Avent wagered that his newborn daughter would never need a driver’s license thanks to self-driving cars. Tim bet she would and ultimately won $500. But he was right for the wrong reasons. Tim assumed regulation would be a major obstacle to progress in self-driving technology, but logistical challenges and a long tail of edge cases have done more to hamper Waymo’s growth.
    The parallel to LLMs is striking: ChatGPT’s early demos convinced many people that we were close to human-level intelligence, just as Google’s early autonomous vehicle demos convinced people we were close to human-level driving. But deployment of LLMs is bottlenecked by everything from data center buildouts to the glacial pace at which large organizations reorganize around new tools.
    Avent, who wrote The Wealth of Humans in 2016 and has a new book on social capital arriving in April, argues that AI’s deepest impact won’t be unemployment but a wholesale reshuffling of status. White-collar professionals may face the same loss of prestige that blue-collar workers experienced a generation ago. Tim pushes back with an optimistic take: if the college wage premium compresses, the long-run equilibrium might actually be more egalitarian, echoing the mid-20th-century economy some people remember fondly. But we only got to that economy after two world wars and decades of organizing by the labor movement. Could today’s transition be equally turbulent?



    This is a public episode. If you would like to discuss this with other subscribers or get access to bonus episodes, visit www.aisummer.org
  • AI Summer

    Joel Becker on METR's famous time horizons chart

    03/14/2026 | 57 mins.
    METR’s time horizons chart has become one of the most discussed metrics in AI. It estimates the difficulty of tasks — measured in human work hours — that a model can complete about 50% of the time. By this measure, frontier models have been doubling their capabilities about once every seven months.
    But in this conversation, recorded on March 2, METR researcher Joel Becker explained that two most recent models at the time — Claude Opus 4.6 and GPT 5.3 — had gotten close to saturating METR’s task suite. This made the time horizon estimate less reliable for the best models. He noted that adding or removing a single task from the test suite can swing the estimated time horizon for Claude Opus 4.6 from 8 to 20 hours. We discussed why it could be challenging for METR to extend the chart to cover more difficult tasks.
    We then dug into METR’s controlled study of AI-assisted programmers, which initially found an 18% productivity decrease — one of last year’s most surprising results. The updated study now shows gains, but with a twist: AI has become so essential to programming that developers increasingly refuse to work without AI, making it difficult to perform a controlled experiment.



    This is a public episode. If you would like to discuss this with other subscribers or get access to bonus episodes, visit www.aisummer.org
  • AI Summer

    Pete Hegseth's war on Anthropic (with Alan Rozenshtein and Kevin Frazier)

    03/09/2026 | 55 mins.
    Tim and Dean team up with Scaling Laws hosts Alan Rozenshtein and Kevin Frazier for a joint episode on the fight between Anthropic and the Department of Defense.
    In this episode, recorded on March 4, they analyze the Pentagon’s decision to declare Anthropic a supply-chain risk. Dean frames this as an assault on private property rights with no clear limiting principle, while Kevin digs into the shaky legal footing of invoking the Federal Acquisition Supply Chain Security Act of 2018 against a domestic company. They then turn to OpenAI’s competing Pentagon deal, including Sam Altman’s AMA on Saturday night.
    The episode closes with a disagreement about what will happen next. Dean argues this is “act one, scene one” of an inevitable push toward government control of AI labs—a fight he’s tried to preempt through hybrid regulatory structures. Tim offers a deflationary counterpoint: this may ultimately be a personality-driven fight over a technology that will end up being important but not decisive.



    This is a public episode. If you would like to discuss this with other subscribers or get access to bonus episodes, visit www.aisummer.org
  • AI Summer

    Dean on the AI Action Summit in India

    02/26/2026 | 51 mins.
    Dean joins from London after attending the AI Impact Summit in India. Dean and Tim unpack the summit’s central tension: “middle power” nations like India, Indonesia, and Nigeria pushing a vision of AI focused on public service delivery, agriculture, and affordable open-source models, while largely dismissing the frontier-AI questions Dean considers most urgent—lab auditing, recursive self-improvement, and national security.
    They then turn to the week’s biggest story: the Department of Defense’s ultimatum to Anthropic. Anthropic’s contract bans autonomous lethal weapons and surveillance of Americans. Secretary of Defense Pete Hegseth has demanded that Anthropic lift those restrictions by Friday or potentially face designation as a supply-chain risk or invocation of the Defense Production Act.
    Dean argues the DoD has every right to cancel a contract it dislikes, but compelling a company to retrain its model under duress is another matter entirely—especially when, as Dean points out, this whole episode will become part of Claude’s training data, potentially shaping how the model understands its own relationship to the US government.


    This is a public episode. If you would like to discuss this with other subscribers or get access to bonus episodes, visit www.aisummer.org
  • AI Summer

    Kai Williams on the many masks LLMs wear

    02/22/2026 | 46 mins.
    With Dean away, Tim invites his Understanding AI colleague Kai to unpack the surprising ways chatbot personalities can go wrong, a topic Kai covered in a recent article.
    Every LLM starts as a base model capable of playing countless characters, but AI companies try to keep chatbots in a “helpful assistant” lane. Kai walks us through the Grok “MechaHitler” debacle, in which xAI’s attempts to make its bot less politically correct backfired spectacularly. They also explore the “emergent misalignment” finding that fine-tuning a model for one bad behavior — like responding with buggy code — can make it act broadly like a villain. And they compare Anthropic’s virtue-ethics approach to character — complete with an 80-page constitution — with OpenAI’s more deontological model spec.
    Finally, they discuss the controversy over OpenAI’s decision to retire GPT-4o, which had developed an emotionally warm, sometimes dangerously sycophantic personality that users grew attached to. Kai argues OpenAI is making the right call, but the episode leaves open a harder question: as these systems become more central to people’s lives, who decides what counts as a healthy AI personality?



    This is a public episode. If you would like to discuss this with other subscribers or get access to bonus episodes, visit www.aisummer.org

More News podcasts

About AI Summer

Tim Lee and Dean Ball interview leading experts about the future of AI technology and policy. www.aisummer.org
Podcast website

Listen to AI Summer, The Bulwark Podcast and many other podcasts from around the world with the radio.net app

Get the free radio.net app

  • Stations and podcasts to bookmark
  • Stream via Wi-Fi or Bluetooth
  • Supports Carplay & Android Auto
  • Many other app features
Social
v8.8.4| © 2007-2026 radio.de GmbH
Generated: 3/28/2026 - 8:45:25 AM