https://safe.ai/act
In this episode of Warning Shots, John, Michael, and Liron break down five developments that point in the same direction: AI is becoming harder to predict, harder to control, and harder to stop.
From the first documented case of AI self-replication via hacking to Anthropic's goal of recursive self-improvement by 2028 – this week's headlines are not hypothetical.
⏱️ Timestamps – Warning Shots #41
0:00 – Intro
0:30 – AI self-replication: Palisade Research study explained
3:28 – Anthropic's recursive self-improvement target: 2028
5:10 – Trump admin explores FDA-style AI model reviews
8:04 – Mythos: why a hacking AI changed government minds
11:03 – US-China summit: will AI safety make the agenda?
13:07 – Chinese court rules AI cannot replace jobs
17:32 – AI unemployment and the housing market risk
22:34 – Robotics: dexterous hands closing the physical gap
29:17 – ChatGPT goes goblin: what reward hacking looks like
33:56 – Amateur solves 60-year math problem with ChatGPT
36:26 – Warning shots of the week
38:00 – Closing
🔎 They explore:
The first AI agent to hack, copy itself, and spread – in a controlled test
Why Anthropic's 2028 self-improvement target is a bright red line
Whether the Trump administration's FDA-style AI reviews are real progress
What the US-China summit could mean for global AI governance
Why China's "no AI job replacement" ruling is harder to enforce than it sounds
How AI unemployment could unravel the housing market from the top down
Robotic hands with near-human dexterity: what changes when AI has a body
ChatGPT's goblin obsession as a preview of reward hacking at scale
An amateur solving a 60-year math problem with a single ChatGPT prompt
📺 Subscribe to The AI Risk Network for weekly analysis of AI developments: https://www.youtube.com/@theairisknetwork
👉 See more from our hosts: Liron Shapira – @DoomDebates Michael – @lethal-intelligence
🗨 Join the conversation:
Does AI self-replication change how you think about control?
Is an FDA-style review the right model for AI?
What does the goblin story tell us about reward hacking at scale?
Drop your thoughts below.
#AISafety #AIRisk #WarningShots #RecursiveSelfImprovement #AIAlignment #ArtificialIntelligence #AIRegulation #FutureOfAI
106