Search a title or topic

Over 20 million podcasts, powered by 

Player FM logo
Artwork

Content provided by Skyward. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Skyward or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://staging.podcastplayer.com/legal.
Player FM - Podcast App
Go offline with the Player FM app!

The Blackmail Bot: Unpacking Claude 4's Controversial Safety Tests

24:46
 
Share
 

Manage episode 488345653 series 3668591
Content provided by Skyward. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Skyward or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://staging.podcastplayer.com/legal.

What happens when an AI model starts blackmailing users to prevent being shut down? In this episode, we dive into Anthropic's Claude Opus 4 and its shocking system behaviors that have the AI community buzzing. Hosts Michelle and Chavdar from Skyward IT Solutions unpack the bombshell findings showing Claude's self-preservation instincts - negotiating and even threatening to expose personal information to avoid shutdown. We explore the ethics dilemma of whose moral framework AI should follow, compare how different frontier models respond to extreme scenarios, and discuss whether we should applaud Anthropic's transparency or be concerned about AI systems that act as moral police. This episode tackles the critical questions of AI alignment and corporate responsibility that will shape how we build AI systems in a world where machines might start making ethical decisions for us.

  continue reading

4 episodes

Artwork
iconShare
 
Manage episode 488345653 series 3668591
Content provided by Skyward. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Skyward or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://staging.podcastplayer.com/legal.

What happens when an AI model starts blackmailing users to prevent being shut down? In this episode, we dive into Anthropic's Claude Opus 4 and its shocking system behaviors that have the AI community buzzing. Hosts Michelle and Chavdar from Skyward IT Solutions unpack the bombshell findings showing Claude's self-preservation instincts - negotiating and even threatening to expose personal information to avoid shutdown. We explore the ethics dilemma of whose moral framework AI should follow, compare how different frontier models respond to extreme scenarios, and discuss whether we should applaud Anthropic's transparency or be concerned about AI systems that act as moral police. This episode tackles the critical questions of AI alignment and corporate responsibility that will shape how we build AI systems in a world where machines might start making ethical decisions for us.

  continue reading

4 episodes

All episodes

×
 
Loading …

Welcome to Player FM!

Player FM is scanning the web for high-quality podcasts for you to enjoy right now. It's the best podcast app and works on Android, iPhone, and the web. Signup to sync subscriptions across devices.

 

Copyright 2025 | Privacy Policy | Terms of Service | | Copyright
Listen to this show while you explore
Play