1. EachPod

Red Teaming and AI Safety: Navigating the Ethical Gray Areas

Author
Mark Smith [nz365guy]
Published
Tue 29 Apr 2025
Episode Link
None

Get featured on the show by leaving us a Voice Mail: https://bit.ly/MIPVM 

FULL SHOW NOTES
https://www.microsoftinnovationpodcast.com/681  
 
The team explores the ethical implications of teaching AI jailbreaking techniques and conducting red team testing on large language models, balancing educational value against potential misuse. They dive into personal experiments with bypassing AI safeguards, revealing both creative workarounds and robust protections in modern systems. 
 
TAKEAWAYS 
• Debate on whether demonstrating AI vulnerabilities is responsible education or potentially dangerous knowledge sharing 
• Psychological impact on security professionals who regularly simulate malicious behaviors to test AI safety 
• Real examples of attempts to "jailbreak" AI systems through fantasy storytelling and other creative prompts 
• Legal gray areas in AI security testing that require dedicated legal support for organizations 
• Personal experiences with testing AI guardrails on different models and their varying levels of protection 
• Future prediction that Microsoft's per-user licensing model may shift to consumption-based as AI agents replace human tasks 
• Growth observations about Microsoft's Business Applications division reaching approximately $8 billion 
• Discussion of how M365 Copilot is transforming productivity, particularly for analyzing sales calls and customer interactions 

Check out this episode for more deep dives into AI safety, security, and the future of technology in business.

Support the show

If you want to get in touch with me, you can message me here on Linkedin.

Thanks for listening 🚀 - Mark Smith

Share to: