
Insights from recent episode analysis
Audience Interest
Podcast Focus
Publishing Consistency
Platform Reach
Insights are generated by CastFox AI using publicly available data, episode content, and proprietary models.
Est. Listeners
Based on iTunes & Spotify (publisher stats).
- Per-Episode Audience
Est. listeners per new episode within ~30 days
1,001 - 10,000 - Monthly Reach
Unique listeners across all episodes (30 days)
5,001 - 25,000 - Active Followers
Loyal subscribers who consistently listen
501 - 5,000
Market Insights
Platform Distribution
Reach across major podcast platforms, updated hourly
Total Followers
—
Total Plays
—
Total Reviews
—
* Data sourced directly from platform APIs and aggregated hourly across all major podcast directories.
On the show
Recent episodes
How to get hired in AI security
Mar 22, 2026
Unknown duration
getting talks accepted into conferences! tips and tricks
Jan 25, 2026
Unknown duration
Do we need to secure model weights?
Jan 18, 2026
Unknown duration
Model Context Protocol and Agent 2 Agent 🤖🕵️
Jan 11, 2026
Unknown duration
Agentic AI Security | case studies by Microsoft, OWASP
Jan 4, 2026
Unknown duration
Social Links & Contact
Official channels & resources
Official Website
Login
RSS Feed
Login
| Date | Episode | Description | Length | ||||||
|---|---|---|---|---|---|---|---|---|---|
| 3/22/26 | How to get hired in AI security | If you’re trying to break into AI security, it can feel confusing — do you need to be a machine learning expert, a cybersecurity professional, or both? In this episode, we break down practical tips for getting hired in AI security, from the skills that actually matter to the types of projects and experience that can help you stand out. We discuss how to build relevant expertise in areas like adversarial machine learning, AI risk, and model security, as well as how to position yourself for roles in startups, research labs, and large tech companies. Whether you’re coming from a cybersecurity, data science, or general tech background, this episode will give you actionable advice on how to start building a career in one of the fastest-growing areas of technology. 🚀 | — | ||||||
| 1/25/26 | getting talks accepted into conferences! tips and tricks | Want to give a great conference talk (and not bore everyone to death)? In this episode, I share practical tips for giving a strong conference talk — from structuring your idea to actually delivering it on stage. #PublicSpeaking #Conferences #CFP #TechTalks #Cybersecurity #AI | — | ||||||
| 1/18/26 | Do we need to secure model weights? | In this episode, we dig into model weight security — what it means, why it’s emerging as a critical issue in AI security, and whether the framing in the recent RAND report on securing AI model weights actually helps defenders and policymakers.We discuss the RAND report Securing AI Model Weights: Preventing Theft and Misuse of Frontier Models — exploring its core findings, including how model weights (the learnable parameters that encode what a model “knows”) are becoming high-value targets and the kinds of attack vectors that threat actors might use to steal or misuse them.#ai #aisecurity #cybersecurity 👉 Read the full RAND report here:https://www.rand.org/pubs/research_reports/RRA2849-1.html | — | ||||||
| 1/11/26 | Model Context Protocol and Agent 2 Agent 🤖🕵️ | In this episode, we dig into Model Context Protocol (MCP) and agent-to-agent (A2A) communication — what they are, why they matter, and where the real risks start to emerge.We cover:- What MCP actually enables beyond “tool calling”- How A2A changes the threat model for AI systems- Where trust boundaries break down when agents talk to each other- Why existing security assumptions don’t hold in agentic systems- What practitioners should be thinking about now (before this ships everywhere)This one’s for anyone working on AI systems, security, or governance who wants to understand what’s coming before it becomes a headline incident.As always: curious to hear your takes — especially where you think the biggest risks (or overblown fears) really are. | — | ||||||
| 1/4/26 | Agentic AI Security | case studies by Microsoft, OWASP | As promised, I’m back with Tania for a deep dive into the wild world of agentic AI security — how modern AI agents break, misbehave, or get exploited, and what real case studies are teaching us. We’re unpacking insights from the Taxonomy of Failure Modes in Agentic AI Systems, the core paper behind today’s discussion, and exploring what these failures look like in practice.We also break down three great resources shaping the conversation right now:Microsoft’s Taxonomy of Failure Modes in Agentic AI Systems — a super clear breakdown of how agent failures emerge across planning, decision-making, and action loops: https://cdn-dynmedia-1.microsoft.com/is/content/microsoftcorp/microsoft/final/en-us/microsoft-brand/documents/Taxonomy-of-Failure-Mode-in-Agentic-AI-Systems-Whitepaper.pdfOWASP’s Agentic AI Threats & Mitigations — a practical, security-team-friendly guide to common attack paths and how to defend against them: https://genai.owasp.org/resource/agentic-ai-threats-and-mitigations/Unit 42’s Agentic AI Threats report — real-world examples of adversarial prompting, privilege escalation, and chain-of-trust issues showing up in deployed systems: https://unit42.paloaltonetworks.com/agentic-ai-threats/Join us as we translate the research, sift through what’s real vs. hype, and talk about what teams should be preparing for next 🚨🛡️. | — | ||||||
| 12/23/25 | a hacky christmas message | A quick end-of-year message to say thanks. Thanks for being part of the channel this year — whether you’ve been watching quietly, sharing, or arguing with me in the comments. I really appreciate it.I hope you have a good Christmas and holiday period, whatever that looks like for you. Take a break if you can. See you in 2026. | — | ||||||
| 12/21/25 | Three Black Hat talks at just 18! My interview with Bandana Kaur. | In this episode, I’m joined by Bandana Kaur — a cybersecurity researcher, speaker, and all-round superstar who somehow managed to do in her teens what most people are still figuring out in their thirties. 🤔Bandana is just 18 years old, entirely self-taught in cybersecurity, already working in the field, and recently gave three talks at Black Hat. Yes, three! 😱We talk about how she taught herself cybersecurity as a teenager, how she broke into the industry without a traditional pathway, and what it’s actually like being young (and very competent) in a field that still struggles with gatekeeping. Bandana shares what she focused on while learning, how she approached opportunities like conference speaking, and what she thinks matters most for people trying to get into security today.This conversation is part career advice, and part reminder that you don’t need permission — or a perfectly linear path — to do meaningful work in cybersecurity.Follow Bandana: @hackwithher | — | ||||||
| 12/14/25 | Effective Altruism and AI with Good Ancestors CEO Greg Sadler | part 2 | Remember that time I invited myself over to Greg's place with my camera? This is part 2 from that great conversation. I'm curious to hear whether you've heard a lot about EA? It's something really big in the AI world but I'm conscious a lot of people outside the bubble haven't heard of it. Let me know in the comments! Check out Greg's work here: https://www.goodancestors.org.au/MIT AI Risk Repository: https://airisk.mit.edu/The Life You Can Save (book): https://www.thelifeyoucansave.org/book/80,000 hours: https://80000hours.org/Learn more about AI capability and impacts: https://bluedot.org/ | — | ||||||
| 12/7/25 | AI Safety with CEO of Good Ancestors Greg Sadler | part 1 | This week I invited myself over to Greg Sadler's place, the CEO of Good Ancestors, about AI safety. I brought sushi but I didn't have lunch so I ate most of it, and then I almost made him late for his next meeting. We specifically chat through AI capabilities, his work in policy, and building a not-profit. Greg is the kind of person who is so smart and cool that I feel like an absolute dummy interviewing him - so I know you're all going to like this episode. Stay tuned for part 2 where we dive into effective altruism and its intersection with AI!Check out Greg's work here: https://www.goodancestors.org.au/MIT AI Risk Repository: https://airisk.mit.edu/The Life You Can Save (book): https://www.thelifeyoucansave.org/book/80,000 hours: https://80000hours.org/Learn more about AI capability and impacts: https://bluedot.org/ | — | ||||||
| 11/24/25 | The United States AI Action Plan | will they win the AI race against China? 🤔 | Hi! 👋 In this episode, we’re diving into the US AI Action Plan — the White House’s new roadmap for how America plans to lead in AI.. and beat China.We’ll look at what’s inside the plan, what it really means for AI security and regulation, and whether it’s more of a policy promise… or a political one.📄 You can read the full plan here:https://www.whitehouse.gov/wp-content/uploads/2025/07/Americas-AI-Action-Plan.pdfLet me know what you think — is this the kind of leadership AI needs, or a dangerous framing of AI capability? | — | ||||||
Want analysis for the episodes below?Free for Pro Submit a request, we'll have your selected episodes analyzed within an hour. Free, at no cost to you, for Pro users. | |||||||||
| 11/9/25 | AI Security vs Application Security | Welcome back! 👋After taking a little break to reset and redesign everything behind the scenes, I’m back — and consolidating all my content. This episode is part of both The AI Security Podcast (on Spotify and Apple Podcasts) and my YouTube channel, HarrietHacks — so whether you prefer to listen or watch, you’ll get the same great conversations (and bad jokes) across both platforms.From now on, I’ll be posting at least fortnightly (with the occasional bonus episode when something big happens… like when I announced the book!).I’ve been in a few conversations lately where people have tried to convince me that AI Security is just Application Security in disguise. Naturally, I disagree. 🤷♀️ So in this episode, we dive into AI Security vs Application Security — how they overlap, where they diverge, and why securing AI systems demands new thinking beyond traditional AppSec.💌 Sign up for the newsletter: http://eepurl.com/i7RgRM📘 Pre-order The AI Security Handbook: [link coming soon]🎥 Watch this episode and more on YouTube: https://www.youtube.com/@HarrietHacks🔗 Useful LinksSQL Injection Examples (W3Schools): https://www.w3schools.com/sql/sql_injection.aspApplication Security Blog (Medium): https://medium.com/@pixelprecisionengineering1/application-security-appsec-in-cybersecurity-855ad9ce5e5eEcholeak Zero-Click Copilot Exploit (Dark Reading): https://www.darkreading.com/application-security/researchers-detail-zero-click-copilot-exploit-echoleakTraditional AppSec vs AI Security (Pillar Security): https://www.pillar.security/blog/traditional-appsec-vs-ai-security-addressing-modern-risks | — | ||||||
| 8/12/25 | Agentic AI Security: A Primer | For a while we've been wanting to talk about Agentic AI Security.. the thing is that we could spend multiple episodes talking about it! So we decided to do just that. This is part 1 - a primer - where we talk about exactly what AI agents are and why we may need to consider their security a bit differently. Stay tuned for the rest of the series! | — | ||||||
| 8/4/25 | How Likely Are AI Security Incidents? Updates From Our Final Report! | Six months ago Tania and I made an episode about the interim report for our AI Security Likelihood Project.. and it is finally time to discuss the final report! You'll see it live at this link shortly: https://www.aisecurityfundamentals.com/The premise was simple: are AI security incidents happening in the wild? What can we learn about future incidents from these historic ones? We answer some of these questions. | — | ||||||
| 7/23/25 | To open or close model weights? | In this episode, Tania and I discuss the debate around closed or open model weights. What do you think?The RAND report we mention: https://www.rand.org/pubs/research_reports/RRA2849-1.html | — | ||||||
| 7/15/25 | Creative prompt injection in the wild | In this episode, Tania and I talk through some creative examples of prompt injection/engineering we've seen in the wild.. think prompts hidden in papers, red-teaming and web-scraping.Your Brain on ChatGPT: https://arxiv.org/pdf/2506.08872Paper with hidden text (p. 12): https://arxiv.org/abs/2502.19918v2Interesting overview: https://www.theregister.com/2025/07/07/scholars_try_to_fool_llm_reviewers/Echoleak blog post: https://www.aim.security/lp/aim-labs-echoleak-m365 | — | ||||||
| 6/24/25 | Threat intel digest: 23 June 2025 | This week we discussed multiple AI vulnerabilities, including Echolink in M365 Copilot, Agent Smith in Langchain, and a SQL injection flaw in Llama Index, all of which have been patched. We also covered a data exposure bug in Asana's MCP server and OWASP's project to create an AI vulnerability scoring system, while also outlining Google's defense layers for Gemini, Thomas Roccia's Proximity tool for MCP server security, news regarding AI and legal/security concerns, and research on AI hacking AI, prompt compression, multi-agent security protocols, and the security of reasoning models versus LLMs. | — | ||||||
| 6/16/25 | AI safety evaluations with Inspect | I'm back from holiday, and this week Tania and I talk about a project she completed as part of the ARENA AI safety curriculum to replicate the findings of evaluations on frontier AI capabilities.Link to reasoning paper: https://arxiv.org/abs/2502.09696Link to the Inspect dashboard: https://inspect-evals-dashboard.streamlit.app/ARENA AI Safety course: https://www.arena.education/ | — | ||||||
| 6/10/25 | Threat intel digest: 9 June 2025 | This week we try a new condensed format for the AI security digest! we covered critical CVEs, including vulnerabilities in AWS MCP, Llama Index, GitHub MCP integration, and tool poisoning attacks. We also reported on malware campaigns using spoofed AI installers, a supply chain attack via fake PyTorch models, and the AI-guided discovery of a Linux kernel vulnerability by Sean Healin using OpenAI's 03 model. We addressed OpenAI's actions against malicious use of their models, Reddit's lawsuit against Anthropic for data scraping, the creation of an AI model for reconstructing 3D faces from DNA by Chinese researchers, a zero-trust framework for AI agent identity management proposed by the Cloud Security Alliance, research on an agent-based red teaming framework, the impact of context length on LLM vulnerability, and CSIRO's technique for improving deep fake detection. We also highlighted the vulnerablemcp.info project and the ongoing evolution of AI security best practices.Sign up to get the digest in your inbox: http://eepurl.com/i7RgRM | — | ||||||
| 5/30/25 | Threat intel digest: 26 May 2025 | Sign up to receive in your inbox: http://eepurl.com/i7RgRMTania Sadhani and Miranda R discussed various AI security topics, including critical CVEs affecting platforms like ChatGPT and Hugging Face, the potential for SharePoint Copilot in internal reconnaissance, and malicious npm packages targeting Cursor developers. They also covered the OASP Gen AI security initiative's Agent Name Service (ANS), the proposed AI.txt for controlling AI agent interactions, and Unit 42's framework for agentic AI attacks. Furthermore, Miranda highlighted security guidance from international agencies, Anthropic triggering ASL 3 for Claude Opus 4, Microsoft's AI red teaming playground, a significant data leak from an AI vendor, and the Israeli police's use of AI-hallucinated laws. | — | ||||||
| 5/20/25 | AI Vulnerability Research with Aditya Rana | Ever wondered how security vulnerabilities are found in AI? Join us as we chat with Aditya, a Vulnerability Researcher at Mileva Security Labs! | — | ||||||
| 5/12/25 | Threat intel digest: 12 May 2025 | Sign up to receive in your inbox: http://eepurl.com/i7RgRMThis week we note regular CVEs in AI libraries such as Nvidia TensorFlow and PyTorch. We discuss a novel prompt injection technique called "policy puppetry", along with malware dispersal through fake AI video generators and Meta's release of an open-source AI security tool set including Llama Firewall. We also covered Israel's experimental use of AI in warfare, Russia's AI-enabled drones in Ukraine, China's crackdown on AI misuse, Dreadnode's research on AI in red teaming, geolocation doxing via multimodal LLMs, safety research on autonomous vehicle attacks targeting inference time, Config Scan for analyzing malicious configurations on Hugging Face, Spotlight as a physical solution against deepfakes, and Reply Bench for benchmarking autonomous replication of LLM agents. | — | ||||||
| 5/7/25 | The evolution of data science and AI ethics with Dr Alberto Chierici | This week I'm joined by my friend Alberto, he has an incredible storied career - from data science, insurance, AI risk, advising Tesla.. check out his book here! https://www.amazon.com.au/Ethics-I-Facts-Fictions-Forecasts/dp/1636763650 | — | ||||||
| 4/30/25 | Stanford's 2025 AI Index Report | We talk about Stanford Human-Centred AI's latest AI Index report, check it out here: https://hai.stanford.edu/ai-index/2025-ai-index-report | — | ||||||
| 4/28/25 | Threat intel digest: 28 April 2025 | Did you know we have a fortnightly threat intel newsletter? We decided there was so much good research in there we have to talk about it here! We're joined by threat intel lead Miranda for this fortnight's biggest AI security news, coming out in this week's digest! http://eepurl.com/i7RgRM | — | ||||||
| 4/23/25 | Episode 28: the OWASP ML Security Top 10 with co-lead Shain Singh | My friend Shain joins me on the podcast to talk about his work with the OWASP MLSec Top 10 list and organisational guidance, as well as how he got here!For info about the list and how to contribute, check out the link: https://owasp.org/www-project-machine-learning-security-top-10/ | — | ||||||
Showing 25 of 52
Sponsor Intelligence
Sign in to see which brands sponsor this podcast, their ad offers, and promo codes.
Chart Positions
1 placement across 1 market.
Chart Positions
1 placement across 1 market.

























