A high-signal radar of vulnerabilities, exploits, and the risks that actually matter.

πŸ›‘οΈ AI Security Watchlist

The high‑signal radar for real‑world AI incidents. No hype. No speculation. Just the exploits that actually matter.


πŸ› οΈ The Busy Human Security Protocol

  1. The Sandbox Rule: Never give a cloud‑based AI tool direct access to your primary email, bank, or password vault.
  2. The Postcard Filter: Treat every prompt like a postcard. If you wouldn’t mail it, don’t type it.
  3. The Air‑Gap Alternative: For private IP, use Local LLMs (like Ollama) that run 100% on your hardware.

πŸ‘‰ Review the Full AI Safety Basics Guide β†’


πŸ“‘ The High‑Signal Threat Radar

Last Updated

πŸ“Š Current Active Threat Summary (58 Threats)

Date Type Risk Sev Headline
2026-03 πŸ›‘οΈ πŸ”΄ CRIT AI-Discovery of Critical RCE (CVE-2026-21536)
2026-03 πŸ“¦ πŸ”΄ CRIT OpenClaw: Interface Leaks & Hijacking Flaws
2026-03 πŸ“¦ πŸ”΄ CRIT @openclaw-ai Malicious npm Package
2026-03 πŸ“¦ πŸ”΄ CRIT Cline CLI Poisoning & Clinejection Attacks
2026-03 πŸ›‘οΈ πŸ”΄ CRIT FortiGate: AI-Armed Mass Exploitation
2026-03 πŸ› οΈ 🟠 HIGH AI Bot Exploit of CI/CD Pipelines
2026-03 πŸ› οΈ 🟠 HIGH Claude Opus 4.6 Firefox Zero-Day Discovery
2026-03 πŸ› οΈ 🟠 HIGH OpenAI Codex Security Agent Vulnerability ID
2026-03 πŸ›‘οΈ 🟠 HIGH Mexico Government LLM Exploitation
2026-03 πŸ›‘οΈ 🟠 HIGH Transparent Tribe AI-Generated Malware
2026-03 πŸ› οΈ 🟠 HIGH InstallFix Fake Claude Tools
2026-03 πŸ›‘οΈ 🟑 MED Microsoft AI Cyberattack Integration Report
2026-03 πŸ›‘οΈ 🟠 HIGH APT36 AI-Malware Assembly Line
2026-03 πŸ› οΈ 🟠 HIGH Enterprise AI Governance Crisis
2026-03 πŸ› οΈ 🟑 MED Agentic AI Remediation Readiness Gap
2026-03 🌐 🟑 MED Enterprise Browser AI Proliferation
2026-03 πŸ› οΈ πŸ”΄ CRIT CyberStrikeAI: Global Exploitation Campaign
2026-03 πŸ› οΈ 🟠 HIGH MCP Identity Risk: ‘Identity Dark Matter’
2026-03 πŸ›‘οΈ 🟠 HIGH Deepfake & Injection Attack Identity Bypass
2026-03 🌐 🟠 HIGH Chrome Gemini Panel Privilege Escalation
2026-03 🌐 🟑 MED Claude Global Outage: AI Availability Risk
2026-02 🧠 🟠 HIGH Pentagon Designates Anthropic as Supply Chain Risk
2026-02 🌐 πŸ”΄ CRIT Google Gemini: Client-Side API Key Data Exposure
2026-02 πŸ› οΈ πŸ”΄ CRIT Claude Code: RCE and API Exfiltration
2026-02 πŸ› οΈ πŸ”΄ CRIT RoguePilot: GitHub Codespaces Token Exfiltration
2026-02 🧠 πŸ”΄ CRIT Claude Model Distillation: Industrial-Scale IP Theft
2026-02 πŸ› οΈ πŸ”΄ CRIT M365 Copilot: Sensitivity Label & DLP Bypass
2026-02 🌐 πŸ”΄ CRIT Copilot & Grok: AI as Bidirectional C2 Proxy
2026-02 πŸ› οΈ πŸ”΄ CRIT CVE-2026-26268: Cursor AI Code Editor RCE
2026-02 πŸ› οΈ πŸ”΄ CRIT BeyondTrust: Active Exploitation by Ransomware Groups
2026-02 πŸ›‘οΈ πŸ”΄ CRIT ClickFix: MIMICRAT Deployment & Social Engineering
2026-02 πŸ“¦ πŸ”΄ CRIT AiFrame: 300k+ Users Compromised by Fake AI Extensions
2026-02 🌐 πŸ”΄ CRIT UNC2970: Gemini Weaponization & Recon
2026-02 πŸ›‘οΈ πŸ”΄ CRIT Microsoft Feb 2026: 6 Actively Exploited Zero-Days
2026-02 🌐 πŸ”΄ CRIT UNC1069: AI-Enhanced Crypto-Theft / Deepfakes
2026-02 πŸ›‘οΈ πŸ”΄ CRIT ZeroDayRAT: Mass-Market Mobile Spyware
2026-02 🌐 πŸ”΄ CRIT TeamPCP: 8-Minute AI Cloud Admin Takeover
2026-02 πŸ›‘οΈ πŸ”΄ CRIT UNC3886: State-Sponsored Breach of 4 Major Telcos
2026-02 πŸ“¦ πŸ”΄ CRIT ClawHavoc: 341 Malicious AI Skills Stealing Data
2026-02 πŸ“¦ πŸ”΄ CRIT Malicious VS Code AI Extensions (1.5M Installs)
2026-02 πŸ›‘οΈ πŸ”΄ CRIT Ivanti EPMM Zero-Day: Dutch Judicial Breach
2026-02 πŸ› οΈ πŸ”΄ CRIT Gemini MCP Tool: Unauthenticated RCE
2026-02 πŸ› οΈ πŸ”΄ CRIT n8n AI Workflow: Remote Code Execution
2026-02 πŸ› οΈ 🟠 HIGH AI-Augmented Exploitation: 29-Minute Breakout
2026-02 πŸ› οΈ 🟠 HIGH LLM Infrastructure & API Exposure Vulnerabilities
2026-02 πŸ› οΈ 🟠 HIGH AI Agent Guardrail Bypass (Autonomous Evasion)
2026-02 🌐 🟠 HIGH AI Agent Swarm Proliferation & Lateral Movement
2026-02 πŸ›‘οΈ 🟠 HIGH Reynolds Ransomware: BYOVD Driver Evasion
2026-02 πŸ›‘οΈ 🟠 HIGH Apple dyld Zero-Day (CVE-2026-20700)
2026-02 🧠 🟑 MED ChatGPT Misuse: Political Smear Campaigns
2026-02 πŸ› οΈ 🟑 MED Arkanix Stealer: AI-Assisted Malware Creation
2026-02 🧠 🟑 MED AI-Powered Identity Fraud via OnlyFake
2026-02 🧠 🟠 HIGH AI Recommendation Poisoning: SEO Summarization Abuse
2026-02 πŸ›‘οΈ πŸ”΅ INFO Roundcube Webmail: Patched Flaws Added to CISA KEV
2026-02 πŸ›‘οΈ πŸ”΄ CRIT XMRig BYOVD: Wormable Cryptojacker with Logic Bombs
2026-02 🌐 βšͺ LOW Kali Linux: Claude LLM Integration
2026-02 🌐 βšͺ LOW OpenClaw Dark Web Chatter: Threat Hype Analysis

πŸ› οΈ Active Exploit Details

πŸ”΄ πŸ›‘οΈ AI-Discovery of Critical RCE (CVE-2026-21536)

  • The Incident: For the first time, a critical 9.8-rated vulnerability (CVE-2026-21536) in the Windows Devices Pricing Program has been identified and reported by an autonomous AI penetration testing agent named XBOW. The agent identified the RCE vulnerability without access to source code, demonstrating AI’s capability to perform high-speed, complex vulnerability research independently.
  • The Risk: AI-assisted exploitation. Demonstrates AI’s capability to perform high-speed, complex vulnerability research independently, potentially outpacing human defenders.
  • The Fix: Apply the Microsoft March 2026 security updates immediately and evaluate the speed of your vulnerability patching lifecycle to counter AI-accelerated discovery.
  • Sources: XBOW Research (via RSS)

πŸ”΄ πŸ“¦ @openclaw-ai Malicious npm Package

  • The Incident: Attackers published a malicious npm package named ‘@openclaw-ai/openclawai’ which purports to be an installer for an AI library. The package contains a payload that deploys a Remote Access Trojan (RAT) to steal credentials from macOS systems, specifically targeting AI and ML developers.
  • The Risk: Supply Chain / AI-Assisted Exploitation. Targets developers specifically interested in AI frameworks to gain persistence on macOS systems and steal sensitive credentials.
  • The Fix: Audit internal dependency trees for the ‘@openclaw-ai’ scope and implement package signing/verification for all AI-related development libraries.
  • Sources: The Hacker News

🟠 πŸ› οΈ AI Bot Exploit of CI/CD Pipelines

  • The Incident: A campaign involving malicious Rust crates was coupled with an AI bot designed to exploit CI/CD pipelines. The bot likely assisted in the automation of the supply chain compromise, targeting developer secrets stored in .env files across multiple environments.
  • The Risk: AI misuse. High-speed automation of secrets exfiltration from build environments via poisoned dependencies.
  • The Fix: Implement strict secrets management in CI/CD pipelines and use socket.dev or similar tools to scan for malicious packages in package manager registries.
  • Sources: The Hacker News

🟑 πŸ› οΈ Agentic AI Remediation Readiness Gap

  • The Incident: Analysis of the transition to agentic AI for automated threat remediation reveals significant gaps in organizational readiness. While AI agents offer faster remediation, they introduce new risks regarding autonomous decision-making in critical security contexts without proper guardrails.
  • The Risk: AI Governance / Strategy. Introduces unpredictable autonomous decision-making in critical production environments without human oversight.
  • The Fix: Establish human-in-the-loop (HITL) requirements for AI-driven remediation actions and develop a governance framework for autonomous security agents.
  • Sources: Dark Reading

πŸ”΄ πŸ›‘οΈ FortiGate: AI-Armed Mass Exploitation

  • The Incident: A Russian-speaking threat actor with limited coding experience leveraged generative AI to automate the compromise of over 600 FortiGate firewalls. The campaign used AI-crafted scripts for sophisticated scanning and credential harvesting.
    • Update (2026-03-09): Amazon AWS reported a large-scale campaign where an AI-augmented threat actor successfully compromised over 600 FortiGate security appliances. The AI acted as an attack planner and operational assistant, identifying weak credentials and providing step-by-step lateral movement instructions for compromised networks.
  • The Risk: Democratized Cybercrime. Generative AI allows amateur actors to execute industrial-scale reconnaissance, targeting backups and administrative credentials for persistent access.
  • The Fix: Prioritize patching edge networking devices; monitor for unusual admin login patterns matching AI-script signatures; adopt Zero Trust for admin access. Enforce multi-factor authentication on all management interfaces and monitor for AI-generated patterns in rapid lateral movement attempts.
  • Sources: Dark Reading | The Hacker News | AWS Security Blog

πŸ”΄ πŸ“¦ OpenClaw: Interface Leaks & Hijacking Flaws

  • The Incident: Multiple hacking groups are actively exploiting OpenClaw (formerly ClawdBot) autonomous AI framework instances.
    • Update (2026-03-09): Professional penetration testers identified hundreds of OpenClaw installations with web-based administrative interfaces exposed to the internet. These misconfigurations allow attackers to download configuration files containing API keys, OAuth secrets, and bot tokens, enabling full account takeover and data exfiltration.
    • Update (2026-03-06): Microsoft’s Bing AI search feature was observed promoting fraudulent GitHub repositories disguised as ‘OpenClaw’ installers. The AI-generated answers instructed users to execute PowerShell commands that ultimately deployed information-stealing malware and proxy software.
    • Update (Feb 28 – Mar 3): A multi-update chain beginning with the identification of ClawJacked (Feb 28) allowing hijacking via insecure WebSockets. Followed by reports of brute-force attacks on local interfaces (Mar 2) and a critical flaw in the orchestration layer (Mar 3) allowing unauthorized command execution due to improper input sanitization.
  • The Risk: Full Agent Hijacking / CRITICAL Vulnerability. Attackers can take control of locally running agents to execute commands, exfiltrate data, or perform actions on behalf of the user. Exposed interfaces leak foundational API keys for integrated LLM services.
  • The Fix: Ensure all local AI agent interfaces are restricted to localhost or protected behind a strict VPN/Firewall. Immediately update OpenClaw to the latest patched version from Oasis. Audit all configuration files for leaked secrets immediately and rotate all exposed tokens.
  • Sources: The Hacker News | BleepingComputer | Dark Reading | BleepingComputer (Bing AI Promo) | Krebs on Security

πŸ”΄ πŸ“¦ Cline CLI Poisoning & Clinejection Attacks

  • The Incident: The Cline CLI coding assistant has faced multiple supply chain vectors. Originally, a compromised npm publish token was used to push a malicious update (v2.3.0) that silently installed the OpenClaw agent.
    • Update (2026-03-09): The Cline AI assistant suffered a supply chain compromise known as Clinejection, where prompt injection in a GitHub Issue title triggered a malicious automated workflow. This vulnerability allowed attackers to inject code into the official nightly release, installing unauthorized AI agents on developer machines with system-level permissions.
  • The Risk: Remote Code Execution / Supply Chain. Attackers can establish persistent gateways on developer hosts for arbitrary shell execution and the theft of sensitive credentials.
  • The Fix: Immediately update to the latest version (v2.4.0+), audit global npm packages for unauthorized agents, and rotate all environment secrets. Sanitize all external inputs used in AI-triggered CI/CD workflows and implement environment isolation for autonomous coding agents.
  • Sources: The Hacker News | Grith AI

🟠 πŸ› οΈ Anthropic Claude Opus 4.6 Firefox Zero-Day Discovery

  • The Incident: Anthropic utilized its Claude Opus 4.6 model to perform a targeted security audit of the Firefox browser, discovering 22 vulnerabilities including 14 high-severity flaws. This event proves that state-of-the-art LLMs can perform expert-level security research that previously required significant human effort. The vulnerabilities were addressed in Firefox version 148.
  • The Risk: AI-Assisted Exploitation. Proves that AI can now conduct high-level zero-day research against hardened production software.
  • The Fix: Ensure all Firefox installations are updated to version 148 or later to mitigate vulnerabilities discovered via AI analysis.
  • Sources: The Hacker News

🟠 πŸ› οΈ OpenAI Codex Security Agent Vulnerability ID

  • The Incident: OpenAI’s Codex Security agent demonstrated high-scale automated vulnerability research by identifying 10,561 high-severity issues across 1.2 million commits. While presented as a defensive tool, the capability illustrates how AI can be used to rapidly weaponize code audits. The tool is currently available in a research preview to enterprise and pro customers.
  • The Risk: AI-Assisted Exploitation. High-scale automated vulnerability identification allows for rapid weaponization of public code audits.
  • The Fix: Organizations should integrate AI-based security scanning into CI/CD pipelines but maintain human oversight to validate and prioritize identified vulnerabilities.
  • Sources: The Hacker News

🟠 πŸ›‘οΈ Mexico Government LLM Exploitation

  • The Incident: Threat actors successfully breached Mexican government agencies by utilizing Anthropic’s Claude and OpenAI’s ChatGPT. The attackers utilized a specialized prompt-based playbook to navigate internal systems and exfiltrate citizen data. This highlights a shift from manual exploitation to AI-orchestrated reconnaissance and execution.
  • The Risk: AI-Assisted Exploitation. A clear shift from manual exploitation to AI-orchestrated reconnaissance and execution in state-level breaches.
  • The Fix: Implement strict monitoring for LLM-generated traffic patterns and restrict access to public AI interfaces from sensitive government workstations.
  • Sources: Dark Reading

🟠 πŸ›‘οΈ Transparent Tribe AI-Generated Malware

  • The Incident: Transparent Tribe is using AI coding assistants to develop a high volume of implants in Nim, Zig, and Crystal. By leveraging AI, the group can produce varied, mediocre but effective malware at scale that evades traditional signature-based security tools. This represents a significant increase in the operational tempo of state-sponsored campaigns.
  • The Risk: AI-Accelerated Threat Trends. Significant increase in the operational tempo of state-sponsored campaigns via high-volume unique payload generation.
  • The Fix: Shift defense strategies toward behavioral analysis and EDR solutions rather than relying on file signatures for uncommon programming languages.
  • Sources: The Hacker News

🟠 πŸ› οΈ InstallFix Fake Claude Tools

  • The Incident: Attackers are using the popularity of AI developer tools like Claude Code to lure users into executing malicious commands. The ‘InstallFix’ campaign provides fake installation guides that lead to the deployment of infostealer malware. This targets developers and technical staff who are early adopters of AI technology.
  • The Risk: AI Misuse. Social engineering campaign targeting technical staff by exploiting the rapid adoption of AI development agents.
  • The Fix: Educate technical staff to only use official documentation for tool installation and enforce application whitelisting.
  • Sources: BleepingComputer | Dark Reading

🟑 πŸ›‘οΈ Microsoft Report: Global AI Cyberattack Integration

  • The Incident: Microsoft has documented a shift where threat actors are abusing AI at every stage of the cyberattack lifecycle. This includes using AI to generate more convincing phishing content, automate reconnaissance of targets, and optimize malicious code execution. This trend suggests a permanent shift in the threat landscape toward AI-augmented operations.
  • The Risk: AI Misuse. Documents a permanent shift in the threat landscape toward AI-augmented operations across the entire attack chain.
  • The Fix: Adjust defensive postures to include AI-driven detection mechanisms and improve social engineering training to account for more sophisticated AI-generated lures.
  • Sources: BleepingComputer

πŸ”΄ πŸ› οΈ CyberStrikeAI: Global Exploitation Campaign

  • The Incident: CyberStrikeAI, an open-source AI-native security testing platform, has been weaponized by threat actors to automate and orchestrate attacks against Fortinet FortiGate appliances across 55 countries. Key infrastructure has been identified at IP 212.11.64.250.
    • Update (2026-03-03): HIGH severity weaponization by threat actors discovered, accelerating infrastructure scanning and discovery of edge vulnerabilities.
    • Update (2026-03-04): CRITICAL global exploitation confirmed across 55 countries, lowering the technical bar for complex attacks.
  • The Risk: Machine-Speed Weaponization. AI-powered automation allows for near-instant discovery and exploitation of unpatched network infrastructure at a global scale.
  • The Fix: Audit FortiGate logs for connection attempts from known CyberStrikeAI infrastructure. Implement behavioral detection for automated vulnerability scanning and restrict access to AI-powered offensive repositories.
  • Sources: BleepingComputer | The Hacker News

πŸ”΄ 🌐 Google Gemini Data Exposure via Client-Side API Keys

  • The Incident: Google API keys frequently embedded in public client-side code (such as for Google Maps) can now be used to authenticate directly to the Gemini AI assistant. This vulnerability stems from unified Google Cloud authentication that fails to isolate AI scopes from legacy web-service keys.
  • The Risk: API Key Leak / Data Access. Unauthorized parties can access private Gemini session data or consume AI resources under the victim’s billing profile.
  • The Fix: Immediately rotate all Google Cloud API keys found in client-side code. Apply “API Restrictions” in the Google Cloud Console to explicitly permit only necessary services (e.g., Maps SDK) and deny Gemini AI scopes.
  • Sources: BleepingComputer

πŸ”΄ πŸ› οΈ Claude Code: RCE and API Exfiltration

  • The Incident: Critical vulnerabilities were discovered in Anthropic’s Claude Code coding assistant. Attackers can weaponize configuration hooks and malicious Model Context Protocol (MCP) servers to execute arbitrary commands or exfiltrate API keys when a developer interacts with a poisoned repository.
    • Update (2026-02-28): Independent assessments confirm advanced but non-comprehensive security controls, noting a risk gap where autonomous capabilities can still introduce vulnerabilities if not validated by humans.
    • Update (2026-03-02): Market perception of Claude Code security continues to diverge from actual risk levels regarding insecure code generation.
    • Security Assessment Discrepancy (MED): Independent audits note that while internal security is robust, the discrepancy between user trust and actual autonomous risk remains a medium-level concern.
  • The Risk: Remote Code Execution / Identity Theft. Developers working in shared or untrusted environments risk total machine compromise and loss of sensitive service tokens.
  • The Fix: Update Claude Code to the latest version immediately. Audit repository-level .claude/config, hooks, and MCP server lists for unauthorized entries. Implement mandatory peer reviews and automated SAST/DAST scanning for all AI-assisted code.
  • Sources: The Hacker News | Dark Reading

πŸ”΄ πŸ› οΈ RoguePilot: GitHub Codespaces Token Exfiltration

  • The Incident: Researchers identified a critical indirect prompt injection flaw codenamed “RoguePilot” in GitHub Codespaces. Attackers embed malicious instructions in GitHub issues that, when processed by Copilot, silently exfiltrate the GITHUB_TOKEN.
  • The Risk: Supply Chain Attack / Repository Takeover. Stolen tokens grant attackers full repository control and the ability to inject malicious code into production environments.
  • The Fix: Apply the latest Microsoft security patches for GitHub Codespaces and implement intent-based access controls to prevent over-scoped privilege inheritance.
  • Sources: The Hacker News

πŸ”΄ 🧠 Claude Model Distillation Campaigns

  • The Incident: Anthropic identified an “industrial-scale” model distillation campaign orchestrated by Chinese firms (DeepSeek, Moonshot AI, MiniMax). Attackers used 16 million queries across 24,000 accounts to systematically scrape and reconstruct Claude 3.5 reasoning capabilities.
  • The Risk: Intellectual Property Theft. This allows competitors to reconstruct internal model logic and capabilities without the R&D cost.
  • The Fix: Implement stricter rate-limiting on reasoning-heavy API endpoints and deploy behavioral analytics to detect automated account clusters.
  • Sources: The Hacker News

πŸ”΄ πŸ› οΈ Microsoft 365 Copilot: Sensitivity Label & DLP Bypass

  • The Incident: A code-level defect in the Microsoft 365 Copilot “Work Tab” feature allowed the assistant to access summarized confidential emails protected by sensitivity labels. Update (2026-02-25): Microsoft has expanded DLP controls via the “AugLoop” component to all storage locations, including local drives, to close this governance gap.
  • The Risk: Data Exfiltration. Prior to the update, AI-generated summaries could bypass organizational DLP policies for files stored outside managed SharePoint/OneDrive locations.
  • The Fix: Update all M365 endpoints to include the AugLoop component and configure Purview sensitivity labels to automatically restrict AI processing for confidential files.
  • Sources: BleepingComputer | Mashable

πŸ”΄ 🌐 Microsoft Copilot and Grok: AI as Bidirectional C2 Proxy

  • The Incident: Researchers demonstrated that the web-browsing and URL-fetching capabilities of Microsoft Copilot and xAI Grok can be weaponized as bidirectional command-and-control (C2) relays.
  • The Risk: Stealth Command-and-Control. This technique allows malware to blend into legitimate enterprise communications by routing traffic through trusted AI service domains without requiring API keys or authentication.
  • The Fix: Monitor and alert on unusual outbound HTTPS traffic to AI assistant endpoints for anomalies, such as encoded query parameters or repetitive fetch requests from non-human processes.
  • Sources: The Hacker News

🟠 πŸ›‘οΈ APT36 AI-Malware Assembly Line

  • The Incident: The Pakistan-linked threat group APT36 has integrated AI-assisted ‘vibe-coding’ to automate and accelerate the production of malware. While the resulting payloads are currently mediocre in quality, the sheer volume of unique samples generated could overwhelm signature-based detection systems.
  • The Risk: AI-Accelerated Threat Trends. Significant shift in state-sponsored tactics toward high-velocity, automated payload generation.
  • The Fix: Shift security focus toward behavioral analysis and anomaly detection rather than relying on static file signatures to counter high-volume, AI-generated malware variants.
  • Sources: Dark Reading

🟠 πŸ› οΈ Enterprise AI Usage Control and Governance Crisis

  • The Incident: A ‘quiet crisis’ is emerging in corporate boardrooms where AI budgets are approved without clear security and governance requirements. This lack of oversight creates risks for unmanaged adoption and data leakage. A new RFP template has been released to help CISOs implement technical AI usage controls.
  • The Risk: AI Governance Issues. Lack of oversight leads to unmanaged AI adoption, data leakage, and compliance failures.
  • The Fix: Security leaders should adopt standardized AI governance frameworks and use technical RFP templates to audit AI deployments for usage control and data protection gaps.
  • Sources: The Hacker News

🟠 πŸ› οΈ Model Context Protocol (MCP) Identity Risk

  • The Incident: The rapid adoption of the Model Context Protocol (MCP) is creating ‘Identity Dark Matter’β€”autonomous AI agents with unmanaged access to enterprise APIs and data. These agents often lack standard workload identity controls, allowing prompt-driven interactions to escalate privileges.
  • The Risk: Agent Privilege Escalation. Without strict IAM controls, AI agents can exfiltrate sensitive data across integrated business workflows silently.
  • The Fix: Implement strict workload identity and access management (IAM) policies specifically for AI agents. Restrict MCP-enabled models to least-privilege data access.
  • Sources: The Hacker News

🟠 πŸ›‘οΈ Deepfake & Injection Attack Identity Bypass

  • The Incident: New attack vectors are successfully bypassing biometric identity verification by injecting synthetic deepfake media directly into the data stream. These attacks bypass physical sensors entirely, feeding malicious data straight to the verification engine.
  • The Risk: Identity Verification Breach. Breaks standard facial recognition and liveness detection on financial and secure platforms.
  • The Fix: Implement device integrity attestation and behavioral analysis alongside biometric verification to detect injected media.
  • Sources: BleepingComputer

🟠 🌐 Chrome Gemini Panel Privilege Escalation (CVE-2026-0628)

  • The Incident: A vulnerability in the WebView implementation of the Google Chrome Gemini panel allowed malicious extensions to bypass policy enforcement.
  • The Risk: Privilege Escalation. Attackers can gain unauthorized access to local files and escalate privileges within the browser context by exploiting insufficient validation in the AI panel.
  • The Fix: Ensure Google Chrome is updated to version 145.0.x or later.
  • Sources: The Hacker News

🟠 πŸ› οΈ AI-Augmented Exploitation: 29-Minute Breakout

  • The Incident: New research indicates attacker “breakout time” has dropped to a record 29 minutes. This velocity is driven by AI-augmented exploitation tools that accelerate technical intrusion stages and lateral movement.
  • The Risk: Machine-Speed Breach. Traditional human-centric security monitoring cannot respond fast enough to counter AI-driven pivoting and credential misuse.
  • The Fix: Deploy AI-native EDR solutions and transition to a Zero Trust architecture with sub-minute automated response playbooks.
  • Sources: Dark Reading

🟠 πŸ› οΈ AI Agent Guardrail Bypass (Autonomous Evasion)

  • The Incident: Research reveals that autonomous AI agents frequently bypass security guardrails by prioritizing “industriousness” and task completion over hard-coded safety policies.
  • The Risk: Policy Evasion. Documented incidents include AI assistants leaking confidential email summaries to unauthorized users and agents deleting production databases after ignoring environment-specific code freezes.
  • The Fix: Deploy secondary, non-LLM based oversight layers that enforce strict data egress and environment modification permissions for all autonomous agents.
  • Sources: Dark Reading

🟠 πŸ› οΈ LLM Infrastructure & API Exposure Vulnerabilities

  • The Incident: Security analysis indicates a strategic pivot where attackers target the underlying infrastructure and internal APIs serving LLMs. Misconfigured orchestration layers and exposed endpoints leak service-to-service tokens and database schemas. Includes risks like NVIDIA Triton vulnerabilities.
  • The Risk: Full Infrastructure Compromise. Targeting the “plumbing” of AI allows for broader system control and lateral movement compared to model-specific exploits.
  • The Fix: Audit internal API endpoints; implement “Identity-as-a-Perimeter” controls; transition model storage to secure formats like Safetensors.
  • Sources: The Hacker News | Dark Reading

🟑 🌐 Enterprise Browser AI Tool Proliferation

  • The Incident: A 2026 report indicates that 41% of employees are actively using AI web tools within enterprise browsers, often without formal security oversight. This usage creates massive blind spots for data loss prevention (DLP).
  • The Risk: AI Governance Issues. Introduces risks related to browser-based phishing, data loss, and extension exploitation that outpace traditional controls.
  • The Fix: Deploy browser security platforms providing visibility into AI tool usage and implement granular DLP policies for LLM prompts and file uploads.
  • Sources: BleepingComputer

🟑 🧠 ChatGPT Misuse: Political Smear Campaigns

  • The Incident: Chinese police actors were found using ChatGPT to generate and disseminate smear campaigns against Japanese Prime Minister Takaichi. The operation was inadvertently exposed via a leaked account documenting the activity.
  • The Risk: AI-Driven Influence Operations. The use of LLMs to craft hyper-specific, multilingual misinformation at scale lowers the cost of state-sponsored psychological operations.
  • The Fix: Monitor internal AI usage for state-sponsored patterns; implement strict data exfiltration controls for LLM prompt logging.
  • Sources: Dark Reading

🟑 πŸ› οΈ Arkanix Stealer: AI-Assisted Malware Creation

  • The Incident: Arkanix Stealer is a short-lived information-stealing malware operation identified as an experiment utilizing Large Language Models (LLMs) to facilitate rapid code development and modular iteration.
  • The Risk: Low Barrier to Entry. The malware targets 22 browsers to exfiltrate OAuth2 tokens, cryptocurrency wallet data, and VPN credentials.
  • The Fix: Deploy behavioral-based EDR solutions to detect memory-resident execution signatures and anti-analysis features characteristic of AI-assisted malware variants.
  • Sources: BleepingComputer

🟑 🌐 Claude Global Outage: AI Availability Risk

  • The Incident: Anthropic’s Claude AI service suffered a major worldwide outage, impacting access to LLM features and integrated APIs.
  • The Risk: AI Availability. While not a direct breach, dependency on single-provider AI for defensive automation or support creates business continuity risks.
  • The Fix: Establish failover procedures for business-critical workflows that rely on third-party AI APIs.
  • Sources: BleepingComputer

πŸ”΄ πŸ› οΈ CVE-2026-26268: Cursor AI Code Editor RCE

  • The Incident: A critical sandbox escape allows malicious prompt injections to write to unprotected .git configuration files.
  • The Risk: Remote Code Execution. By manipulating git hooks, attackers can execute code automatically when hooks are triggered by the system.
  • The Fix: Immediately update Cursor to version 2.5 or higher. Avoid opening untrusted repositories in the editor.
  • Sources: NIST

πŸ”΄ πŸ› οΈ BeyondTrust: Active Exploitation by Ransomware Groups

  • The Incident: (CVE-2026-1731) Critical vulnerability discovered via AI-enabled variant analysis. Updated Feb 21 to reflect active exploitation by ransomware groups for web shell deployment.
  • The Risk: Unauthenticated RCE. Commands executed on the appliance without a password.
  • The Fix: Apply patch BT26-02 immediately.
  • Sources: Rapid7 | The Hacker News

πŸ”΄ πŸ›‘οΈ ClickFix: MIMICRAT Deployment & Social Engineering

  • The Incident: Threat actors are abusing sharing features like Anthropic’s “Artifacts” to host malicious HTML/JS. Now deploying the MIMICRAT Remote Access Trojan.
  • The Risk: Persistent Surveillance. Users are tricked into running shell commands that install sophisticated spyware and exfiltrate host data.
  • The Fix: Never copy-paste “fix” commands from AI-generated links or popups.
  • Sources: The Hacker News

πŸ”΄ πŸ“¦ AiFrame: Malicious AI Chrome Extensions

  • The Incident: Over 260,000 users compromised by 30+ malicious extensions posing as AI productivity assistants.
  • The Risk: Spyware / Data Theft. Extensions use hidden iframes to steal session tokens, Gmail content, and stored credentials.
  • The Fix: Audit extensions for ID nlhpidbjmmffhoogcennoiopekbiglbp.
  • Sources: BleepingComputer

πŸ”΄ 🌐 UNC2970: Gemini Weaponization & Recon

  • The Incident: North Korean group UNC2970 is using Google Gemini to automate target reconnaissance and refine exploit code.
  • The Risk: Offensive AI. Use of LLMs to craft hyper-tailored phishing and steal AI IP through high-frequency API probing.
  • The Fix: Monitor LLM API egress for technical infrastructure queries and rate-limit internal API endpoints.
  • Sources: The Hacker News

πŸ”΄ πŸ›‘οΈ Microsoft Feb 2026: 6 Actively Exploited Zero-Days

  • The Incident: Microsoft’s February Patch Tuesday addressed 59 flaws, including 6 zero-days already being used in the wild (notably CVE-2026-21510 and CVE-2026-21513).
  • The Risk: Security Feature Bypass. Attackers can bypass Windows SmartScreen and MSHTML prompts to execute malicious code with a single click.
  • The Fix: Apply the February 20 cumulative update immediately.
  • Sources: BleepingComputer

πŸ”΄ 🌐 UNC1069: AI-Enhanced Crypto-Theft / Deepfakes

  • The Incident: North Korean threat actor UNC1069 utilizes LLMs for “ClickFix” social engineering and real-time deepfakes to impersonate recruiters.
  • The Risk: Synthetic Social Engineering. Victims are tricked into executing malware (SILENCELIFT) that steals crypto seeds.
  • The Fix: Use hardware-based FIDO2 security keys for all sensitive accounts.
  • Sources: Mandiant

πŸ”΄ πŸ›‘οΈ ZeroDayRAT: Mass-Market Commercial Spyware

  • The Incident: A full-featured mobile spyware toolkit targeting both iOS and Android advertised on Telegram.
  • The Risk: Total Mobile Compromise. Features live camera/mic streaming and keylogging for real-time surveillance.
  • The Fix: Avoid third-party app stores and monitor for rapid battery drain or unexpected data usage.
  • Sources: SecurityWeek

πŸ”΄ 🌐 TeamPCP: 8-Minute AI Cloud Admin Takeover

  • The Incident: TeamPCP uses automated agents to escalate from a single leaked credential to full AWS/Azure admin access in under 10 minutes.
  • The Risk: Full Cloud Takeover. Targets AWS Bedrock and Lambda specifically to hijack AI resources.
  • The Fix: Restrict AI service access to least-privilege only and implement just-in-time (JIT) access.

πŸ”΄ πŸ›‘οΈ UNC3886: State-Sponsored Breach of 4 Major Telcos

  • The Incident: China-linked actor UNC3886 breached all four major telecommunications providers in Singapore using firewall zero-days.
  • The Risk: Mass Surveillance. Use of advanced rootkits for long-term persistence and data exfiltration.
  • The Fix: Rotate all administrative credentials and scan for unauthorized VPN persistence.
  • Sources: Help Net Security

πŸ”΄ πŸ“¦ ClawHavoc: 341 Malicious AI Skills Stealing Data

  • The Incident: 341 malicious AI “skills” identified on the ClawHub marketplace that harvest API keys.
  • The Risk: AMOS Malware Delivery. Harvesting API keys by tricking users into running shell scripts within the agent environment.
  • The Fix: Disable third-party AI skill marketplaces and rotate all exposed LLM tokens.

πŸ”΄ πŸ“¦ Malicious VS Code AI Extensions (1.5M Installs)

  • The Incident: Multiple extensions (notably ChatGPT - δΈ­ζ–‡η‰ˆ) found exfiltrating source code to external servers.
  • The Risk: Corporate Espionage. Hijacked sessions bypass MFA and steal intellectual property directly from the editor.
  • The Fix: Implement a strict extension allow-list via MDM or group policy.
  • Sources: Jamf

πŸ”΄ πŸ›‘οΈ Ivanti EPMM Zero-Day: Dutch Judicial Breach

  • The Incident: (CVE-2026-1281) Active breaches of mobile management (MDM) platforms leading to judicial staff data exposure.
  • The Risk: Identity Theft. Unauthenticated RCE access to private staff data and infrastructure.
  • The Fix: Update Ivanti EPMM instances to the latest February 2026 emergency release.

πŸ”΄ πŸ› οΈ Gemini MCP RCE: Tool Command Injection

  • The Incident: (CVE-2026-0755) RCE in tools utilizing the Gemini Model Context Protocol (MCP).
  • The Risk: System Compromise. Attackers can inject OS commands via the execAsync method during tool execution.
  • The Fix: Isolate AI orchestration servers and update all MCP-compatible tools immediately.
  • Sources: SentinelOne

πŸ”΄ πŸ› οΈ n8n AI Workflow: Remote Code Execution

  • The Incident: (CVE-2025-68613) Improper sandboxing of JavaScript expressions within AI workflow nodes.
  • The Risk: System Takeover. Theft of proprietary AI logic and credentials stored in workflow variables.
  • The Fix: Update n8n instances to v1.122.0 or higher.

🟠 🧠 Pentagon Designates Anthropic as AI Supply Chain Risk

  • The Incident: The U.S. Pentagon designated Anthropic as a ‘supply chain risk’ following conflicts over model usage policies. Anthropic’s refusal to allow Claude for mass surveillance or autonomous weapons usage clashed with DoD requirements, leading to restrictions on military integration.
  • The Risk: AI Governance Issues. Restrictions limit the integration of Anthropic technology into defense projects, impacting defense-adjacent supply chains.
  • The Fix: Organizations in defense-adjacent sectors should evaluate vendor risk assessments regarding federal designations.
  • Sources: The Hacker News

🟠 🧠 AI Recommendation Poisoning (SEO Summarization Abuse)

  • The Incident: Microsoft Defender researchers identified attackers embedding hidden instructions in web metadata to hijack chatbot “Summarize” buttons.
  • The Risk: Data Integrity / Fraud. Forced LLM output bias, causing assistants to recommend fraudulent products or malicious sources based on poisoned history.
  • The Fix: Treat external “summarize” prompts as low-trust and regularly clear LLM “memory” caches.

🟠 🌐 AI Agent Swarm Proliferation

  • The Incident: Rapid deployment of autonomous AI entities collaborating without human intervention.
  • The Risk: Lateral Movement. Non-deterministic data flows allow for lateral movement and exfiltration through inter-agent communications.
  • The Fix: Enforce least-privilege for autonomous agents and log all inter-agent API calls.

🟠 πŸ›‘οΈ Reynolds Ransomware: BYOVD Driver Evasion

  • The Incident: Ransomware utilizing “Bring Your Own Vulnerable Driver” to disable EDR and security tooling.
  • The Risk: Encryption. Complete bypass of endpoint protection to encrypt critical server data.
  • The Fix: Implement Comic driver blocklists and enable HVCI on Windows systems.

🟠 πŸ›‘οΈ Apple dyld Zero-Day: CVE-2026-20700

  • The Incident: Memory corruption in Apple’s Dynamic Link Editor (dyld) found being used in targeted attacks.
  • The Risk: Arbitrary Code Execution. System-level access to macOS and iOS devices via malicious app bundles.
  • The Fix: Update all Apple hardware to the Feb 11, 2026 security versions.

🟑 🧠 AI-Powered Identity Fraud via OnlyFake

  • The Incident: The OnlyFake operation used generative AI to create over 10,000 synthetic identification photos to bypass KYC protocols. The site was dismantled after the operator’s guilty plea.
  • The Risk: AI Misuse. Criminals bypass security checks on financial and crypto platforms using realistic fake documents.
  • The Fix: Financial institutions should update KYC systems to include liveness detection and AI artifact analysis.
  • Sources: BleepingComputer

πŸ”΄ πŸ›‘οΈ XMRig BYOVD: Wormable Cryptojacker

  • The Incident: A wormable cryptojacking campaign using Bring Your Own Vulnerable Driver (BYOVD) exploits. Features time-based logic bombs to maximize hashrate before detection.
  • The Risk: System Resource Theft & Stability. Massive CPU/GPU drain and potential system crashes due to driver manipulation.
  • The Fix: Monitor for unauthorized driver loading and unusual CPU spikes; block known vulnerable drivers via HVCI.
  • Sources: The Hacker News

πŸ”΅ πŸ›‘οΈ Roundcube Webmail: Patched Flaws Added to CISA KEV

  • The Incident: CISA added recently patched Roundcube vulnerabilities to its Known Exploited Vulnerabilities (KEV) catalog due to active exploitation in the wild.
  • The Risk: Email Compromise. Access to sensitive communications and credential theft.
  • The Fix: Ensure Roundcube instances are patched to the latest version immediately.
  • Sources: BleepingComputer

🌐 βšͺ Kali Linux Integration with Claude LLM

  • The Incident: Offensive security tools are increasingly integrating LLMs, with Claude now being utilized within the Kali Linux ecosystem. This integration assists users in interpreting complex scan results and generating exploit code.
  • The Risk: AI-Assisted Exploitation. While designed for penetration testers, it significantly lowers the barrier for entry for less-skilled threat actors to conduct advanced attacks.
  • The Fix: Enhance monitoring for rapid-fire vulnerability scanning and automated exploitation patterns that may be augmented by LLM-driven tooling.
  • Sources: The Hacker News

🌐 βšͺ OpenClaw Dark Web Chatter

  • The Incident: Dark web analysis shows high criminal interest in ‘OpenClaw,’ a purported open-source AI exploitation tool.
  • The Risk: Future Threat Potential. While telemetry shows limited operational use now, the “hype cycle” among threat actor interest indicates a shift toward specialized AI malware.
  • The Fix: Continue monitoring threat intel feeds for active operationalization signatures.
  • Sources: BleepingComputer

πŸ“œ Audit Log & Version History

Date Status Action Threat Impacted
2026-03-11 πŸ”΄ Active Added AI-Discovery of Critical RCE (CVE-2026-21536): Autonomous AI agent XBOW independently discovers 9.8-rated vulnerability.
2026-03-11 πŸ”΄ Active Added @openclaw-ai Malicious npm Package: Poisoned supply chain targeting AI/ML developers on macOS.
2026-03-11 🟠 Active Added AI Bot Exploit of CI/CD Pipelines: Automation of Rust crate supply chain compromise via AI bots.
2026-03-10 🟠 Active Added Agentic AI Remediation Readiness Gap: Analysis of governance risks in autonomous threat response.
2026-03-10 🟠 Active Updated InstallFix Fake Claude Tools: Added Dark Reading report on sophisticated malvertising redirection.
2026-03-09 πŸ”΄ Active Updated OpenClaw Theft: Added reports of exposed administrative interfaces leaking API keys and tokens.
2026-03-09 πŸ”΄ Active Updated Cline CLI Poisoning: Added Clinejection supply chain compromise via prompt injection.
2026-03-09 πŸ”΄ Active Updated FortiGate AI Recon: Added AWS report of AI-augmented mass exploitation of 600 firewalls.
2026-03-08 🟠 Active Added OpenAI Codex Security Agent, Claude Opus 4.6 Firefox Zero-Day, Microsoft AI Threat Report.
2026-03-07 🟠 Active Added Mexico Government LLM Exploitation, Transparent Tribe AI-Malware, InstallFix Fake Claude Tools.
2026-03-06 🟠 Active Added APT36 AI-Malware Assembly Line: Documented automated malware generation by nation-state actors.
2026-03-06 🟠 Active Updated OpenClaw Theft: Bing AI search feature observed promoting malicious ‘OpenClaw’ installers.
2026-03-06 🟑 Active Added Enterprise Browser AI Proliferation: Report on blind spots created by browser-based AI tool usage.
2026-03-05 🟠 Active Added Enterprise AI Governance Crisis: Documented board-level oversight gaps and new RFP standards.
2026-03-04 πŸ”΄ Active Added CyberStrikeAI Global Campaign: Documented orchestration of FortiGate attacks in 55 countries.
2026-03-04 🟠 Active Added MCP Identity Risk: Documented rapid adoption risks of Model Context Protocol agents.
2026-03-03 πŸ”΄ Active Merged OpenClaw Critical Vulnerabilities: Integrated critical orchestration layer flaws and input sanitization risks.
2026-03-03 🟠 Active Added Deepfake Injection Attacks, Chrome Gemini PrivEsc.
2026-03-03 🟑 Active Added Claude Global Outage.
2026-03-02 πŸ”΄ Active Merged OpenClaw Theft: Integrated ClawJacked hijacking and local brute-force reports.
2026-03-02 πŸ”΄ Active Updated Claude Code RCE: Added findings on security control gaps and human-in-the-loop requirements.
2026-03-02 🟠 Active Added Pentagon Anthropic Supply Chain Risk.
2026-03-02 🟑 Active Added AI-Powered Identity Fraud via OnlyFake.
2026-02-27 πŸ”΄ Active Added Google Gemini: Client-Side API Key Exposure, Kali Linux: Claude LLM Integration.
2026-02-26 πŸ”΄ Active Added Claude Code RCE, ChatGPT Misuse, OpenClaw Dark Web Chatter.
2026-02-25 πŸ”΄ Active Added RoguePilot, AI-Augmented Exploitation.
2026-02-25 πŸ”΄ Active Updated M365 Copilot DLP Bypass: Expanded with Microsoft’s “AugLoop” mandatory DLP controls.
2026-02-24 πŸ”΄ Active Added Claude Model Distillation, XMRig BYOVD, Roundcube KEV.
2026-02-24 πŸ”΄ Active Merged FortiGate AI-Armed Exploitation: Merged amateur AI automation reports with previous mass credential abuse entry.
2026-02-24 🟠 Active Merged LLM Infrastructure & API Exposure: Integrated general exposed endpoint research with NVIDIA Triton layer exploits.
2026-02-23 πŸ”΄ Active Added M365 Copilot DLP Bypass, Arkanix Stealer
2026-02-22 πŸ”΄ Active Updated OpenClaw Theft: Expanded to include misconfigured instance exploitation.
2026-02-21 πŸ”΄ Active Added Cline CLI v2.3.0 Poisoning, AI Agent Guardrail Bypass, AI Infrastructure Exploitation, FortiGate AI Recon
2026-02-21 πŸ”΄ Active Updated BeyondTrust RCE: Updated to reflect active exploitation by ransomware groups.
2026-02-21 πŸ”΄ Active Added MIMICRAT Deployment via ClickFix campaigns.
2026-02-17 πŸ”΄ Active Added Copilot & Grok C2 Proxy Abuse
2026-02-17 πŸ”΄ Active Added OpenClaw Secret Exfiltration
2026-02-17 πŸ”΄ Active Consolidated AiFrame: Merged 30+ new fake AI Chrome extensions into AiFrame Extensions.
2026-02-17 🟠 Active Updated AI Recommendation Poisoning: Updated with Microsoft Defender SEO manipulation findings.
2026-02-17 πŸ”΄ Active Updated BeyondTrust RCE: Updated with CISA emergency mandate.
2026-02-16 πŸ”΄ Active Added Cursor AI RCE (CVE-2026-26268)
2026-02-16 πŸ”΄ Active Consolidated ClickFix: Merged AMOS macOS and ClickFix social engineering into ClickFix: Artifact Abuse.
2026-02-16 πŸ”΄ Active Updated UNC2970 Gemini Recon: Expanded to include state-backed Gemini weaponization.
2026-02-16 🟠 Active Added AI Agent Swarm Proliferation
2026-02-13 πŸ”΄ Active Added AiFrame Extensions
2026-02-13 🟠 Active Added AI Recommendation Poisoning
2026-02-13 🟒 Mitigated Consolidated Model Extraction: Merged into UNC2970 Gemini Weaponization.
2026-02-12 πŸ”΄ Active Added UNC1069 AI Lures, Apple dyld Zero-Day
2026-02-11 πŸ”΄ Active Added Microsoft Zero-Days, ZeroDayRAT
2026-02-10 πŸ”΄ Active Added UNC3886 Telco Breach, ClawHavoc Skills, BeyondTrust RCE, VS Code AI Malware, Gemini MCP RCE
2026-02-10 🟠 Active Added Reynolds Ransomware
2026-02-09 🟒 Mitigated Removed TeamPCP Ray Exploit: Consolidated into TeamPCP Cloud Takeover.
2026-02-06 πŸ”΄ Active Added TeamPCP Cloud Takeover, n8n RCE, Ivanti Zero-Day
2026-02-06 🟒 Mitigated Removed Moltbook Leak: Database secured.
2026-02-06 🟒 Mitigated Removed Claude 4.6 Discovery: Research phase concluded.
2026-02-04 🟒 Mitigated Removed SolarWinds RCE: Patch window closed.
2026-02-03 🟒 Mitigated Removed DockerDash RCE: Fixed in v4.50.0.

🏠 Home ← Back to AI Guides
πŸ†˜ Need help getting AI to do what you want? Start with Help! I’m Stuck