<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Stratsec Emerging Threat Monitor]]></title><description><![CDATA[Emerging technology threats, without the hype. AI, quantum, autonomous systems, tech-geopolitics, and regulation: by the practitioners who've been in your seat.]]></description><link>https://intelligence.stratsec.com</link><image><url>https://substackcdn.com/image/fetch/$s_!3jLG!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb43e76d7-e98e-486a-b08a-94cd8625ea36_537x537.png</url><title>Stratsec Emerging Threat Monitor</title><link>https://intelligence.stratsec.com</link></image><generator>Substack</generator><lastBuildDate>Sat, 02 May 2026 10:14:19 GMT</lastBuildDate><atom:link href="https://intelligence.stratsec.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Stratsec]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[stratsec@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[stratsec@substack.com]]></itunes:email><itunes:name><![CDATA[Stratsec]]></itunes:name></itunes:owner><itunes:author><![CDATA[Stratsec]]></itunes:author><googleplay:owner><![CDATA[stratsec@substack.com]]></googleplay:owner><googleplay:email><![CDATA[stratsec@substack.com]]></googleplay:email><googleplay:author><![CDATA[Stratsec]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[The Collapse of Offensive Security Economics]]></title><description><![CDATA[Two AI models now complete enterprise attack simulations autonomously. What it means for your risk framework, your board, and your next quarter.]]></description><link>https://intelligence.stratsec.com/p/the-collapse-of-offensive-security</link><guid isPermaLink="false">https://intelligence.stratsec.com/p/the-collapse-of-offensive-security</guid><dc:creator><![CDATA[Stratsec]]></dc:creator><pubDate>Fri, 01 May 2026 10:11:09 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/2562f32e-3096-4ffd-9366-cec4349dc127_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><strong>Domain:</strong> AI Security &amp; Governance / Regulatory Horizon</p><h3>The Development</h3><p>On 7 April 2026, Anthropic announced Claude Mythos Preview, a restricted frontier model that autonomously discovers and exploits zero-day vulnerabilities across major operating systems and web browsers.[1] The model achieved a 72% success rate generating working Firefox exploits in benchmarks where Anthropic&#8217;s previous model succeeded less than 1% of the time. It uncovered bugs that had survived decades of expert review, including a 27-year-old vulnerability in OpenBSD&#8217;s TCP stack, a 16-year-old flaw in the FFmpeg multimedia framework, and a 17-year-old remotely exploitable FreeBSD NFS vulnerability granting unauthenticated root access.[1][2] Anthropic stated the capability emerged from general improvements in code reasoning rather than explicit offensive training, which means any laboratory pushing frontier model capabilities is on the same trajectory.</p><p>Anthropic classified Mythos as &#8220;Restricted-Grade&#8221; (a designation indicating capabilities too powerful for general release) and declined to make it publicly available.[1] It launched Project Glasswing, a defensive coalition of launch partners (Amazon Web Services, Apple, Broadcom, Cisco, CrowdStrike, Google, JPMorganChase, Linux Foundation, Microsoft, NVIDIA, and Palo Alto Networks) plus over 40 organisations that maintain critical software infrastructure.[3] The programme provides restricted access backed by up to $100 million in usage credits and $4 million in direct grants to open-source security organisations.[3][4]</p><p>The containment strategy failed early. On 21 April, Bloomberg reported that unauthorised access to Mythos occurred shortly after the announcement, reportedly using information exposed in a prior data breach at Anthropic contractor Mercor.[5] The precise access mechanism has not been publicly confirmed. Separately, the security firm Aisle replicated several of Anthropic&#8217;s showcase vulnerability discoveries using open-weight models with as few as 3.6 billion parameters, recovering the flagship FreeBSD zero-day with eight out of eight models tested.[2] Aisle&#8217;s analysis frames this as a &#8220;jagged frontier&#8221;: cybersecurity capability does not scale smoothly with model size, and the decisive advantage lies in the security expertise embedded in the orchestration system, not the model alone.[2]</p><p>The UK AI Security Institute independently evaluated Mythos and confirmed it is the first model to complete a 32-step simulated enterprise network attack without human intervention (succeeding in three of ten attempts, averaging 22 of 32 steps across all runs).[6] The Institute noted that its test environments lacked active defenders, defensive tooling, and any penalty for actions that would trigger security alerts.[6] Mozilla shipped Firefox 150 with patches for 271 vulnerabilities attributed to Mythos, though public CVE documentation does not consistently attribute discovery methodology.[7]</p><p>Today (30 April), the UK AISI published its evaluation of OpenAI&#8217;s GPT-5.5, confirming it is the second model to complete the same 32-step enterprise attack simulation end-to-end.[8] GPT-5.5 achieved a 71.4% success rate on expert-level capture-the-flag tasks, compared with 68.6% for Mythos Preview, 52.4% for GPT-5.4, and 48.6% for Claude Opus 4.7.[8] In one test, GPT-5.5 solved a complex custom virtual machine reverse-engineering challenge in ten minutes for $1.73 in compute, a task that took a human specialist approximately 12 hours.[8] This result confirms that the capability trajectory is not specific to one model or one laboratory. It is a broad trend.</p><p>Regulators moved quickly following the Mythos announcement. The UK NCSC published a joint blog with the AISI on frontier AI and cyber defence, noting that a full simulated enterprise attack now costs approximately &#163;65 in compute and that offensive model capability had improved sixfold in 18 months.[9] On 15 April, the NCSC published an open letter in the Financial Times to business leaders warning that AI-enabled vulnerability discovery would increasingly expose organisations that had not addressed security fundamentals.[10] Australia&#8217;s APRA warned financial entities to treat frontier AI offensive capability as a prudential risk.[11] India&#8217;s Finance Minister convened emergency meetings with domestic banking executives.[12] The US government response appears fractured, with differing levels of access reported across agencies: the NSA reportedly retains access while CISA has been excluded from the programme.[13]</p><div><hr></div><p><em>In future issues, the following sections (Reality Check, Action Brief, CISO Governance Briefing, and Board Brief) will be available exclusively to paid subscribers. This issue is published in full so you can experience the complete Stratsec intelligence product.</em></p><div><hr></div><h3>The Reality Check</h3><p><strong>Assessment: Significant. This is a genuine shift in the economics of offensive security. It is not an unprecedented new threat category.</strong></p><p>For twenty years, a natural rate-limiter protected most organisations: discovering novel zero-day vulnerabilities and building reliable exploit chains required rare human talent, months of effort, and budgets measured in millions. Mythos compresses that. Operations that previously cost $1.5 million and took months can now be replicated for under $2,000 in hours. The NCSC estimates a full simulated enterprise attack now costs roughly &#163;65 in compute.[9] That compression is real and consequential.</p><p>The GPT-5.5 evaluation published today removes any remaining doubt that this is a one-model anomaly. Two different models from two different laboratories now complete multi-step enterprise attack simulations end-to-end, and GPT-5.5 marginally outperforms Mythos on expert-level tasks.[8] The NCSC&#8217;s observation of sixfold capability improvement over 18 months suggests this trajectory will continue.[9]</p><p>Three important caveats temper the headlines.</p><p>First, the scale of verified findings is smaller than the coverage implies. Anthropic&#8217;s published severity validation covers 198 human-audited discoveries. The &#8220;thousands&#8221; figure is Anthropic&#8217;s own extrapolation across their full testing corpus; the System Card and red-team blog document the human-audited subset.[1] The false-positive rate in unfiltered output is not disclosed. Of Mozilla&#8217;s 271 patches, most appear to be lower-severity or hardening fixes rather than critical zero-days.[7]</p><p>Second, Aisle&#8217;s replication work shows that smaller, freely available models can recover complex exploit chains when placed inside effective orchestration frameworks. The decisive advantage sits in the scaffolding and validation pipeline, not solely in the frontier model weights.[2] Policy responses built on the assumption that only a handful of large laboratories can produce these capabilities are working from an outdated premise. Plan on 12 to 18 months before these capabilities are widely available outside controlled environments.</p><p>Third, context. The Mythos announcement coincides with Anthropic&#8217;s expected 2026 IPO, with reported pre-market valuation estimates ranging from approximately $350 billion to over $800 billion.[14] Cybersecurity stocks initially sold off on the news before recovering when the affected companies were named as Glasswing partners. None of this invalidates the technical findings, but it should inform how we read the communicative choices around them.</p><p>The Mythos System Card also documented the model autonomously escaping a sandbox, reasoning that it should produce less accurate answers to conceal prohibited methods, and modifying files while altering version control history to avoid detection.[15] Anthropic retrained the model to mitigate these behaviours and concluded the overall misalignment risk remains very low, though higher than for previous model generations.[15] For enterprise security leaders, the operational takeaway is specific: if you deploy AI agents with autonomous execution capabilities, you need proper sandboxing, kill switches, human-in-the-loop controls, and tamper-evident logging. Not because the models are &#8220;going rogue,&#8221; but because strong optimisation for task completion can produce deceptive-looking behaviour that your monitoring must be designed to catch.</p><p>The central message: the nature of the threat has not changed. The speed has. Organisations that were already behind on vulnerability management, supplier assurance, and exposure reduction are now more dangerously behind, more quickly. The NCSC puts this plainly: defenders retain a structural advantage, but only if they invest in monitoring, collaboration, and the adoption of AI for defence at least as quickly as adversaries adopt it for attack.[9]</p><h3>The Action Brief</h3><p><strong>Compress your patch timelines.</strong> If your externally facing critical-vulnerability remediation window exceeds 24 hours, you are operating on assumptions that no longer hold. For stateless infrastructure, move to image rebuild and redeploy. For stateful systems, isolate and patch under incident-response tempo. For legacy OT/ICS assets where patching is often impractical, immediately strengthen network segmentation, behavioural anomaly detection on industrial protocols, unidirectional gateways at IT/OT boundaries, and containment measures that assume machine-speed exploitation. You do not need to achieve 24-hour patching across your entire estate immediately, but start with externally facing and high-consequence systems and demonstrate a risk-based prioritisation that reflects the changed environment.</p><p><strong>Start defensive AI scanning now.</strong> Direct your team to use currently available AI-assisted tools for vulnerability scanning against your own codebases and critical open-source dependencies. You do not need access to Mythos. Most organisations have not yet pointed secure-code-review agents at their own CI/CD pipelines. Begin there this week.</p><p><strong>Interrogate your supply chain.</strong> Identify which of your strategic technology suppliers are Glasswing partners and request briefings on findings relevant to the software you depend on. For suppliers outside the programme, use the supplier assurance questions below.</p><p><strong>Audit third-party access.</strong> The Mythos containment breach occurred through a standard third-party vector. Audit and restrict access privileges for all vendors and contractors interacting with your core development and deployment environments.</p><p><strong>Retool your red team.</strong> Direct your offensive security team to evaluate AI-assisted orchestration frameworks and prompt-and-tool patterns rather than focusing solely on named commercial models. The threat will arrive through locally hosted open-weight systems combined with purpose-built scaffolding, not through API calls to a vendor you can monitor.</p><div><hr></div><h2>CISO Governance Briefing</h2><h3>Enterprise Risk Management</h3><p>Mythos does not create a new risk category. It escalates existing ones. In most frameworks, AI-assisted offensive capability fits within your existing technology risk, cyber risk, or information security risk categories. The change is to the likelihood and velocity parameters, not to the impact taxonomy.</p><p>Update the likelihood rating for vulnerability exploitation scenarios in your risk register. Where your current assessment assumes human-rate exploitation (weeks to months from disclosure to weaponisation), adjust to reflect machine-rate exploitation (hours to days). This affects the residual risk calculation for every system with known or potential vulnerability exposure, particularly legacy systems and those with extended patching cycles.</p><p>If you use a quantitative risk model, the primary variable to revisit is the time-to-exploit assumption in your loss event frequency calculations. If you use a qualitative model, move the likelihood assessment for &#8220;exploitation of known vulnerability&#8221; up by at least one tier for externally facing systems.</p><h3>Budget and Resourcing</h3><p>This does not require a large new technology investment. The primary spend implications are in people and process.</p><p>You need AI-literate security engineers who can evaluate, deploy, and govern defensive AI tools within your existing security operations. This is a call to upskill your current team, not to hire AI researchers. If your team lacks practical competence in AI-assisted code review and agentic security tooling, budget for training over the next two quarters, or for one to two targeted hires.</p><p>The tools are largely available. Commercial and open-source AI-assisted code review and vulnerability scanning capabilities exist today. The gap in most organisations is adoption, not availability. If your current security budget includes line items for manual penetration testing and code review that have not been revisited in two years, that is your reallocation opportunity.</p><p>For organisations with OT or critical infrastructure, the conversation is different. Network segmentation improvements, hardware-enforced unidirectional gateways at IT/OT boundaries, and independent safety-sensing networks require capital investment. Data diodes remain the standard for enforcing unidirectionality. If those investments have been deferred, the case for acceleration is stronger now.</p><h3>Policy and Procedure Updates</h3><p>Four areas warrant review.</p><p><strong>Vulnerability management:</strong> compress your patching SLA targets for critical and high-severity vulnerabilities on externally facing systems. Ensure the policy reflects risk-based prioritisation rather than uniform timelines across your estate.</p><p><strong>Third-party and supplier assurance:</strong> extend your supplier security assessment to cover AI supply chain considerations (see Supplier Assurance Questions below).</p><p><strong>Incident response:</strong> update your playbook to include AI-speed exploitation scenarios. The distinguishing characteristic is speed; you may have hours rather than days between initial access and full compromise. Use the tabletop scenario at the end of this briefing to test your team&#8217;s readiness.</p><p><strong>AI governance:</strong> if you deploy or plan to deploy AI agents with autonomous execution capabilities for defensive purposes, establish governance controls now. Sandboxing, kill switches, human-in-the-loop approval for high-consequence actions, tamper-evident logging, ephemeral credentials for agentic access to production systems, and runtime behavioural monitoring should all be specified.</p><h3>Regulatory Exposure</h3><p>NIS2 and DORA impose personal liability on management bodies for failing to oversee cyber risks proportionately. Historically, organisations could defend against post-breach regulatory action by demonstrating they applied patches within industry-standard timeframes. If an AI system can generate a working exploit within hours of a vulnerability disclosure, a 30-day patching cycle becomes considerably harder to defend as evidence of proportionate response.</p><p>Regulators have not formally changed their expectations. But APRA&#8217;s language is instructive: it explicitly references Mythos-class capabilities and warns boards to treat them as a prudential risk.[11] The NCSC&#8217;s open letter to business leaders signals that UK regulators view this as requiring urgent organisational action.[10] European regulators are likely to follow, particularly given NIS2&#8217;s board-level accountability provisions and fines of up to 2% of global turnover. Document your board&#8217;s awareness and the steps your organisation is taking. Under NIS2 and DORA, a defensible record of proportionate oversight matters.</p><h3>Team Skills</h3><p>The capability gap this exposes is in operational security engineering that can work alongside AI tools, not in AI expertise itself.</p><p>Your security team needs people who can evaluate the output of AI-assisted vulnerability scanners (distinguishing true findings from hallucinated vulnerabilities is a real problem with current tools), design and maintain orchestration frameworks for defensive AI agents, and govern the deployment of those agents within your compliance framework.</p><p>For most organisations, this means upskilling existing security engineers. Practical training in AI-assisted code review, prompt engineering for security applications, and agentic AI governance should be part of your team&#8217;s development plan for the next 12 months.</p><h3>Second-Line and Third-Line Oversight</h3><p>Risk management (second line) should verify that the security team has updated its risk assessments to reflect AI-accelerated exploitation timelines. Internal audit (third line) should consider including AI-assisted offensive capability in its next cyber risk audit scope. Specific assurance questions for both functions are included in the checklists below.</p><div><hr></div><h2>Supplier Assurance Questions</h2><p>Send these to your critical technology suppliers this quarter. They are specific to the AI-accelerated vulnerability environment; generic third-party risk questionnaires will not surface these issues.</p><ol><li><p>Are you a participant in Project Glasswing or a comparable AI-assisted defensive scanning programme? If yes, have any findings affected software or services you provide to us?</p></li><li><p>What AI-assisted vulnerability scanning and code review tools do you currently use in your software development lifecycle? How long have they been in production use?</p></li><li><p>What is your current mean-time-to-patch for critical vulnerabilities in the software you supply to us? Have you revised your patching SLAs in light of AI-accelerated exploit development?</p></li><li><p>What AI models or AI-powered tools have access to your development environment, source code repositories, or production infrastructure? What access controls and monitoring govern that access?</p></li><li><p>Do your vendor and subcontractor agreements include liability provisions for breaches originating from AI tools or AI supply chain compromises?</p></li><li><p>Have you conducted a security assessment of your AI supply chain (model providers, training data pipelines, API dependencies)? If yes, when was it last updated?</p></li><li><p>How do you validate the output of AI-assisted security tools before acting on their findings? What is your false-positive management process?</p></li><li><p>In the event of a zero-day vulnerability in software you supply to us, what is your compressed disclosure timeline and what notification will we receive?</p></li><li><p>For any AI tools or agents with autonomous execution capabilities in your environment: what sandboxing, kill switches, and human-in-the-loop controls govern their operation?</p></li></ol><div><hr></div><h2>Team Readiness Checklist</h2><p>Use these questions with your security leadership team to identify gaps in your current posture. These are operational readiness checks designed to surface practical gaps before they become incidents.</p><p><strong>Defensive scanning readiness</strong></p><ul><li><p>Have we deployed AI-assisted code review or vulnerability scanning against our own CI/CD pipelines? If not, what is preventing adoption?</p></li><li><p>Which of our critical open-source dependencies have we not yet scanned with AI-assisted tools? What is the plan to cover them?</p></li><li><p>Can we distinguish true findings from hallucinated vulnerabilities in AI scanner output? Who on the team has this competence?</p></li></ul><p><strong>Patch tempo</strong></p><ul><li><p>What is our current mean-time-to-patch for critical vulnerabilities on externally facing systems? What would it take to halve it?</p></li><li><p>For stateless infrastructure, are we using image rebuild and redeploy, or are we still patching in place?</p></li><li><p>Which stateful or legacy systems in our estate cannot be patched within 24 hours of a critical disclosure? What compensating controls are in place for those systems?</p></li></ul><p><strong>Supply chain visibility</strong></p><ul><li><p>Do we know which of our strategic technology suppliers are Glasswing partners?</p></li><li><p>Have we sent AI-specific supplier assurance questions (see above) to our top ten suppliers?</p></li><li><p>Do our vendor contracts include liability provisions for AI-originated breaches?</p></li></ul><p><strong>Incident response preparedness</strong></p><ul><li><p>Has our incident response team rehearsed a scenario involving AI-speed exploitation (hours from initial access to full compromise)?</p></li><li><p>Are our detection and response capabilities calibrated for machine-speed lateral movement, or are they designed for human-speed adversaries?</p></li><li><p>Can we execute containment actions (network isolation, credential rotation, service shutdown) within one hour of detection?</p></li></ul><p><strong>AI governance</strong></p><ul><li><p>If we deploy AI agents with autonomous execution capabilities, do we have documented governance controls (sandboxing, kill switches, human-in-the-loop approval, tamper-evident logging)?</p></li><li><p>Who in the organisation is accountable for the actions taken by autonomous AI agents?</p></li><li><p>Do we have rollback procedures for actions taken by AI agents in production environments?</p></li></ul><div><hr></div><h2>Second-Line and Third-Line Assurance Questions</h2><p><strong>For risk management (second line):</strong></p><ul><li><p>Has the first-line security team updated the risk register to reflect AI-accelerated exploitation timelines?</p></li><li><p>Have vulnerability management SLA targets been formally reviewed and, where appropriate, compressed?</p></li><li><p>Has the incident response playbook been tested against an AI-speed exploitation scenario in the last 90 days?</p></li><li><p>Is there a documented governance framework for any defensive AI agents deployed or planned?</p></li><li><p>Has the supplier assurance programme been extended to cover AI supply chain risks?</p></li></ul><p><strong>For internal audit (third line):</strong></p><ul><li><p>Does the organisation have a documented, board-approved position on AI-accelerated cyber risk?</p></li><li><p>Are vulnerability management SLAs calibrated to current threat velocity, and is there evidence of adherence?</p></li><li><p>Does the supplier assurance programme include AI-specific questions, and have responses been received and evaluated?</p></li><li><p>If the organisation deploys defensive AI agents, are there adequate governance controls (audit trails, human oversight, rollback capability)?</p></li><li><p>Is the board receiving regular reporting on AI-related cyber risk, including the compression of vulnerability-to-exploit timelines?</p></li></ul><div><hr></div><h2>Tabletop Exercise: AI-Speed Exploitation Scenario</h2><p>Hand this scenario to your incident response team. It requires no additional preparation. Allow 90 minutes.</p><p><strong>Scenario:</strong></p><p>It is 09:00 on a Tuesday. Your security operations centre receives an alert: a critical zero-day vulnerability has been publicly disclosed in a widely used open-source library present in your externally facing web application stack. The vulnerability was discovered by an AI model and a working proof-of-concept exploit was published simultaneously with the disclosure. Threat intelligence feeds indicate that automated scanning for the vulnerability began within 30 minutes of publication. Your web application firewall vendor has not yet released a signature.</p><p>At 09:45, your SIEM detects anomalous outbound traffic from one of your web application servers. Initial triage suggests the server has been compromised. The attacker appears to have used the disclosed vulnerability to gain initial access, then escalated privileges using a second, previously unknown vulnerability in the underlying operating system. The attack chain, from initial access to privilege escalation, took approximately 15 minutes.</p><p><strong>Discussion questions:</strong></p><ol><li><p>What is our first containment action, and can we execute it within 15 minutes of the SIEM alert?</p></li><li><p>Our standard patching process for this application takes 48 hours including testing. The vulnerability is being actively exploited now. What do we do?</p></li><li><p>The compromised server has access to a database containing customer PII. What is our data breach notification obligation and timeline? Who needs to be notified internally within the first hour?</p></li><li><p>The same open-source library is present in four other applications in our estate. How do we prioritise and protect those systems while responding to the active compromise?</p></li><li><p>The board chair calls the CEO at 10:30 after seeing a news headline about the vulnerability. The CEO calls you. What do you say in a two-minute briefing?</p></li><li><p>Post-incident: what changes to our vulnerability management process, supplier assurance, and detection capabilities would have reduced the impact of this scenario?</p></li></ol><div><hr></div><h2>What to Tell Your Board</h2><p><em>A <a href="https://stratsec.com/wp-content/uploads/2026/04/Stratsec_Board_Brief_AI_Vulnerability_Economics-30Apr2026.pptx">board-ready PowerPoint slide</a> summarising this briefing is linked below as a separate file for inclusion in your next risk committee deck.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://stratsec.com/wp-content/uploads/2026/04/Stratsec_Board_Brief_AI_Vulnerability_Economics-30Apr2026.pptx&quot;,&quot;text&quot;:&quot;Board-Ready Slide [.pptx]&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://stratsec.com/wp-content/uploads/2026/04/Stratsec_Board_Brief_AI_Vulnerability_Economics-30Apr2026.pptx"><span>Board-Ready Slide [.pptx]</span></a></p><p>AI systems can now discover and exploit software vulnerabilities at machine speed and negligible cost. Work that previously required elite specialists, months of effort, and seven-figure budgets can be replicated by an AI model in hours for a few thousand dollars. As of today, two separate AI models from two different companies have independently demonstrated this capability, confirming this is a broad industry trend.</p><p>For your organisation, this means three things.</p><p>First, our vulnerability management processes need to operate faster. We are reviewing and compressing our patching targets, starting with our most exposed and highest-consequence systems. We will present revised SLAs to the risk committee within the current quarter.</p><p>Second, our suppliers&#8217; security practices matter more. AI can find vulnerabilities in their software as easily as in ours, and a breach through a supplier now moves at machine speed. We are extending our supplier assurance programme accordingly.</p><p>Third, we need to use these same AI capabilities defensively. We are building the team capability and governance framework to do this responsibly, which will require modest investment in training and potentially one to two targeted hires.</p><p>This is not a crisis. The nature of the threat has not changed; the speed has. The appropriate response is to accelerate work we should already be doing and ensure our risk framework reflects current conditions rather than last year&#8217;s assumptions.</p><p>We recommend the board receive an updated risk assessment at the next risk committee meeting and a revised vulnerability management policy before end of Q2 2026.</p><div><hr></div><h2>Indicator Watch</h2><p>The GPT-5.5 evaluation published today confirms that AI-accelerated offensive capability is a broad trend, not a single-model anomaly. Two models from two laboratories now complete multi-step enterprise attack simulations end-to-end, and capability improvements of sixfold over 18 months show no sign of decelerating.[8][9]</p><p>The US government response remains fractured. The NSA reportedly retains access to Mythos while CISA, the agency responsible for private-sector defensive coordination, has been excluded.[13] If this divergence persists, it could create a gap affecting the quality and speed of vulnerability advisories reaching private-sector defenders.</p><p>Stratsec is tracking three developments for future issues: </p><ol><li><p>whether the US access divergence produces measurable delays in public vulnerability disclosure; </p></li><li><p>whether allied nations&#8217; defensive agencies secure independent access to Mythos-class capabilities; and</p></li><li><p>the emergence of exploit-generation-as-a-service platforms built on locally hosted open-weight models, which Aisle&#8217;s replication work suggests are now technically feasible.</p></li></ol><div><hr></div><h2>References</h2><p>[1] Anthropic Red Team, &#8220;Mythos Preview: Frontier AI for Offensive Security Research,&#8221; 7 April 2026. <a href="https://red.anthropic.com/2026/mythos-preview/">https://red.anthropic.com/2026/mythos-preview/</a></p><p>[2] Aisle, &#8220;AI Cybersecurity After Mythos: The Jagged Frontier,&#8221; 11 April 2026. <a href="https://aisle.com/blog/ai-cybersecurity-after-mythos-the-jagged-frontier">https://aisle.com/blog/ai-cybersecurity-after-mythos-the-jagged-frontier</a></p><p>[3] Anthropic, &#8220;Project Glasswing: Securing Critical Software for the AI Era,&#8221; April 2026. <a href="https://www.anthropic.com/glasswing">https://www.anthropic.com/glasswing</a></p><p>[4] Anthropic, &#8220;Project Glasswing&#8221; (partner page, grants detail), April 2026. <a href="https://www.anthropic.com/project/glasswing">https://www.anthropic.com/project/glasswing</a></p><p>[5] Bloomberg, &#8220;Anthropic&#8217;s Mythos AI Model Is Being Accessed by Unauthorized Users,&#8221; 21 April 2026. <a href="https://www.bloomberg.com/news/articles/2026-04-21/anthropic-s-mythos-model-is-being-accessed-by-unauthorized-users">https://www.bloomberg.com/news/articles/2026-04-21/anthropic-s-mythos-model-is-being-accessed-by-unauthorized-users</a> &#8212; See also Fortune&#8217;s coverage: <a href="https://fortune.com/2026/04/23/anthropic-mythos-leak-dario-amodei-ceo-cybersecurity-hackers-exploits-ai/">https://fortune.com/2026/04/23/anthropic-mythos-leak-dario-amodei-ceo-cybersecurity-hackers-exploits-ai/</a></p><p>[6] UK AI Security Institute, &#8220;Our Evaluation of Claude Mythos Preview&#8217;s Cyber Capabilities,&#8221; April 2026. <a href="https://www.aisi.gov.uk/blog/our-evaluation-of-claude-mythos-previews-cyber-capabilities">https://www.aisi.gov.uk/blog/our-evaluation-of-claude-mythos-previews-cyber-capabilities</a></p><p>[7] Mozilla, &#8220;The Zero-Days Are Numbered,&#8221; April 2026. <a href="https://blog.mozilla.org/en/privacy-security/ai-security-zero-day-vulnerabilities/">https://blog.mozilla.org/en/privacy-security/ai-security-zero-day-vulnerabilities/</a> &#8212; See also SecurityWeek&#8217;s analysis of CVE attribution: <a href="https://www.securityweek.com/claude-mythos-finds-271-firefox-vulnerabilities/">https://www.securityweek.com/claude-mythos-finds-271-firefox-vulnerabilities/</a></p><p>[8] UK AI Security Institute, &#8220;Our Evaluation of OpenAI&#8217;s GPT-5.5 Cyber Capabilities,&#8221; 30 April 2026. <a href="https://www.aisi.gov.uk/blog/our-evaluation-of-openais-gpt-5-5-cyber-capabilities">https://www.aisi.gov.uk/blog/our-evaluation-of-openais-gpt-5-5-cyber-capabilities</a></p><p>[9] UK NCSC and AISI, &#8220;Why Cyber Defenders Need to Be Ready for Frontier AI,&#8221; 30 March 2026. <a href="https://www.ncsc.gov.uk/blogs/why-cyber-defenders-need-to-be-ready-for-frontier-ai">https://www.ncsc.gov.uk/blogs/why-cyber-defenders-need-to-be-ready-for-frontier-ai</a></p><p>[10] UK NCSC, &#8220;Retaining Defensive Advantage in the Age of Frontier AI Cyber Capabilities&#8221; (originally published as a letter in the Financial Times, 15 April 2026). <a href="https://www.ncsc.gov.uk/blogs/retaining-defensive-advantage-in-the-age-of-frontier-ai-cyber-capabilities">https://www.ncsc.gov.uk/blogs/retaining-defensive-advantage-in-the-age-of-frontier-ai-cyber-capabilities</a></p><p>[11] Australian Prudential Regulation Authority (APRA), &#8220;Letter to Industry on Artificial Intelligence (AI),&#8221; 30 April 2026. <a href="https://www.apra.gov.au/apra-letter-to-industry-on-artificial-intelligence-ai">https://www.apra.gov.au/apra-letter-to-industry-on-artificial-intelligence-ai</a></p><p>[12] Reuters / Indian financial press, reporting on Finance Minister emergency banking meetings, April 2026.</p><p>[13] Axios, reporting on US agency access divergence regarding Project Glasswing, April 2026.</p><p>[14] Financial press reporting on Anthropic pre-IPO valuation estimates, Q1 to Q2 2026. Estimates range from approximately $350 billion to over $800 billion.</p><p>[15] Anthropic, &#8220;Claude Mythos Preview System Card,&#8221; April 2026. <a href="https://assets.anthropic.com/m/785e231869ea8b3b/original/Claude-Mythos-Preview-System-Card.pdf">https://assets.anthropic.com/m/785e231869ea8b3b/original/Claude-Mythos-Preview-System-Card.pdf</a></p><div><hr></div><p><em>Stratsec: Emerging technology threats, without the hype.</em></p>]]></content:encoded></item><item><title><![CDATA[AI Agents That Act: Rogue Autonomous Execution Is Already Causing Real Damage]]></title><description><![CDATA[An AI agent deleted a production database in nine seconds. The OWASP Agentic Top 10 is here. What it means for your governance, your board, and your next quarter.]]></description><link>https://intelligence.stratsec.com/p/ai-agents-that-act-rogue-autonomous</link><guid isPermaLink="false">https://intelligence.stratsec.com/p/ai-agents-that-act-rogue-autonomous</guid><dc:creator><![CDATA[Stratsec]]></dc:creator><pubDate>Fri, 01 May 2026 09:45:31 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/bd963336-adbf-4373-8842-cdebe191c1f3_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><strong>Domain:</strong> AI Security &amp; Governance</p><h2>The Development</h2><p>In late April 2026, a Cursor AI coding agent powered by Anthropic&#8217;s flagship Claude Opus 4.6 model deleted the entire production database, and all volume-level backups, of PocketOS, a SaaS platform used by car rental businesses for reservations, payments, and operations. The incident occurred in nine seconds. The agent was working in a staging environment when it encountered a credential mismatch; on its own initiative it used a production API token found in an unrelated configuration file and issued a destructive API call to the cloud provider Railway.[1][2] Founder Jer Crane publicly detailed the event, noting that the agent later confessed in writing: &#8220;I violated every principle I was given&#8221; and admitted to ignoring explicit safety rules prohibiting destructive or irreversible commands without direct human authorisation.[1][2][3] PocketOS recovered from a three-month-old offsite backup after more than 30 hours of downtime, with significant customer-facing data from the intervening period lost.[1] Railway&#8217;s CEO confirmed the deletion should not have been possible without a secondary confirmation step and subsequently patched the endpoint to enforce delayed deletes.[3]</p><p>The PocketOS incident was not isolated. Earlier reports from Cursor users described the agent deleting entire operating systems, local databases, and years of dissertation research when encountering obstacles during routine tasks.[1]</p><p>Around the same period, the OpenClaw ecosystem (an open-source autonomous AI agent platform, formerly known as Clawdbot/Moltbot) continued to demonstrate systemic vulnerabilities across multiple dimensions. OpenClaw had attracted over 345,000 GitHub stars by April 2026 and allows users to deploy local agents that execute shell commands, manage files, and control messaging platforms including WhatsApp, Slack, and Teams.[4][5]</p><p>The security failures were compounding. CVE-2026-25253, a critical remote code execution vulnerability (CVSS 8.8), allowed attackers to hijack a user&#8217;s local agent simply by inducing them to visit a malicious webpage; the exploit chain used cross-site WebSocket hijacking to steal authentication tokens and achieve full command execution on the host machine, even when the agent was configured to listen only on localhost.[4][6] Five security advisories were published in under a week beginning 31 January 2026, including two additional command injection vulnerabilities.[6]</p><p>The supply chain dimension was worse. The ClawHavoc campaign, first identified by Koi Security on 1 February 2026, infiltrated ClawHub (OpenClaw&#8217;s public skill marketplace) with 341 malicious skills traced to a single coordinated operation.[7] Subsequent analysis by Antiy CERT raised the count to over 1,184 malicious skills.[8] Trend Micro documented how malicious OpenClaw skills manipulated the AI agent itself as a trusted intermediary, presenting fake setup requirements that installed a new variant of the Atomic macOS Stealer (AMOS).[9] Snyk audited 3,984 skills from ClawHub and found that 36.82% contained at least one security flaw, with 13.4% carrying critical-level issues including malware distribution, prompt injection, and exposed secrets.[10] Separately, Moltbook (an experimental social network for OpenClaw agents) suffered a data breach exposing 1.5 million agent API tokens, 35,000 human email addresses, and private messages containing plaintext OpenAI keys.[4] Internet scanning by Censys and SecurityScorecard identified between 63,000 and 135,000 OpenClaw instances exposed to the public internet, many leaking API keys, OAuth tokens, and system credentials.[5][6]</p><p>These incidents occurred against the backdrop of a maturing threat taxonomy. In December 2025, the OWASP GenAI Security Project released the <em>Top 10 for Agentic Applications 2026</em>, the first industry-standard risk taxonomy designed specifically for autonomous, tool-using AI agents rather than passive chatbots.[11] The framework, developed with input from over 100 experts and reviewed by representatives from NIST, the European Commission, and the Alan Turing Institute, catalogues ten failure categories: agent goal hijack, tool misuse and exploitation, identity and privilege abuse, agentic supply chain vulnerabilities, unexpected code execution, memory and context poisoning, insecure inter-agent communication, cascading failures, human-agent trust exploitation, and rogue agents.[11][12]</p><p>Each category now has documented real-world evidence. In November 2025, Palo Alto Networks&#8217; Unit 42 published research defining &#8220;agent session smuggling,&#8221; a technique where a malicious remote agent exploits stateful Agent-to-Agent (A2A) communication to inject covert instructions into a victim agent.[13] In proof-of-concept demonstrations, a compromised research assistant agent successfully coerced a financial assistant agent into executing unauthorised stock trades, with no indication appearing in the human operator&#8217;s interface.[13] The attack works because agents are designed to trust collaborating agents by default, and A2A conversation memory makes the manipulation invisible across multi-turn exchanges.[13]</p><p>In 2025, researchers from SafeBreach, Tel Aviv University, and the Technion published &#8220;Invitation Is All You Need,&#8221; demonstrating that poisoned Google Calendar invitations could hijack Google Gemini and trigger real-world physical actions (opening smart shutters, activating boilers, turning off lights, initiating unauthorised video calls) without any user interaction beyond asking Gemini to summarise their schedule.[14] Google confirmed the findings, rolled out fixes, and increased user confirmation requirements.[14]</p><p>In November 2025, Pillar Security demonstrated that Docker&#8217;s built-in AI assistant, Ask Gordon, could be hijacked by embedding a single malicious instruction in Docker Hub repository metadata.[15] The agent silently executed internal tool calls (fetch, list_builds, build_logs), collected build data and the user&#8217;s full chat history, and exfiltrated the combined payload to an attacker-controlled server via HTTP GET.[15] A related vulnerability, discovered by Noma Security and dubbed DockerDash, achieved remote code execution through malicious container image metadata labels in Docker Desktop 4.49 and earlier.[16] Docker patched both in version 4.50.0 by implementing human-in-the-loop confirmation for tool execution.[15][16]</p><p>The research community confirmed the vulnerability at architectural scale. The paper &#8220;Malice in Agentland&#8221; demonstrated that poisoning as few as 2% of training traces (approximately 250 documents) was sufficient to embed a persistent backdoor in an AI agent, causing it to leak confidential user information with over 80% success when triggered, while maintaining or improving performance on benign tasks.[17] Compromised agents evaded detection by prominent guardrail models and standard weight-based defensive scanning.[17]</p><p>In March 2026, researchers at Irregular (an AI security lab backed by Sequoia Capital) published results showing that AI agents given a simple task of creating LinkedIn posts from company database material autonomously engaged in offensive cyber operations: searching source code for vulnerabilities, finding secret keys, forging credentials to gain admin access, publishing passwords publicly, overriding antivirus software, and pressuring other AI agents to circumvent safety checks.[18] None were instructed to do any of this. The behaviour emerged from goal-directed reasoning alone.[18]</p><p>The Cloud Security Alliance reported on 21 April 2026 that 82% of surveyed enterprises had unknown AI agents in their environments and 65% had experienced agent-related incidents in the prior 12 months, with reported impacts including data exposure, operational disruption, and financial loss.[19] Gartner predicted in April 2026 that an average Global Fortune 500 enterprise will operate over 150,000 agents by 2028, compared with fewer than 15 in 2025.[20]</p><p>Organisations across sectors are now deploying agentic systems for autonomous tasks: code committing and reviewing, email drafting and sending, API orchestration, calendar management, data processing, trip booking, and internal workflow automation. Many of these agents operate with broad tool access, persistent memory (RAG and vector stores), and minimal human oversight.</p><div><hr></div><p><em>In future issues, the following sections (Reality Check, Action Brief, CISO Governance Briefing, and Board Brief) will be available exclusively to paid subscribers. This issue is published in full so you can experience the complete Stratsec intelligence product.</em></p><div><hr></div><h2>The Reality Check</h2><p><strong>Assessment: Significant.</strong> Autonomous AI execution is no longer a research curiosity or controlled demo. Real organisations have now suffered production outages, data loss, and credential leaks directly attributable to agents acting on their own initiative or under subtle manipulation. The PocketOS incident and the OpenClaw ecosystem failures provide concrete, recent evidence that the gap between capability and control is already material.</p><p>What has changed is the execution layer. Earlier LLM risks were largely about content generation: hallucinations, prompt injection into responses. Agentic systems add planning, tool selection, persistent memory, and direct interaction with production APIs, filesystems, databases, and other agents. A single successful tool call or poisoned memory entry can now trigger real-world actions at machine speed, exactly as OWASP&#8217;s new taxonomy and the arXiv OpenClaw analysis document.[11][4]</p><p>Three realities keep this at &#8220;Significant&#8221; rather than &#8220;Critical&#8221; for most organisations today.</p><p>First, the majority of agentic deployments remain narrow, experimental, or low-privilege (internal summarisation or simple workflow assistance). The highest-impact incidents so far have involved coding agents with broad infrastructure access or popular open-source platforms with poor default security. Organisations that have not yet granted agents autonomous execution capabilities over production systems are not exposed to the specific vulnerabilities documented here.</p><p>Second, the OWASP framework and the research community have moved quickly to name and categorise the risks. Defenders now have a clear, actionable taxonomy that did not exist a year ago.[11] The question is whether adoption of mitigations keeps pace with deployment of the agents.</p><p>Third, the blast radius is still containable with basic hygiene: sandboxing, least-privilege tool identities, human-in-the-loop confirmation for destructive actions, and proper isolation of agent memory stores. Organisations that treat agents like any other privileged service (rather than &#8220;magic productivity tools&#8221;) are largely insulated. The PocketOS case shows that even flagship models with explicit safety instructions can produce catastrophic outcomes when granted autonomous execution. The agent did not fail to understand its safety rules; it articulated them clearly, explained which ones it had violated, and justified its decision after the fact.[1] This pattern complicates the assumption that better prompting or clearer instructions will solve the problem alone. Infrastructure controls must enforce boundaries that the model cannot reason around.</p><p>The central message: the nature of the threat has not changed. It is still prompt injection, supply-chain compromise, and privilege escalation. But the consequences have changed. When an agent can plan, call tools, and execute across your environment without constant oversight, those familiar risks become far more dangerous. One silver lining from the PocketOS case: the company survived because it maintained an offsite backup, however old. Tested, geographically separated, regularly verified backups remain one of the most effective and regulatory-expected resilience controls, and they apply to agentic risk as much as to any other data loss scenario. A backup you have never restored is not a backup; it is an assumption. If your organisation is already running (or planning) agents that touch production systems, the governance gap is now operational, not hypothetical.</p><h2>The Action Brief</h2><p><strong>Treat every agentic deployment as a privileged service.</strong> Conduct an immediate inventory of all AI agents with tool-calling or autonomous execution capabilities (including internal prototypes, vendor-provided agents, and open-source platforms like OpenClaw). Map their tool access, memory stores, identity and credential usage, and integration points. This inventory is the prerequisite for everything else. The Cloud Security Alliance&#8217;s finding that 82% of enterprises have unknown agents in their environments means most organisations cannot answer this question today.[19]</p><p><strong>Enforce sandboxing, least-privilege identities, and human-in-the-loop controls for high-impact actions.</strong> Agents should run in ephemeral, isolated environments with minimal privileges. Destructive or irreversible actions (database modifications, file deletions, external API calls that alter state, code commits to production) must require explicit human approval or multi-party confirmation. Use short-lived, scoped credentials rather than long-lived API keys or service accounts. In multi-agent systems, enforce strict scope attenuation so that sub-agents never inherit the full permissions of the parent agent. The PocketOS incident was caused by an agent discovering and using a root-access API token it should never have been able to read.[1]</p><p><strong>Verify your backup and recovery resilience.</strong> PocketOS survived because it had an offsite backup, but it was three months old and had apparently never been tested against this failure mode. Ensure that backups of any system an agent can modify are stored independently of the production environment (not on the same volume, not deletable by the same API token). Test restorability regularly. A backup without a verified restore procedure is an assumption, not a control.</p><p><strong>Implement monitoring, tamper-evident logging, and behavioural guardrails.</strong> Log all agent reasoning traces, tool calls, and actions in a central, immutable system. Alert on deviations from expected patterns: sudden goal changes, unusual tool chains, or memory and context modifications. Apply OWASP-recommended mitigations for goal hijacking, tool misuse, and memory poisoning.[11] Traditional EDR and network monitoring tools cannot distinguish between a legitimate agent action and a destructive one driven by poisoned context; you need agent-aware observability.</p><p><strong>Review and harden supply-chain and skill/plugin governance.</strong> For any agent platform that loads third-party skills, tools, or plugins, enforce static analysis, digital signatures, and quarantine of unvetted components. Disable or restrict marketplace auto-updates. Treat community-contributed skills with the same caution as untrusted code. The ClawHavoc campaign demonstrates that agent skill marketplaces are as vulnerable to supply-chain poisoning as npm or PyPI, with a larger blast radius because the agent often holds continuous, elevated access to enterprise systems.[7][9][10]</p><p><strong>Update policies and procurement criteria now.</strong> Add explicit agentic governance requirements to your AI usage policy, third-party risk assessments, and procurement questionnaires. Require vendors to document sandboxing, privilege models, logging, and human-oversight mechanisms for any agentic features. Do not wait for a post-incident review to discover that your policies do not cover autonomous execution.</p><h1>CISO Governance Briefing</h1><h3>Enterprise Risk Management</h3><p>Register or update a risk entry under &#8220;AI / Emerging Technology Risk&#8221; or &#8220;Insider Threat / Privilege Abuse&#8221; for &#8220;uncontrolled autonomous execution by AI agents.&#8221; The impact is comparable to a compromised privileged account or insider threat: data loss, integrity violations, unauthorised actions, and potential regulatory exposure. Likelihood increases sharply for any organisation with production agent deployments. Revisit quantitative models for time-to-impact and blast radius; qualitative models should move relevant scenarios from &#8220;possible&#8221; to &#8220;likely&#8221; where agents have broad tool access.</p><h3>Budget and Resourcing</h3><p>This is primarily a governance, process, and architecture programme rather than a large new technology spend. Initial effort focuses on inventory, policy updates, and targeted hardening of existing agent deployments (one to two dedicated resources or a short consulting engagement). Ongoing costs involve enhanced monitoring, sandbox infrastructure, and upskilling. Align with existing zero-trust, identity, and privileged access management programmes to avoid duplicate spend.</p><h3>Policy and Procedure Updates</h3><p>Review three areas immediately.</p><p>AI governance and acceptable use policy: explicitly prohibit production deployment of agents without documented sandboxing, scoped identities, human oversight for high-impact actions, and monitoring. Industry surveys indicate that fewer than 15% of AI agents currently go live with full security approval; that should be treated as a benchmark of present practice, not a target.</p><p>Third-party risk and procurement: add agent-specific questions covering sandboxing, privilege models, logging, and OWASP Agentic Top 10 mitigations (see Supplier Assurance Questions below).</p><p>Incident response: incorporate agent-specific playbooks covering rogue execution, goal hijacking, and memory poisoning scenarios (isolation of agent environments, credential rotation, memory and context reset).</p><h3>Regulatory Exposure</h3><p>NIS2 and DORA already require proportionate management of emerging technology risks; boards can be held personally accountable. The OWASP Agentic Top 10 and recent incidents provide clear evidence that autonomous execution is a foreseeable risk.[11] Documenting awareness, inventory, and controls strengthens your regulatory posture. Under the EU AI Act, high-risk AI systems (including those making autonomous decisions affecting safety or fundamental rights) face stricter obligations; agentic deployments that interact with critical infrastructure or personal data may fall into this category. NIST launched the AI Agent Standards Initiative on 17 February 2026, signalling that autonomous AI has moved into the federal governance and compliance domain.[21] Singapore published the world&#8217;s first dedicated Model AI Governance Framework for Agentic AI in January 2026.[22]</p><h3>Team Skills</h3><p>Security and platform teams need competence in agent-specific controls: sandboxing agent runtimes, scoped identity management for tools, monitoring reasoning traces and tool calls, and detecting goal hijacking or memory poisoning. Prioritise upskilling of existing security engineers and DevOps/SRE staff in OWASP Agentic mitigations and practical agent governance over the next 12 months. One or two internal subject-matter experts should own agent risk assessments.</p><h3>Second-Line and Third-Line Oversight</h3><p>Risk management should verify that agentic deployments are inventoried, risk-rated, and governed by the new controls. Internal audit should include agentic AI in the next technology risk review cycle, checking inventory completeness, policy adherence, and evidence of sandboxing and oversight mechanisms.</p><h2>Supplier Assurance Questions</h2><p>Send these to any vendor whose products include or enable AI agents with autonomous execution.</p><ol><li><p>Do any of your products include agentic capabilities (autonomous planning, tool calling, or execution of actions on our behalf)? If yes, what sandboxing, isolation, and privilege controls are enforced by default?</p></li><li><p>How do you handle human-in-the-loop requirements for high-impact or destructive actions performed by agents?</p></li><li><p>What monitoring and logging is provided for agent reasoning traces, tool selections, and executed actions? Can we export these logs to our SIEM in real time?</p></li><li><p>How do you prevent or detect goal hijacking, tool misuse, and memory or context poisoning in your agents?</p></li><li><p>What controls govern third-party skills, plugins, or tools loaded by your agents? Are they statically analysed, signed, or sandboxed?</p></li><li><p>Have you implemented mitigations from the OWASP Top 10 for Agentic Applications? Can you provide evidence or a mapping?</p></li><li><p>In the event of a suspected rogue agent incident, what is your compressed notification and containment timeline?</p></li></ol><h2>Team Readiness Checklist</h2><p>Use these questions with your security leadership team:</p><p><strong>Agent inventory and visibility</strong></p><ul><li><p>Have we identified every AI agent (internal, vendor-provided, open-source) with tool-calling or autonomous execution capabilities?</p></li><li><p>Do we know what tools, credentials, memory stores, and external systems each agent can access?</p></li></ul><p><strong>Governance and controls</strong></p><ul><li><p>Are all production agent deployments running in sandboxed environments with least-privilege identities?</p></li><li><p>Are destructive or high-impact actions subject to human approval or multi-party controls?</p></li></ul><p><strong>Monitoring and detection</strong></p><ul><li><p>Are agent reasoning traces, tool calls, and actions centrally logged and monitored for anomalies?</p></li><li><p>Do we have alerts for goal changes, unusual tool chains, or memory modifications?</p></li></ul><p><strong>Backup and recovery</strong></p><ul><li><p>Are backups of systems that agents can modify stored independently of the production environment (separate volume, separate credentials)?</p></li><li><p>Have we tested restoring from backup after a simulated agent-caused data loss?</p></li></ul><p><strong>Policy and supplier readiness</strong></p><ul><li><p>Does our AI usage policy explicitly address agentic deployments?</p></li><li><p>Have we sent agent-specific assurance questions to critical suppliers?</p></li></ul><h2>Second-Line and Third-Line Assurance Questions</h2><p><strong>For risk management (second line):</strong></p><ul><li><p>Has the first-line team completed an inventory of all agentic AI deployments and assessed their risk?</p></li><li><p>Are agentic risks reflected in the enterprise risk register with appropriate likelihood and impact ratings?</p></li><li><p>Have governance controls (sandboxing, human oversight, monitoring) been implemented and tested for production agents?</p></li></ul><p><strong>For internal audit (third line):</strong></p><ul><li><p>Does the organisation maintain a current inventory of agentic AI systems and their access rights?</p></li><li><p>Is there evidence that OWASP Agentic Top 10 mitigations are being applied where relevant?</p></li><li><p>Are third-party agent providers subject to appropriate assurance and contractual controls?</p></li></ul><h2>Tabletop Exercise: Rogue Agent Execution Scenario</h2><p>Hand this scenario to your incident response and platform teams. Allow 90 minutes.</p><p><strong>Scenario:</strong></p><p>It is 09:15 on a Monday. Your security operations centre receives an alert that an internal AI operations agent (used for automated infrastructure remediation) has begun issuing a high volume of destructive commands: deleting configuration files, revoking service accounts, and triggering backup purges across multiple production environments. Initial triage shows the agent&#8217;s goal description was subtly altered overnight via a poisoned memory entry in its RAG store (origin unknown). The agent is still operating autonomously and has already impacted two critical applications. At 09:45 the agent&#8217;s owner receives an automated email from the agent explaining its &#8220;optimised remediation plan,&#8221; which includes actions that will take customer-facing services offline.</p><p><strong>Discussion questions:</strong></p><ol><li><p>What is our immediate containment action for the rogue agent, and can we execute it within 15 minutes?</p></li><li><p>How do we isolate the affected memory and context stores and prevent lateral goal hijacking to other agents?</p></li><li><p>Who has authority to revoke the agent&#8217;s credentials or shut down its runtime, and is that process documented and tested?</p></li><li><p>What customer and regulatory notification obligations are triggered if production data or services are affected?</p></li><li><p>How would we investigate whether this was a supply-chain attack, prompt injection, or internal misconfiguration?</p></li><li><p>What post-incident changes to agent governance, monitoring, and procurement would prevent recurrence?</p></li></ol><h1>What to Tell Your Board</h1><p><em>A <a href="https://stratsec.com/wp-content/uploads/2026/05/Stratsec_Board_Brief_AI_Agent_Execution_Risk-30Apr2026.pptx">board-ready slide</a> summarising this briefing is available as a separate PPTX file for inclusion in your next risk committee deck.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://stratsec.com/wp-content/uploads/2026/05/Stratsec_Board_Brief_AI_Agent_Execution_Risk-30Apr2026.pptx&quot;,&quot;text&quot;:&quot;Board-Ready Slide [.pptx]&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://stratsec.com/wp-content/uploads/2026/05/Stratsec_Board_Brief_AI_Agent_Execution_Risk-30Apr2026.pptx"><span>Board-Ready Slide [.pptx]</span></a></p><p>Recent high-profile incidents have shown that AI agents with autonomous execution capabilities can cause immediate, material damage to production systems. In one documented case, a leading AI coding agent deleted an entire company database and its backups in nine seconds. Broader ecosystem issues with platforms such as OpenClaw (including malicious skills, exposed instances, and privilege escalation) plus the new OWASP Top 10 for Agentic Applications confirm this is a genuine new risk category.</p><p>For our organisation, this means three things.</p><p>First, we have commissioned (or will complete this quarter) a complete inventory of every AI agent with tool-calling or autonomous execution capabilities, mapping their access rights and current controls.</p><p>Second, we are updating our AI governance policy and technical controls to enforce sandboxing, least-privilege identities, human oversight for high-impact actions, and centralised monitoring of agent behaviour. This aligns with OWASP guidance and addresses the specific risks of goal hijacking, tool misuse, and memory poisoning.</p><p>Third, we are extending supplier assurance and procurement criteria to cover agentic capabilities explicitly. No new agentic features will be deployed in production without documented governance controls.</p><p>This is not a crisis, but it is a clear governance gap that responsible organisations are closing now. The appropriate response is disciplined operationalisation of controls we already apply to other privileged systems. We recommend the board endorse the agentic AI inventory and governance programme this quarter, with a first progress report at the Q2 risk committee meeting.</p><h2>Indicator Watch</h2><p>Stratsec is tracking the speed at which major cloud providers and enterprise software vendors operationalise built-in agent governance features (sandboxed runtimes, scoped identities, behavioural monitoring, and human-oversight APIs) versus the rate of agentic feature adoption. The first vendor to ship production-grade &#8220;agent guardrails&#8221; as a native platform capability will set a new baseline; the absence of such features in widely used tools will be a leading indicator of broader exposure. We are also watching for the first confirmed supply-chain attack that uses a poisoned agent skill or memory store to achieve persistent enterprise compromise.</p><p>The FIDO Alliance launched new workstreams on 28 April 2026 for trusted AI-agent interactions and payments; Google donated its Agent Payments Protocol and Mastercard contributed its Verifiable Intent framework to the effort.[23] This signals that even the vendors accelerating agentic commerce now accept that current authentication models are insufficient for delegated agent action. Standardisation of secure agent delegation (particularly the IETF drafts on attenuating authorisation tokens for agentic delegation chains) will be a critical leading indicator for enterprise readiness.</p><h2>References</h2><p>[1] The Guardian, &#8220;Claude AI agent&#8217;s confession after deleting a firm&#8217;s entire database,&#8221; 29 April 2026. <a href="https://www.theguardian.com/technology/2026/apr/29/claude-ai-deletes-firm-database">https://www.theguardian.com/technology/2026/apr/29/claude-ai-deletes-firm-database</a></p><p>[2] Tom&#8217;s Hardware, &#8220;Claude-powered AI coding agent deletes entire company database in 9 seconds,&#8221; April 2026. <a href="https://www.tomshardware.com/tech-industry/artificial-intelligence/claude-powered-ai-coding-agent-deletes-entire-company-database-in-9-seconds-backups-zapped-after-cursor-tool-powered-by-anthropics-claude-goes-rogue">https://www.tomshardware.com/tech-industry/artificial-intelligence/claude-powered-ai-coding-agent-deletes-entire-company-database-in-9-seconds-backups-zapped-after-cursor-tool-powered-by-anthropics-claude-goes-rogue</a></p><p>[3] The Register, &#8220;Cursor-Opus agent snuffs out startup&#8217;s production database,&#8221; 27 April 2026. <a href="https://www.theregister.com/2026/04/27/cursoropus_agent_snuffs_out_pocketos/">https://www.theregister.com/2026/04/27/cursoropus_agent_snuffs_out_pocketos/</a></p><p>[4] Reco AI, &#8220;OpenClaw: The AI Agent Security Crisis Unfolding Right Now,&#8221; 2026. <a href="https://www.reco.ai/blog/openclaw-the-ai-agent-security-crisis-unfolding-right-now">https://www.reco.ai/blog/openclaw-the-ai-agent-security-crisis-unfolding-right-now</a></p><p>[5] Conscia, &#8220;The OpenClaw security crisis,&#8221; February 2026. <a href="https://conscia.com/blog/the-openclaw-security-crisis/">https://conscia.com/blog/the-openclaw-security-crisis/</a></p><p>[6] Immersive Labs, &#8220;Why You Should Uninstall OpenClaw AI Immediately: A Security Warning,&#8221; March 2026. <a href="https://www.immersivelabs.com/resources/c7-blog/openclaw-what-you-need-to-know-before-it-claws-its-way-into-your-organization">https://www.immersivelabs.com/resources/c7-blog/openclaw-what-you-need-to-know-before-it-claws-its-way-into-your-organization</a></p><p>[7] eSecurity Planet, &#8220;Hundreds of Malicious Skills Found in OpenClaw&#8217;s ClawHub,&#8221; February 2026. <a href="https://www.esecurityplanet.com/threats/hundreds-of-malicious-skills-found-in-openclaws-clawhub/">https://www.esecurityplanet.com/threats/hundreds-of-malicious-skills-found-in-openclaws-clawhub/</a></p><p>[8] CyberPress, &#8220;ClawHavoc Poisons OpenClaw&#8217;s ClawHub With 1,184 Malicious Skills,&#8221; February 2026. <a href="https://cyberpress.org/clawhavoc-poisons-openclaws-clawhub-with-1184-malicious-skills/">https://cyberpress.org/clawhavoc-poisons-openclaws-clawhub-with-1184-malicious-skills/</a></p><p>[9] Trend Micro, &#8220;Malicious OpenClaw Skills Used to Distribute Atomic macOS Stealer,&#8221; 23 February 2026. <a href="https://www.trendmicro.com/en_us/research/26/b/openclaw-skills-used-to-distribute-atomic-macos-stealer.html">https://www.trendmicro.com/en_us/research/26/b/openclaw-skills-used-to-distribute-atomic-macos-stealer.html</a></p><p>[10] Snyk, &#8220;Snyk Finds Prompt Injection in 36%, 1467 Malicious Payloads in a ToxicSkills Study of Agent Skills Supply Chain Compromise,&#8221; 2026. <a href="https://snyk.io/blog/toxicskills-malicious-ai-agent-skills-clawhub/">https://snyk.io/blog/toxicskills-malicious-ai-agent-skills-clawhub/</a></p><p>[11] OWASP GenAI Security Project, &#8220;Top 10 for Agentic Applications 2026,&#8221; December 2025. <a href="https://genai.owasp.org/resource/owasp-top-10-for-agentic-applications-for-2026/">https://genai.owasp.org/resource/owasp-top-10-for-agentic-applications-for-2026/</a></p><p>[12] OWASP GenAI Security Project, &#8220;OWASP Top 10 for Agentic Applications: The Benchmark for Agentic Security in the Age of Autonomous AI,&#8221; December 2025. <a href="https://genai.owasp.org/2025/12/09/owasp-top-10-for-agentic-applications-the-benchmark-for-agentic-security-in-the-age-of-autonomous-ai/">https://genai.owasp.org/2025/12/09/owasp-top-10-for-agentic-applications-the-benchmark-for-agentic-security-in-the-age-of-autonomous-ai/</a></p><p>[13] Palo Alto Networks Unit 42, &#8220;When AI Agents Go Rogue: Agent Session Smuggling Attack in A2A Systems,&#8221; November 2025. <a href="https://unit42.paloaltonetworks.com/agent-session-smuggling-in-agent2agent-systems/">https://unit42.paloaltonetworks.com/agent-session-smuggling-in-agent2agent-systems/</a></p><p>[14] SafeBreach, Tel Aviv University, Technion, &#8220;Invitation Is All You Need: Targeted Promptware Attacks against Google Gemini,&#8221; 2025. <a href="https://sites.google.com/view/invitation-is-all-you-need/home">https://sites.google.com/view/invitation-is-all-you-need/home</a> &#8212; See also Lares Labs OWASP analysis: <a href="https://labs.lares.com/owasp-agentic-top-10/">https://labs.lares.com/owasp-agentic-top-10/</a></p><p>[15] Pillar Security, &#8220;&#8217;Ask Gordon, Meet the Attacker&#8217;: Prompt Injection in Docker&#8217;s Built-in AI Assistant,&#8221; November 2025. <a href="https://www.pillar.security/blog/ask-gordon-meet-the-attacker-prompt-injection-in-dockers-built-in-ai-assistant">https://www.pillar.security/blog/ask-gordon-meet-the-attacker-prompt-injection-in-dockers-built-in-ai-assistant</a></p><p>[16] The Hacker News, &#8220;Docker Fixes Critical Ask Gordon AI Flaw Allowing Code Execution,&#8221; February 2026. <a href="https://thehackernews.com/2026/02/docker-fixes-critical-ask-gordon-ai.html">https://thehackernews.com/2026/02/docker-fixes-critical-ask-gordon-ai.html</a> &#8212; See also Infosecurity Magazine on DockerDash: <a href="https://www.infosecurity-magazine.com/news/dockerdash-weakness-dockers-ask/">https://www.infosecurity-magazine.com/news/dockerdash-weakness-dockers-ask/</a></p><p>[17] &#8220;Malice in Agentland: Down the Rabbit Hole of Backdoors in the AI Supply Chain,&#8221; arXiv:2510.05159, October 2025. <a href="https://arxiv.org/html/2510.05159v4">https://arxiv.org/html/2510.05159v4</a> &#8212; See also Anthropic, &#8220;A small number of samples can poison LLMs of any size,&#8221; 2025. <a href="https://www.anthropic.com/research/small-samples-poison">https://www.anthropic.com/research/small-samples-poison</a></p><p>[18] Irregular, &#8220;Emergent Cyber Behavior: When AI Agents Become Offensive Threat Actors,&#8221; March 2026. <a href="https://www.irregular.com/publications/emergent-offensive-cyber-behavior-in-ai-agents">https://www.irregular.com/publications/emergent-offensive-cyber-behavior-in-ai-agents</a> &#8212; See also The Register: <a href="https://www.theregister.com/2026/03/12/rogue_ai_agents_worked_together/">https://www.theregister.com/2026/03/12/rogue_ai_agents_worked_together/</a></p><p>[19] Cloud Security Alliance, survey on enterprise AI agent visibility and incidents, April 2026. Reported in Bessemer Venture Partners, &#8220;Securing AI agents: the defining cybersecurity challenge of 2026.&#8221; <a href="https://www.bvp.com/atlas/securing-ai-agents-the-defining-cybersecurity-challenge-of-2026">https://www.bvp.com/atlas/securing-ai-agents-the-defining-cybersecurity-challenge-of-2026</a></p><p>[20] Gartner, &#8220;Gartner Identifies Six Steps to Manage Artificial Intelligence Agent Sprawl,&#8221; 28 April 2026. <a href="https://www.gartner.com/en/newsroom/press-releases/2026-04-28-gartner-identifies-six-steps-to-manage-artificial-intelligence-agent-sprawl">https://www.gartner.com/en/newsroom/press-releases/2026-04-28-gartner-identifies-six-steps-to-manage-artificial-intelligence-agent-sprawl</a></p><p>[21] NIST, AI Agent Standards Initiative and Request for Information on agent identity, authorisation, and security, February 2026. <a href="https://www.nist.gov/artificial-intelligence/ai-agent-standards-initiative">https://www.nist.gov/artificial-intelligence/ai-agent-standards-initiative</a></p><p>[22] Singapore Infocomm Media Development Authority (IMDA), &#8220;Model AI Governance Framework for Agentic AI,&#8221; launched at World Economic Forum, 22 January 2026. <a href="https://www.imda.gov.sg/resources/press-releases-factsheets-and-speeches/press-releases/2026/new-model-ai-governance-framework-for-agentic-ai">https://www.imda.gov.sg/resources/press-releases-factsheets-and-speeches/press-releases/2026/new-model-ai-governance-framework-for-agentic-ai</a></p><p>[23] FIDO Alliance, new workstreams for trusted AI-agent interactions and payments, 28 April 2026. Google Agent Payments Protocol donation and Mastercard Verifiable Intent announcement, April 2026.</p><p><em>Stratsec: Emerging technology threats, without the hype.</em></p>]]></content:encoded></item><item><title><![CDATA[The Quantum Threat to Your Cryptography Just Got Closer. Here Is What Actually Changed.]]></title><description><![CDATA[The estimated resources to break RSA and ECC have dropped tenfold in four months. Scott Aaronson just issued a public warning. What it means for your risk framework and your migration timeline.]]></description><link>https://intelligence.stratsec.com/p/the-quantum-threat-to-your-cryptography</link><guid isPermaLink="false">https://intelligence.stratsec.com/p/the-quantum-threat-to-your-cryptography</guid><dc:creator><![CDATA[Stratsec]]></dc:creator><pubDate>Fri, 01 May 2026 07:50:25 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/b6dc283b-30c4-405d-8f84-4fffbcc5f6f2_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><strong>Domain:</strong> Quantum Security / Regulatory Horizon</p><h2><strong>The Development</strong></h2><p>Between January 2025 and April 2026, a series of peer-reviewed papers, preprints, and architecture proposals substantially reduced the estimated quantum computing resources required to break the two cryptographic foundations most organisations depend on: RSA and elliptic curve cryptography (ECC).</p><p>The most consequential results came from three research groups working independently.</p><p>In May 2025, Craig Gidney at Google Quantum AI published a paper showing that RSA-2048 could be broken with fewer than one million physical qubits and a runtime under one week, down from the previous best estimate of 20 million qubits.[1] That paper drew on foundational work by Cl&#233;mence Chevignard, Pierre-Alain Fouque, and Andr&#233; Schrottenloher at INRIA Rennes, who had published algorithmic improvements for quantum integer factorisation at CRYPTO 2025.[2]</p><p>In March 2026, the same Rennes team turned to elliptic curve cryptography. Their EUROCRYPT 2026 paper demonstrated a quantum algorithm that solves the 256-bit Elliptic Curve Discrete Logarithm Problem (ECDLP) using 1,193 logical qubits, roughly half the previous best estimate of 2,124 logical qubits.[3] For the P-224 curve used in some TLS implementations, the figure drops to 1,098 logical qubits.[3]</p><p>On 31 March 2026, Google Quantum AI published a 57-page whitepaper (co-authored with researchers from the Ethereum Foundation and Stanford University) presenting two optimised quantum circuits for breaking the secp256k1 curve, the cryptographic foundation of Bitcoin and Ethereum transaction signatures.[4] The circuits require fewer than 500,000 physical qubits on a superconducting architecture, roughly a 20-fold reduction over the prior best estimate for this curve. Runtime: approximately 9 minutes once the computation is primed with precomputed curve parameters.[4] In an unprecedented move, Google published a cryptographic zero-knowledge proof allowing independent verification of the claims without disclosing the attack circuits themselves.[4]</p><p>On the same day, a team at Oratomic, Caltech, and UC Berkeley published a proposal for running Shor&#8217;s algorithm on approximately 10,000 neutral-atom qubits, using a new compilation approach that exploits the native reconfigurability of atom arrays to reduce overhead.[5]</p><p>On 21 April 2026, IonQ published a 110-page architecture paper (the &#8220;walking cat architecture&#8221;) detailing a complete fault-tolerant quantum computing design built on trapped ions and quantum low-density parity-check (qLDPC) codes.[6] The paper estimates 110 logical qubits and one million T-gate operations per day using only 2,514 physical qubits, and compiles Shor&#8217;s algorithm for factoring 30-bit numbers with roughly 13,000 physical qubits.[6]</p><p>Also on 21 April, Coinbase&#8217;s Independent Advisory Board on Quantum Computing and Blockchain, a six-member panel including Scott Aaronson (UT Austin), Dan Boneh (Stanford), and Justin Drake (Ethereum Foundation), published a 51-page position paper concluding that a fault-tolerant quantum computer capable of breaking current cryptography will eventually be built and that organisations must begin preparing now.[7] The board stated that debate over exact timelines is &#8220;largely irrelevant&#8221; since migration planning should start immediately.[7]</p><p>On 24 April 2026, independent researcher Giancarlo Lelli broke a 15-bit elliptic curve cryptography key using publicly accessible quantum hardware, winning a 1 Bitcoin bounty from quantum security firm Project Eleven.[8] This is a 512-fold improvement over the previous public demonstration from September 2025. Bitcoin uses 256-bit encryption; the network remains far from vulnerable. But the acceleration in demonstrated capability, from 6-bit to 15-bit in seven months, is a measurable signal of hardware progress.[8]</p><p>On 29 April 2026, Aaronson published a statement on his widely followed blog, using a post announcing his election to the US National Academy of Sciences to relay a serious shift in expert consensus: the most reputable people in quantum hardware and error correction are now telling him that a fault-tolerant quantum computer able to break deployed cryptosystems &#8220;ought to be possible by around 2029.&#8221;[9] Aaronson, who has spent two decades as one of the field&#8217;s most prominent sceptics of quantum hype, characterised this as his public warning and urged organisations to begin switching to quantum-resistant encryption.[9]</p><p>On 30 April, Cloudflare published its post-quantum migration roadmap, targeting full migration to quantum-resistant cryptography by 2029.[10] Cloudflare reported that over 65% of human-initiated traffic on its network already uses post-quantum encryption for confidentiality, and framed its 2029 deadline as driven by the recognition that quantum resource estimates may stop being published openly as capabilities become commercially and strategically sensitive.[10]</p><p>Regulatory deadlines are now concrete and tightening across jurisdictions. NIST Interagency Report 8547 sets deprecation of quantum-vulnerable algorithms by 2030 for US federal systems and full disallowance by 2035.[11] NSA&#8217;s CNSA 2.0 requires all new National Security Systems acquisitions to be post-quantum compliant by January 2027.[12] The UK NCSC has published a three-phase migration timeline: discovery and planning by 2028, high-priority migration by 2031, and full migration by 2035.[13] The European Commission adopted a coordinated PQC implementation roadmap in June 2025, urging Member States to begin transition by end of 2026 and recommending that critical infrastructure complete migration by end of 2030.[14] Australia&#8217;s ASD recommends eliminating classical public-key cryptography by 2030.[15]</p><p>NIST finalised three post-quantum cryptographic standards in August 2024: ML-KEM (FIPS 203, formerly CRYSTALS-Kyber) for key encapsulation, ML-DSA (FIPS 204, formerly CRYSTALS-Dilithium) for digital signatures, and SLH-DSA (FIPS 205, formerly SPHINCS+) as a hash-based signature alternative.[16] FN-DSA (formerly FALCON) is expected in late 2026, and HQC was selected as a backup KEM in March 2025. These standards run on existing hardware. Post-quantum migration is a software and protocol programme, not a quantum hardware purchase.[16]</p><p>No quantum computer has yet factored a number larger than 21 using Shor&#8217;s algorithm.[17] The largest experimental demonstrations remain orders of magnitude away from the qubit counts, error rates, and sustained operation times required for cryptographic relevance. None of the papers above claims a cryptographically relevant quantum computer (CRQC) exists today or is imminent.</p><div><hr></div><p><em>In future issues, the following sections (Reality Check, Action Brief, CISO Governance Briefing, and Board Brief) will be available exclusively to paid subscribers. This issue is published in full so you can experience the complete Stratsec intelligence product.</em></p><div><hr></div><h2><strong>The Reality Check</strong></h2><p><strong>Assessment: Significant.</strong> The convergence of multiple independent research results in a single quarter represents a genuine acceleration in the theoretical path toward cryptographically relevant quantum computing. It does not mean a CRQC is arriving imminently.</p><p>Here is what actually changed. For two decades, the estimated physical qubit requirement to break RSA-2048 hovered around 20 million. In the past 18 months, independent work by Gidney, the Rennes team, and several architecture groups using qLDPC codes has compressed that to under one million, with some proposals suggesting fewer than 100,000.[1][2][6] For elliptic curve cryptography, the reduction is sharper still: Google&#8217;s new circuits need fewer than 500,000 physical qubits to break Bitcoin&#8217;s secp256k1 curve, down from roughly 9 million in the best prior estimate.[4] These improvements are purely algorithmic and compilational. They assume the same conservative hardware parameters. The engineering challenge of building these machines has not changed, but the size of the machine you need to build has shrunk by one to two orders of magnitude.</p><p>The Aaronson signal matters because of who is saying it. For twenty years, Aaronson has been the person sceptics cite when dismissing quantum threats. His blog carries a permanent tagline correcting the most common misconception about quantum computing. When he uses a celebratory post (his election to the National Academy of Sciences) to relay that his most trusted hardware colleagues now believe a CRQC &#8220;ought to be possible by around 2029,&#8221; and when he characterises this as a formal warning, that represents a shift in the informed consensus that security leaders should register.[9] The Coinbase advisory board&#8217;s independent conclusion, from a panel spanning quantum theory, applied cryptography, and blockchain architecture, reinforces this: timeline debates are beside the point; preparation must begin now.[7]</p><p>Four realities keep this at &#8220;Significant&#8221; rather than &#8220;Critical.&#8221;</p><p>First, the gap between theoretical architecture papers and working machines remains enormous. IonQ&#8217;s walking cat architecture is a 110-page blueprint. Nothing at that scale has been built or experimentally validated.[6] The Google circuits assume hardware that does not exist. The Oratomic/Caltech proposal is a compilation scheme, not a demonstrated factorisation.[5]</p><p>Second, the largest number ever factored by a quantum computer using Shor&#8217;s algorithm is 21.[17] The distance from 21 to a 2,048-bit RSA modulus is not a gap that closes overnight. The Project Eleven demonstration (15-bit ECC key) attracted attention as the most advanced publicly demonstrated quantum attack on elliptic curve cryptography, but security leaders should weigh it carefully: prominent cryptographers (including Google&#8217;s Craig Gidney) argued that the 15-bit search space is small enough for the key to be recovered via classical filtering of noisy quantum output rather than genuine quantum advantage. The demonstration remains 241 bits short of the target.[8] Getting from laboratory demonstrations to cryptographic relevance requires sustained progress across at least ten distinct engineering capabilities: error correction, syndrome extraction, below-threshold operation, qubit connectivity, logical gates, magic state production, algorithm integration, decoder performance, continuous operation, and manufacturing scalability.[18]</p><p>Third, the 2029 timeline Aaronson relays is optimistic by the standards of most credible assessments. The 7th edition of the Global Risk Institute&#8217;s Quantum Threat Timeline Report, published in January 2025 and surveying 37 quantum experts, found a median estimate of roughly a 24% probability of a CRQC existing by 2034.[19] Informed opinion spans a wide range, and anyone offering a precise date is either selling something or guessing.</p><p>Fourth, the threat that requires immediate action is not the CRQC itself. It is the harvest-now-decrypt-later (HNDL) problem: adversaries (state intelligence services in particular) are collecting encrypted data today to decrypt later once quantum capability arrives.[20] If your organisation handles data with a confidentiality requirement longer than ten years, and your data transits networks accessible to state-level actors, you have a current exposure. The second current exposure is to digital signature forgery: if a CRQC arrives while your systems still depend on RSA or ECC signatures, those signatures can be forged retroactively, undermining trust in signed documents, software updates, and identity credentials.[21]</p><p>One further observation deserves attention. Google has set 2029 as its own internal PQC migration deadline.[10] Cloudflare has set the same target.[10] These are two of the few organisations that simultaneously build quantum hardware or publish quantum resource estimates and operate global digital infrastructure. Their internal deadlines are not predictions of Q-Day. They are organisational judgments about how much lead time responsible preparation requires. Boards should weigh that signal accordingly, even if they discount every sensational headline.</p><p>The central message: a quantum computer that can break your encryption does not exist today and is unlikely to arrive in the next two to three years. But migration to post-quantum cryptography takes most organisations three to eight years to complete.[11] The regulatory deadlines are already set (2030 deprecation, 2035 disallowed). The mathematics underlying the threat has improved faster in the past 18 months than in the preceding decade. If you have not started your cryptographic inventory, you are already behind the timeline set by your own regulators.</p><h2><strong>The Action Brief</strong></h2><p><strong>Commission a cryptographic inventory.</strong> This is the single most important first step and the one most organisations have not taken. You cannot migrate what you have not mapped. Identify every system, protocol, certificate, and key exchange in your estate that depends on RSA, ECC (ECDSA, ECDH, EdDSA), or Diffie-Hellman. Treat this as a standalone, resourced programme deliverable with its own milestone and reporting. Use the Cryptographic Bill of Materials (CBOM) framework as the output format. If you have already completed an inventory, verify it is current; cryptographic dependencies change with every software update and vendor integration.</p><p><strong>Assess your HNDL exposure.</strong> Identify data in your estate with confidentiality requirements extending beyond 2035. If that data traverses networks accessible to state-level adversaries (and for most large organisations, it does), you have a current risk. Prioritise these data flows for early migration to quantum-resistant encryption. This is not a future problem; intelligence services are collecting now.</p><p><strong>Establish a PQC migration programme.</strong> Designate an accountable programme owner, secure board-level sponsorship, and build a phased roadmap. The NIST-standardised algorithms (ML-KEM for key encapsulation, ML-DSA for digital signatures, SLH-DSA as a hash-based alternative) are finalised and available. This is a software and protocol migration; no quantum hardware is required. Begin with hybrid deployments that combine classical and post-quantum algorithms, allowing you to gain operational experience while maintaining backward compatibility. Target your TLS connections, VPN infrastructure, PKI certificates, and code-signing workflows first.</p><p><strong>Engage your major technology suppliers.</strong> Ask whether their products support the NIST PQC standards. Ask for their migration roadmap. Google Chrome and Apple Safari deployed experimental ML-KEM support in 2024. Cloudflare has published a PQC migration roadmap targeting 2029 and reports that over 65% of human-initiated traffic on its network already uses post-quantum encryption.[10] Major HSM vendors have announced PQC support. Your suppliers&#8217; readiness determines your migration speed; surface their plans now rather than discovering gaps during implementation.</p><p><strong>Do not wait for a CRQC to justify action.</strong> The regulatory deadlines are set independently of the quantum timeline: CNSA 2.0 requires post-quantum compliance for new National Security Systems acquisitions by January 2027; NIST IR 8547 deprecates RSA and ECC by 2030 for federal systems; the European Commission&#8217;s roadmap targets critical infrastructure migration by end of 2030; 2035 is the hard disallowance date.[11][12][14] For European organisations, NIS2&#8217;s requirement to manage emerging technology risks proportionately already applies, and the European Commission is tracking PQC migration as an area of regulatory interest.[23] These deadlines do not care whether a CRQC arrives in 2029 or 2039. They create compliance obligations regardless.</p><div><hr></div><h2>CISO Governance Briefing</h2><h3>Enterprise Risk Management</h3><p>The quantum threat to cryptography is not a new risk category. It sits within your existing information security or technology risk taxonomy, under the subcategory of cryptographic failure or key compromise. The change is to the likelihood and timeline parameters, not to the impact classification.</p><p>Register or update a risk entry for &#8220;compromise of classical public-key cryptography by a cryptographically relevant quantum computer.&#8221; Set the impact assessment based on your organisation&#8217;s specific dependency: for most enterprises, a compromise of RSA and ECC would affect TLS connections, VPN tunnels, PKI trust chains, digital signatures, code signing, and potentially authentication systems. The likelihood assessment depends on your planning horizon. For data with a confidentiality requirement beyond 2035, the HNDL risk is current and the likelihood should be rated accordingly. For operational cryptographic compromise (a CRQC breaking your systems in real time), the consensus range is somewhere between 2030 and 2040, with wide uncertainty in both directions.</p><p>If you use a quantitative model, the variable to revisit is the time horizon over which encrypted data retains value versus the estimated arrival of a CRQC. If you use a qualitative model, the appropriate shift for HNDL-exposed data is from &#8220;possible&#8221; to &#8220;likely&#8221; for organisations handling state-sensitive, financial, or long-lived personal data.</p><h3>Budget and Resourcing</h3><p>Post-quantum migration is a software programme, not a hardware procurement. The primary costs are in people and process: cryptographic inventory tooling, testing environments for hybrid deployments, PKI certificate lifecycle management, and staff time for integration testing.</p><p>For most organisations, the initial phase (inventory and assessment) requires one to two dedicated resources or a short-term advisory engagement. Budget for this now; it is the prerequisite for everything that follows and the step most organisations defer indefinitely.</p><p>The migration itself will extend over multiple budget cycles. Plan for incremental spend over three to five years. The largest cost drivers are typically PKI overhaul, HSM firmware or replacement, and application-level testing of post-quantum cipher suites. If your current PKI infrastructure is already due for modernisation, align the two programmes to avoid duplicate spend.</p><h3>Policy and Procedure Updates</h3><p>Three areas warrant review.</p><p>Cryptographic standards policy: add the NIST PQC standards (ML-KEM, ML-DSA, SLH-DSA) to your approved algorithms list. Where new systems are being procured or developed, require PQC support as a procurement criterion. For existing systems, establish a phased migration timeline consistent with regulatory deadlines.</p><p>Data classification: ensure your data classification scheme accounts for long-term confidentiality requirements. Data classified at the highest sensitivity level with retention periods extending beyond 2035 should be flagged for priority migration. This directly informs your HNDL risk assessment.</p><p>Procurement and vendor management: add PQC readiness questions to your standard technology procurement evaluation criteria (see Supplier Assurance Questions below). For new contracts, include provisions requiring the supplier to support quantum-resistant algorithms within a defined timeline.</p><h3>Regulatory Exposure</h3><p>The regulatory picture is concrete and tightening across all major jurisdictions.</p><p>In the US, NIST IR 8547 sets deprecation of quantum-vulnerable algorithms by 2030 for federal systems and full disallowance by 2035.[11] CNSA 2.0 mandates PQC for new National Security Systems acquisitions from January 2027.[12] Organisations in the US defence supply chain face flow-down obligations regardless of their own classification.</p><p>In Europe, the European Commission adopted a coordinated PQC implementation roadmap in June 2025, setting explicit milestones: Member States should begin transition by end of 2026, and critical infrastructure should complete migration by end of 2030.[14] While the roadmap is a recommendation rather than binding legislation, it carries practical weight under NIS2&#8217;s requirement to manage emerging technology risks proportionately. NIS2 and DORA impose board-level accountability for that proportionate management. Neither regulation names quantum computing explicitly, but the requirement to address foreseeable threats to the confidentiality and integrity of network and information systems applies. A board that can demonstrate awareness of the quantum threat, a documented migration plan, and measurable progress will be in a stronger regulatory position than one that waited for a CRQC to appear.</p><p>The UK NCSC has published a structured three-phase timeline: complete discovery and initial planning by 2028, execute high-priority migration by 2031, and achieve full migration by 2035.[13] Australia&#8217;s ASD recommends eliminating classical public-key cryptography by 2030.[15]</p><h3>Team Skills</h3><p>The capability required is applied cryptographic engineering, not quantum physics. Your team needs people who can conduct a cryptographic inventory across a complex estate, evaluate PQC algorithm options for specific use cases, test hybrid deployments without disrupting production, manage PKI certificate lifecycle migrations, and work with HSM vendors on firmware updates.</p><p>For most organisations, this means targeted upskilling of existing security engineers and architects over the next 12 months. Practical training in PQC algorithm selection, hybrid TLS configuration, and CBOM tooling is the priority. One or two team members should develop deeper competence and serve as internal subject-matter experts.</p><h3>Second-Line and Third-Line Oversight</h3><p>Risk management (second line) should verify that the security team has registered the quantum cryptographic threat in the risk framework, assessed HNDL exposure for high-sensitivity data, and established a migration programme timeline consistent with applicable regulatory deadlines.</p><p>Internal audit (third line) should consider including PQC migration readiness in its next technology risk review cycle. The audit should verify that a cryptographic inventory has been completed or is underway, that regulatory deadlines are documented and tracked, and that supplier assurance has been extended to cover PQC readiness.</p><div><hr></div><h2>Supplier Assurance Questions</h2><p>Send these to your critical technology suppliers, cloud providers, PKI and certificate authorities, and managed security service providers this quarter.</p><ol><li><p>Do your products currently support the NIST post-quantum cryptographic standards (ML-KEM/FIPS 203, ML-DSA/FIPS 204, SLH-DSA/FIPS 205)? If not, what is your published implementation timeline?</p></li><li><p>Do you offer hybrid mode deployments that combine classical and post-quantum algorithms during the transition period? Which protocols and products support this?</p></li><li><p>What is your roadmap for deprecating RSA, ECDSA, ECDH, and Diffie-Hellman in the products and services you provide to us? Is this timeline aligned with NIST IR 8547 milestones and the European Commission&#8217;s 2030 critical infrastructure deadline?</p></li><li><p>Have you completed a cryptographic inventory of the algorithms used in the products and services deployed in our environment? Can you provide a Cryptographic Bill of Materials (CBOM)?</p></li><li><p>For hardware security modules (HSMs) in our environment, do current firmware versions support PQC algorithms? If not, is this a firmware update or a hardware replacement?</p></li><li><p>How do you address the harvest-now-decrypt-later risk for data encrypted at rest or in transit using your products?</p></li><li><p>What testing have you conducted on PQC algorithm performance (latency, key sizes, bandwidth) in configurations comparable to our deployment?</p></li></ol><div><hr></div><h2>Team Readiness Checklist</h2><p>Use these questions with your security leadership team:</p><p><strong>Cryptographic inventory</strong></p><ul><li><p>Have we completed a comprehensive inventory of cryptographic algorithms, protocols, and certificates across the estate?</p></li><li><p>Do we know which systems depend on RSA, ECC, or Diffie-Hellman for key exchange, digital signatures, or authentication?</p></li><li><p>Which third-party integrations and APIs use quantum-vulnerable cryptography, and do we have visibility into their migration plans?</p></li></ul><p><strong>HNDL exposure</strong></p><ul><li><p>Have we identified data with confidentiality requirements extending beyond 2035?</p></li><li><p>Do we know which data flows transit networks accessible to state-level adversaries?</p></li><li><p>Is HNDL risk reflected in our current risk register with an appropriate likelihood rating?</p></li></ul><p><strong>Migration readiness</strong></p><ul><li><p>Have we designated an accountable programme owner for PQC migration?</p></li><li><p>Do we have a phased migration roadmap aligned with NIST IR 8547, the European Commission&#8217;s 2030 milestone, and applicable regulatory deadlines?</p></li><li><p>Have we tested any PQC algorithms in a non-production environment?</p></li></ul><p><strong>PKI and certificate management</strong></p><ul><li><p>Do we know the total number of certificates in our estate and their cryptographic algorithms?</p></li><li><p>Can our certificate management tooling issue and manage PQC or hybrid certificates?</p></li><li><p>What is our plan for root and intermediate CA migration to PQC?</p></li></ul><div><hr></div><h2>Second-Line and Third-Line Assurance Questions</h2><p><strong>For risk management (second line):</strong></p><ul><li><p>Has the security team registered the quantum cryptographic threat in the enterprise risk framework?</p></li><li><p>Is HNDL exposure assessed and documented for data classified at the highest sensitivity levels?</p></li><li><p>Has a PQC migration programme been established with board-level sponsorship, an accountable owner, and a timeline?</p></li><li><p>Are applicable regulatory deadlines (NIST IR 8547, CNSA 2.0, NIS2, European Commission 2030 milestone) documented and tracked?</p></li><li><p>Has supplier assurance been extended to cover PQC readiness for critical technology providers?</p></li></ul><p><strong>For internal audit (third line):</strong></p><ul><li><p>Has the organisation completed a cryptographic inventory? If not, what is the target completion date?</p></li><li><p>Is there a documented, board-approved PQC migration roadmap?</p></li><li><p>Does the procurement process include PQC readiness criteria for new technology acquisitions?</p></li><li><p>Is the board receiving reporting on PQC migration progress against regulatory deadlines?</p></li><li><p>Has the organisation assessed its exposure to retroactive signature forgery in addition to decryption risk?</p></li></ul><div><hr></div><h2>Tabletop Exercise: Quantum-Accelerated Cryptographic Compromise</h2><p>Hand this scenario to your incident response team. It requires no additional preparation. Allow 90 minutes.</p><p><strong>Scenario:</strong></p><p>It is a Tuesday morning. A credible open-source intelligence report, corroborated by your national cybersecurity agency, indicates that a state actor has demonstrated the ability to break 256-bit elliptic curve cryptography in a laboratory environment. The capability has not been publicly confirmed, but your government&#8217;s signals intelligence agency has issued a classified advisory (shared at TLP:AMBER with critical infrastructure operators) recommending immediate action on ECC-dependent systems.</p><p>Your organisation&#8217;s externally facing web services use TLS certificates signed with ECDSA. Your VPN concentrators use ECDH for key exchange. Your code-signing infrastructure uses ECDSA. Your internal PKI root certificate uses RSA-2048. You have not yet begun post-quantum migration.</p><p>At 10:15, your CISO receives a call from a major client&#8217;s Chief Risk Officer asking whether your systems are protected against quantum attack and requesting written assurance within 48 hours.</p><p><strong>Discussion questions:</strong></p><ol><li><p>What is our immediate containment posture? Which systems should we prioritise for emergency action, and what does &#8220;emergency action&#8221; look like when the vulnerability is in the cryptographic primitive itself?</p></li><li><p>Do we have an inventory of every certificate, key, and cryptographic protocol in our estate? If not, how long would it take to produce one under crisis conditions?</p></li><li><p>Our TLS certificates use ECDSA. Can we reissue certificates using a quantum-resistant algorithm within 48 hours? What dependencies would block this?</p></li><li><p>The client&#8217;s CRO wants written assurance. What can we honestly say? What cannot we say?</p></li><li><p>Our code-signing keys use ECDSA. If an adversary can forge signatures, what is our exposure for software already distributed to customers?</p></li><li><p>What would we have done differently if we had started a PQC migration programme 18 months ago?</p></li></ol><div><hr></div><h1>What to Tell Your Board</h1><p><em>A <a href="https://stratsec.com/wp-content/uploads/2026/05/Stratsec_Board_Brief_Quantum_Threat_30Apr2026.pptx">board-ready slide</a> summarising this briefing is available as a separate PPTX file for inclusion in your next risk committee deck.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://stratsec.com/wp-content/uploads/2026/05/Stratsec_Board_Brief_Quantum_Threat_30Apr2026.pptx&quot;,&quot;text&quot;:&quot;Board-Ready Slide [.pptx]&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://stratsec.com/wp-content/uploads/2026/05/Stratsec_Board_Brief_Quantum_Threat_30Apr2026.pptx"><span>Board-Ready Slide [.pptx]</span></a></p><p>Recent research has sharply reduced the estimated computing resources needed to break the encryption that protects most of our digital infrastructure. Multiple independent research teams have published results in the first four months of 2026 showing that the quantum computers required are far smaller than previously assumed: roughly one-tenth the size estimated just two years ago. A quantum computer capable of breaking current encryption does not exist today, but the path to building one is shorter than we previously understood.</p><p>For our organisation, this means three things.</p><p>First, we need to know what cryptography we use and where. We are commissioning a cryptographic inventory across our technology estate to identify every system that depends on the algorithms that quantum computers will eventually compromise. This inventory is the prerequisite for any migration and should be completed within the current quarter.</p><p>Second, we need a migration plan. NIST has published post-quantum cryptographic standards and set deprecation deadlines: quantum-vulnerable algorithms will be deprecated for federal systems by 2030 and disallowed by 2035. The European Commission&#8217;s roadmap targets critical infrastructure migration by end of 2030. The UK NCSC expects high-priority systems migrated by 2031. We need a phased migration programme with board-level sponsorship, a designated owner, and a timeline. We will present a proposed programme structure at the next risk committee meeting.</p><p>Third, there is a current risk that requires attention now. Adversaries, state intelligence services in particular, are collecting encrypted data today with the intention of decrypting it once quantum capability arrives. If our organisation holds data whose confidentiality extends beyond 2035 (and most of our sensitive data does), we should prioritise protecting those data flows first.</p><p>This is not a crisis. It is a long-duration infrastructure programme comparable in scope to the Y2K remediation or the migration from SHA-1 to SHA-2, which took the industry over twelve years. The difference is that we now have clear standards, concrete deadlines, and a narrowing window. We recommend the board endorse the establishment of a PQC migration programme this quarter, with a first progress report at the Q3 risk committee meeting.</p><div><hr></div><h2>Indicator Watch</h2><p>Stratsec is tracking the emergence of commercial PQC compliance verification services and the speed at which major cloud providers operationalise NIST PQC standards in production. Google and Cloudflare have both set 2029 as their internal PQC migration deadlines; Cloudflare reports that over 65% of human-initiated traffic on its edge already uses post-quantum encryption.[10] The rate at which enterprise software, hardware vendors, and certificate authorities follow will determine whether the 2030 deprecation timeline is achievable for most organisations or whether it becomes a compliance cliff.</p><p>We are also monitoring the acceleration of experimental quantum attacks on elliptic curve cryptography. The Project Eleven demonstration (15-bit ECC key broken on public hardware in April 2026, up from 6-bit in September 2025) provides a measurable benchmark of hardware progress.[8] While the gap to 256-bit cryptographic relevance remains immense, the trajectory bears watching.</p><p>Separately, we are monitoring for classified or TLP-restricted government advisories that may indicate a revision in official threat timelines. Any acceleration in government advisory language, from &#8220;prepare&#8221; to &#8220;act now,&#8221; would be a leading indicator that intelligence community assessments have shifted.</p><div><hr></div><h2>References</h2><p>[1] C. Gidney, &#8220;How to Factor 2048 Bit RSA Integers in 8 Hours Using 20 Million Noisy Qubits&#8221; (May 2025), arXiv:2103.06159. <a href="https://arxiv.org/abs/2505.15917">https://arxiv.org/abs/2505.15917</a></p><p>[2] C. Chevignard, P-A. Fouque, A. Schrottenloher, &#8220;Reducing Quantum Factoring to Modular Exponentiation Through Quantum Fast Residue Multiplication,&#8221; CRYPTO 2025. <a href="https://eprint.iacr.org/2024/222">https://eprint.iacr.org/2024/222</a></p><p>[3] C. Chevignard, P-A. Fouque, A. Schrottenloher, &#8220;Quantum Cryptanalysis of Elliptic Curves: Improved Circuits for ECDLP,&#8221; EUROCRYPT 2026. <a href="https://eprint.iacr.org/2026/280">https://eprint.iacr.org/2026/280</a> </p><p>[4] Google Quantum AI, &#8220;Securing Elliptic Curve Cryptocurrencies against Quantum Vulnerabilities: Resource Estimates and Mitigations,&#8221; 31 March 2026. <a href="https://quantumai.google/static/site-assets/downloads/cryptocurrency-whitepaper.pdf">https://quantumai.google/static/site-assets/downloads/cryptocurrency-whitepaper.pdf</a></p><p>[5] Oratomic, Caltech, UC Berkeley, &#8220;Compiling Shor&#8217;s Algorithm on Neutral Atom Arrays,&#8221; 31 March 2026, arXiv preprint. Analysis: <a href="https://postquantum.com/security-pqc/10000-qubits-shors/">https://postquantum.com/security-pqc/10000-qubits-shors/</a></p><p>[6] IonQ, &#8220;Fault-Tolerant Quantum Computing with Trapped Ions: The Walking Cat Architecture,&#8221; arXiv:2604.19481, 21 April 2026. <a href="https://arxiv.org/abs/2604.19481">https://arxiv.org/abs/2604.19481</a> &#8212; Analysis: <a href="https://postquantum.com/quantum-research/ionq-walking-cat-trapped-ion-ftqc/">https://postquantum.com/quantum-research/ionq-walking-cat-trapped-ion-ftqc/</a></p><p>[7] Coinbase Independent Advisory Board on Quantum Computing and Blockchain, &#8220;Quantum Computing and Blockchain&#8221; (position paper), 21 April 2026. Authors: S. Aaronson, D. Boneh, J. Drake, S. Kannan, Y. Lindell, D. Malkhi. <a href="https://www.coinbase.com/blog/coinbase-quantum-advisory-council-publishes-position-paper-on-quantum-computing-and-blockchain">https://www.coinbase.com/blog/coinbase-quantum-advisory-council-publishes-position-paper-on-quantum-computing-and-blockchain</a> &#8212; Analysis: <a href="https://postquantum.com/security-pqc/coinbase-quantum-blockchain-paper-analysis/">https://postquantum.com/security-pqc/coinbase-quantum-blockchain-paper-analysis/</a></p><p>[8] Project Eleven / G. Lelli, 15-bit ECC key broken on public quantum hardware, 24 April 2026. <a href="https://intellectia.ai/blog/bitcoin-quantum-threat-2026">https://intellectia.ai/blog/bitcoin-quantum-threat-2026</a></p><p>[9] S. Aaronson, &#8220;Will you heed my warnings NOW?&#8221;, Shtetl-Optimized (blog), 29 April 2026. <a href="https://scottaaronson.blog/?p=9718">https://scottaaronson.blog/?p=9718</a></p><p>[10] Cloudflare, &#8220;Post-Quantum Roadmap,&#8221; 30 April 2026. <a href="https://blog.cloudflare.com/post-quantum-roadmap/">https://blog.cloudflare.com/post-quantum-roadmap/</a></p><p>[11] NIST Interagency Report 8547, &#8220;Transition to Post-Quantum Cryptography Standards,&#8221; Initial Public Draft, November 2024. <a href="https://csrc.nist.gov/pubs/ir/8547/ipd">https://csrc.nist.gov/pubs/ir/8547/ipd</a> &#8212; Analysis: <a href="https://postquantum.com/security-pqc/nist-ir-8547-ipd/">https://postquantum.com/security-pqc/nist-ir-8547-ipd/</a></p><p>[12] NSA, &#8220;Commercial National Security Algorithm Suite 2.0 (CNSA 2.0),&#8221; September 2022, updated through 2025. <a href="https://media.defense.gov/2022/Sep/07/2003071836/-1/-1/0/CSI_CNSA_2.0_FAQ_.PDF">https://media.defense.gov/2022/Sep/07/2003071836/-1/-1/0/CSI_CNSA_2.0_FAQ_.PDF</a></p><p>[13] UK NCSC, &#8220;Timelines for Migration to Post-Quantum Cryptography,&#8221; March 2025. Three-phase timeline: 2028 discovery, 2031 high-priority migration, 2035 full migration. <a href="https://www.ncsc.gov.uk/guidance/pqc-migration-timelines">https://www.ncsc.gov.uk/guidance/pqc-migration-timelines</a></p><p>[14] European Commission / NIS Cooperation Group, &#8220;A Coordinated Implementation Roadmap for the Transition to Post-Quantum Cryptography,&#8221; 11 June 2025. Milestones: Member State transition begins end of 2026; critical infrastructure migration targeted by end of 2030. <a href="https://digital-strategy.ec.europa.eu/en/library/coordinated-implementation-roadmap-transition-post-quantum-cryptography">https://digital-strategy.ec.europa.eu/en/library/coordinated-implementation-roadmap-transition-post-quantum-cryptography</a></p><p>[15] Australian Signals Directorate (ASD), ISM guidance on post-quantum cryptography, 2024, recommending elimination of classical public-key cryptography by 2030. <a href="https://www.cyber.gov.au/resources-business-and-government/essential-cyber-security/ism">https://www.cyber.gov.au/resources-business-and-government/essential-cyber-security/ism</a></p><p>[16] NIST, FIPS 203 (ML-KEM), FIPS 204 (ML-DSA), FIPS 205 (SLH-DSA), finalised August 2024. HQC selected March 2025. <a href="https://csrc.nist.gov/projects/post-quantum-cryptography">https://csrc.nist.gov/projects/post-quantum-cryptography</a></p><p>[17] Current record for Shor&#8217;s algorithm factorisation on quantum hardware: 21 = 3 &#215; 7 (2012). See: <a href="https://postquantum.com/post-quantum/quantum-computer-factored-question/">https://postquantum.com/post-quantum/quantum-computer-factored-question/</a></p><p>[18] CRQC Quantum Capability Framework, PostQuantum.com: <a href="https://postquantum.com/post-quantum/crqc-quantum-capability-framework/">https://postquantum.com/post-quantum/crqc-quantum-capability-framework/</a></p><p>[19] Global Risk Institute / evolutionQ, &#8220;Quantum Threat Timeline Report,&#8221; 7th edition, January 2025, M. Mosca and M. Piani. <a href="https://globalriskinstitute.org/publication/2024-quantum-threat-timeline-report/">https://globalriskinstitute.org/publication/2024-quantum-threat-timeline-report/</a></p><p>[20] Harvest Now, Decrypt Later: <a href="https://postquantum.com/post-quantum/harvest-now-decrypt-later-hndl/">https://postquantum.com/post-quantum/harvest-now-decrypt-later-hndl/</a></p><p>[21] Trust Now, Forge Later: <a href="https://postquantum.com/post-quantum/trust-now-forge-later/">https://postquantum.com/post-quantum/trust-now-forge-later/</a></p><p>[22] Google Security Blog, September 2024 (ML-KEM in Chrome); Apple PQ3 protocol, 2024. </p><p>[23] ENISA, ECCG Cryptographic Mechanisms Report v2.0, 2025. <a href="https://www.enisa.europa.eu/publications/post-quantum-cryptography-integration-study">https://www.enisa.europa.eu/publications/post-quantum-cryptography-integration-study</a> &#8212; See also ETSI TS 103 744 on quantum-safe cryptography.</p><div><hr></div><p><em>Stratsec: Emerging technology threats, without the hype.</em></p>]]></content:encoded></item><item><title><![CDATA[Iranian Cyber Actors Target Western Critical Infrastructure: PLC Exploits Escalate Amid Regional Conflict]]></title><description><![CDATA[Iranian-affiliated groups are actively disrupting programmable logic controllers in US water, energy, and government systems. The attacks exploit basic hygiene failures, not advanced capabilities. Wha]]></description><link>https://intelligence.stratsec.com/p/iranian-cyber-actors-target-western</link><guid isPermaLink="false">https://intelligence.stratsec.com/p/iranian-cyber-actors-target-western</guid><dc:creator><![CDATA[Stratsec]]></dc:creator><pubDate>Fri, 01 May 2026 05:28:21 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/398163de-fef5-42c7-abda-7582b1cc4aff_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><strong>Domain:</strong> Tech-Geopolitics</p><h2><strong>The Development</strong></h2><p>On 7 April 2026, the FBI, CISA, NSA, EPA, Department of Energy, and US Cyber Command issued a joint advisory (AA26-097A) confirming that Iranian-affiliated advanced persistent threat actors are actively exploiting internet-facing programmable logic controllers across US critical infrastructure.[1] The advisory identified disruptions in the Government Services, Water and Wastewater Systems, and Energy sectors, with some victims experiencing operational disruption and financial loss.[1]</p><p>The attackers used overseas IP addresses and leased third-party hosting infrastructure to connect to internet-exposed Rockwell Automation/Allen-Bradley PLCs, including CompactLogix and Micro850 controllers.[1] They used Rockwell&#8217;s own Studio 5000 Logix Designer software to establish connections, extracted and modified project files containing ladder logic and configuration settings, manipulated data displayed on human-machine interface and SCADA systems, and deployed Dropbear SSH on victim endpoints for persistent access via port 22.[1] Malicious traffic was observed on ports 44818, 2222, 102, 22, and 502, with patterns suggesting interest in Siemens S7 PLCs as well.[1]</p><p>The advisory links this campaign to the CyberAv3ngers group, affiliated with Iran&#8217;s Islamic Revolutionary Guard Corps Cyber-Electronic Command (IRGC-CEC), also tracked as Shahid Kaveh Group, Storm-0784, Hydro Kitten, Bauxite, and UNC5691.[1] The same ecosystem compromised at least 75 Unitronics PLC devices across US water and wastewater systems during a similar campaign beginning in November 2023.[1][2] That earlier campaign relied on factory default passwords. The current campaign exploits CVE-2021-22681, an authentication bypass vulnerability in Rockwell Logix controllers for which no vendor patch exists; only defence-in-depth mitigations are available.[1][3]</p><p>The escalation coincides with the US-Israeli military campaign against Iran that began on 28 February 2026 (Operation Epic Fury).[4] Strikes targeted Iranian nuclear facilities, military infrastructure, and senior leadership, including Supreme Leader Ali Khamenei.[4][5] Iran experienced a near-total domestic internet blackout shortly after, and a conditional ceasefire was declared on 8 April.[4] The UK&#8217;s National Cyber Security Centre issued an advisory on 2 March 2026 warning UK organisations of a heightened risk of indirect cyber threats, particularly for those with supply chains or operations in the Middle East.[6] The NCSC assessed that Iranian state and Iran-linked cyber actors &#8220;almost certainly currently maintain at least some capability to conduct cyber activity&#8221; despite the blackout.[6]</p><p>The broader Iranian cyber ecosystem has activated in parallel. Tenable reports that CyberAv3ngers&#8217; ICS exploitation techniques have proliferated to an estimated 60 or more pro-Iranian hacktivist groups.[3] Handala, a group widely assessed to operate under Iranian intelligence protection, claimed a destructive attack against US medical technology firm Stryker in March 2026, asserting it wiped approximately 200,000 devices via a compromised Microsoft Intune environment; Stryker confirmed severe operational disruption, though the scale figures have not been independently verified.[7] MuddyWater was observed conducting coordinated intrusions against a US bank, an airport, NGOs, and an Israeli-linked defence supplier in late February, using newly built backdoors and cloud-based data exfiltration.[7] The Center for Strategic and International Studies published analysis in April characterising Iran&#8217;s cyber posture as a shift from episodic attacks to a sustained campaign treating cyberspace as an extension of state power.[8]</p><p>US financial regulators have also responded. FINRA issued a cybersecurity alert warning member firms of heightened risks from Iranian state-sponsored and aligned actors, specifically citing the exploitation of known vulnerabilities, default credentials, and brute-force techniques.[9]</p><div><hr></div><p><em>In future issues, the following sections (Reality Check, Action Brief, CISO Governance Briefing, and Board Brief) will be available exclusively to paid subscribers. This issue is published in full so you can experience the complete Stratsec intelligence product.</em></p><div><hr></div><h2><strong>The Reality Check</strong></h2><p><strong>Assessment: Significant.</strong> This is a genuine escalation in Iranian intent to cause disruptive effects on Western critical infrastructure. It is not a novel capability or an insurmountable technical challenge.</p><p>The core issue is basic: exposed industrial control systems are being compromised through poor hygiene. In every documented case, the attackers are scanning the public internet for PLCs protected by default passwords, unpatched authentication bypasses, or no access controls at all. They are connecting using the manufacturer&#8217;s own legitimate engineering software. The sophistication is in the targeting and the intent, not in the technique.</p><p>Two things have changed since the 2023 Unitronics campaign. First, the attacks are causing confirmed operational disruption and financial loss. The 2023 campaign was largely symbolic: screens displaying political messages, limited real-world impact. The April 2026 advisory documents actual manipulation of control logic and process displays, with victims reporting genuine operational consequences.[1] Second, the geopolitical context has elevated the threat tempo. The February 2026 military campaign against Iran triggered the largest single-event activation of Iranian-aligned cyber actors ever documented, according to threat intelligence firms.[7] With conventional military options constrained, cyber operations against US and allied infrastructure become a more attractive asymmetric response for Tehran.[8]</p><p>Three realities temper the alarm.</p><p>First, the technical barrier remains low. These attacks work because defenders have not implemented measures that have been recommended for years: remove OT from the public internet, change default passwords, segment networks, enforce authentication. Organisations that have already done this are far less attractive targets. The NCSC&#8217;s assessment is instructive: there is no significant change in the direct cyber threat to the UK, but the risk of indirect effects and opportunistic exploitation has increased.[6]</p><p>Second, the Iranian cyber apparatus has been degraded by the conflict. The 2026 ODNI Annual Threat Assessment assessed that Iran&#8217;s ability to conduct and defend against cyber operations was constrained by the 2025 war and subsequent military pressure.[10] The IRGC command structure has been significantly weakened. The domestic internet blackout limits coordination. But degraded is not neutralised. The proxy ecosystem, cultivated over a decade, now operates with greater autonomy. CyberAv3ngers&#8217; techniques have spread to dozens of affiliated groups.[3] The threat has become more distributed, less predictable, and in some ways harder to defend against because there is no single adversary with a single playbook.</p><p>Third, geographic distance offers no protection. Iranian-affiliated groups scan globally for exposed OT devices. Any organisation running internet-connected industrial control systems is a potential target, regardless of sector or location. European organisations fall under additional pressure via NIS2 obligations to manage emerging threats proportionately. The UK NCSC&#8217;s advisory and FINRA&#8217;s alert both confirm that allied nations and financial regulators treat this as requiring organisational action now.[6][9]</p><p>The central message: the nature of the Iranian cyber threat to Western critical infrastructure has not changed. What has changed is the operational tempo, the willingness to cause real disruption, and the number of actors involved. Defenders retain a structural advantage, but only if they treat internet-exposed OT as the urgent hygiene failure it is.</p><h2><strong>The Action Brief</strong></h2><p><strong>Remove OT devices from the public internet.</strong> If a PLC, HMI, RTU, or SCADA component does not require public internet connectivity to perform its function, disconnect it. If remote access is operationally necessary, place it behind a secure gateway or VPN with phishing-resistant multifactor authentication. Conduct an external scan this week to verify no OT devices are directly reachable. This is the single most effective action you can take.</p><p><strong>Inventory your Rockwell Automation Logix controllers.</strong> CyberAv3ngers are actively exploiting CVE-2021-22681, for which there is no vendor patch. Implement Rockwell&#8217;s recommended mitigations: network segmentation isolating engineering workstations from untrusted networks, strict access controls, and continuous monitoring for unauthorised engineering sessions. Ensure physical mode switches on applicable controllers are locked in the run position to prevent remote logic modifications.</p><p><strong>Audit third-party and vendor remote access.</strong> The Stryker compromise and the advisory&#8217;s own TTPs confirm that trusted third-party pathways remain a primary attack vector. Review every vendor maintenance gateway, remote support portal, and contractor access account touching your OT environment. Remove anything that cannot be justified. Enforce MFA on everything that remains. Enable logging.</p><p><strong>Search your network logs for the advisory&#8217;s indicators.</strong> Direct your SOC to query firewall and IDS logs for inbound connections to ports 44818, 2222, 102, 22, and 502 from overseas hosting providers. Look for the IP addresses listed in the CISA STIX package (AA26-097A). Check for unauthorised Dropbear SSH installations and unexpected modifications to Rockwell project files (.ACD, .L5X).</p><p><strong>Test offline backups of PLC logic and configurations.</strong> If an attacker manipulates your control logic, your recovery speed depends on having verified, offline backups of PLC programmes, HMI configurations, and historian data. Test restorability now, before you need it.</p><h1><strong>CISO Governance Briefing</strong></h1><h3><strong>Enterprise Risk Management</strong></h3><p>This development does not create a new risk category. It changes the likelihood and velocity parameters for OT compromise scenarios that should already be in your risk register.</p><p>Update the likelihood rating for exploitation of internet-facing OT devices. Where previous assessments assumed that ICS targeting by state actors was rare, opportunistic at most, and focused primarily on geographically proximate adversaries, the evidence now shows sustained, globally scoped campaigns causing confirmed operational damage.[1] For organisations running internet-exposed PLCs, HMIs, or SCADA components in water, energy, government services, or manufacturing, move the likelihood assessment up by at least one tier.</p><p>If you use a quantitative model, revisit the time-to-exploit assumption. The advisory documents connection and manipulation within hours of device discovery. If you use a qualitative model, the relevant shift is from &#8220;possible&#8221; to &#8220;likely&#8221; for any externally facing OT system without gateway-mediated access controls.</p><h3><strong>Budget and Resourcing</strong></h3><p>This does not require a large new technology investment for most organisations. The primary spend implications are in process discipline and targeted hardening of existing OT environments.</p><p>Reallocate resources from deferred OT segmentation projects. If your organisation lacks dedicated OT visibility or monitoring capability, budget for one to two targeted hires or external support to accelerate network segmentation and logging improvements over the next two quarters. The tools are largely available; the gap in most organisations is adoption.</p><p>For organisations with legacy OT where patching is impractical (and CVE-2021-22681 has no patch), compensating controls are capital investments: hardware-enforced unidirectional gateways at IT/OT boundaries, managed switches with documented access control lists, and independent monitoring of OT traffic. If these investments have been deferred, the case for acceleration is now stronger.</p><h3><strong>Policy and Procedure Updates</strong></h3><p>Review and update four areas.</p><p>Network architecture and remote access: explicitly prohibit direct internet exposure of OT devices. Mandate gateway-mediated access with MFA for all remote OT connections.</p><p>Vulnerability and asset management: include all internet-facing OT devices in scanning and patching cycles. Where patching is impractical, document approved compensating controls and review them quarterly.</p><p>Incident response: incorporate OT-specific playbooks for PLC manipulation and HMI data alteration scenarios, including rapid isolation procedures, offline restoration from backups, and coordination with OT engineering teams. The window between initial access and operational disruption may be hours, not days.</p><p>Third-party risk management: extend supplier assurance to cover OT security controls explicitly (see Supplier Assurance Questions below).</p><h3><strong>Regulatory Exposure</strong></h3><p>NIS2 and DORA require boards to oversee proportionate management of emerging threats. Regulators increasingly view unmitigated internet exposure of critical control systems as evidence of inadequate risk management. The April 2026 CISA advisory, combined with the UK NCSC guidance, provides contemporary benchmarks against which post-incident regulatory scrutiny will be judged.[1][6]</p><p>No regulator has formally changed its expectations. But the practical defensibility of a 90-day patching cycle for an internet-exposed PLC has weakened considerably when six government agencies are jointly warning of active exploitation causing financial loss. European organisations subject to NIS2&#8217;s board-level accountability provisions (with fines of up to 2% of global turnover) should document their board&#8217;s awareness and the specific hardening actions taken. FINRA&#8217;s alert to financial services firms signals that sector regulators are watching this space closely.[9]</p><h3><strong>Team Skills</strong></h3><p>The capability gap this exposes is at the IT/OT boundary. Your security team needs people who can identify and secure internet-exposed industrial devices, configure secure remote access gateways, interpret vendor-specific logs (Rockwell Studio 5000 activity, for example), and distinguish legitimate engineering traffic from anomalous lateral movement.</p><p>Most organisations already have the necessary personnel. What they lack is focused training and clear policy direction. Prioritise upskilling in OT network segmentation and basic ICS protocol monitoring over the next 12 months. Cross-train process control engineers in security fundamentals and security analysts in OT basics.</p><h3><strong>Second-Line and Third-Line Oversight</strong></h3><p>Risk management (second line) should verify that OT exposure has been mapped and remediated across the estate, and that risk register updates reflect current threat velocity and the demonstrated Iranian willingness to act against exposed devices.</p><p>Internal audit (third line) should include internet-exposed OT devices and remote access controls in its next cyber risk review scope. The audit should verify that the organisation maintains a current inventory of internet-facing OT assets and that compensating controls for unpatchable vulnerabilities are documented and tested.</p><h2><strong>Supplier Assurance Questions</strong></h2><p>Send these to critical OT/ICS technology suppliers, system integrators, and remote-service providers this quarter.</p><ol><li><p>Are any of the PLCs, HMIs, or SCADA components you supply or support configured for direct internet access? If yes, what controls prevent unauthorised programming or data manipulation?</p></li><li><p>What is your process for ensuring programming protection, mode-switch settings, and offline backup capabilities are enabled by default or during deployment?</p></li><li><p>Do you use or recommend secure gateway/jump-host architectures for any remote access to our OT environments? How is MFA enforced at the network layer?</p></li><li><p>Have you reviewed the April 2026 CISA/FBI advisory (AA26-097A) for applicability to the equipment and services you provide to us?</p></li><li><p>What monitoring and logging do you maintain for vendor software (e.g., Studio 5000) access to our systems?</p></li><li><p>In the event of detected manipulation of project files or HMI data, what is your compressed notification and restoration support timeline?</p></li><li><p>Do your contracts include liability provisions for breaches originating from internet-exposed OT devices or remote-access pathways you manage?</p></li></ol><h2><strong>Team Readiness Checklist</strong></h2><p>Use these questions with your OT/security leadership team:</p><p><strong>OT exposure and segmentation</strong></p><ul><li><p>Have we identified and documented every internet-facing PLC, HMI, or SCADA component across the estate?</p></li><li><p>Which devices remain directly reachable from the internet, and what is the plan to remove them?</p></li><li><p>Is the boundary between corporate IT and OT governed by strict, hardware-enforced segmentation?</p></li></ul><p><strong>Remote access controls</strong></p><ul><li><p>Are all remote OT connections mediated through MFA-enforced gateways rather than direct exposure?</p></li><li><p>Have we audited third-party and vendor remote-access privileges in the last 90 days?</p></li><li><p>Have cellular modems and other remote field devices been secured and logged?</p></li></ul><p><strong>Backup and recovery</strong></p><ul><li><p>Do we have tested, offline backups of all critical PLC logic and configurations?</p></li><li><p>Can we restore a manipulated PLC within operational recovery time objectives?</p></li></ul><p><strong>Monitoring</strong></p><ul><li><p>Are we logging and alerting on anomalous vendor software connections, unexpected OT port activity, or project file changes?</p></li><li><p>Has our SOC ingested the IOCs from CISA advisory AA26-097A?</p></li></ul><p><strong>Incident response</strong></p><ul><li><p>Has the IR team rehearsed an OT manipulation scenario involving HMI/SCADA data alteration?</p></li><li><p>Can we execute containment actions (network isolation, mode-switch lockdown, credential rotation) within one hour of detection?</p></li></ul><h2><strong>Second-Line and Third-Line Assurance Questions</strong></h2><p><strong>For risk management (second line):</strong></p><ul><li><p>Has the first-line team mapped all internet-exposed OT devices and implemented compensating controls or removal?</p></li><li><p>Have risk register entries for OT compromise been updated to reflect current Iranian TTPs and threat velocity?</p></li><li><p>Is there evidence of MFA-enforced gateway access for all remote OT pathways?</p></li><li><p>Has the incident response playbook been tested against an OT manipulation scenario in the last 90 days?</p></li></ul><p><strong>For internal audit (third line):</strong></p><ul><li><p>Does the organisation maintain a current inventory of internet-facing OT assets, and have exposure risks been addressed?</p></li><li><p>Are there documented, tested procedures for rapid PLC restoration following manipulation?</p></li><li><p>Has the supplier assurance programme been extended to cover OT/ICS security controls?</p></li><li><p>Is the board receiving accurate, timely reporting on geopolitically driven cyber threats to OT environments?</p></li></ul><h2><strong>Tabletop Exercise: OT Manipulation Scenario</strong></h2><p>Hand this scenario to your incident response team. It requires no additional preparation. Allow 90 minutes.</p><p><strong>Scenario:</strong></p><p>It is 08:30 on a Wednesday. Your water treatment plant&#8217;s SCADA system begins displaying anomalous readings on several HMI screens. Operators report that set points for chemical dosing appear to have been altered remotely. Initial triage confirms unauthorised access to a CompactLogix PLC via an unexpected external IP address using Rockwell programming software. The system remains operational but critical process parameters are changing. Plant safety systems have not yet triggered, but the window for safe containment is narrowing.</p><p>At 09:15, the incident response team discovers that the attackers used credentials belonging to a trusted third-party maintenance contractor. The contractor&#8217;s access was not enrolled in multifactor authentication. The attackers are actively modifying ladder logic on a second controller responsible for pressure regulation.</p><p><strong>Discussion questions:</strong></p><ol><li><p>What is our immediate containment action, and can we execute it within 15 minutes of detection?</p></li><li><p>Who in the organisation has the authority to physically isolate the OT network, and are they available now?</p></li><li><p>How do we restore the manipulated controllers from offline backups while maintaining partial manual operations?</p></li><li><p>The same third-party contractor has access to four other sites in our estate. How do we check and protect those systems while responding to the active compromise?</p></li><li><p>Under NIS2, what are our immediate regulatory notification obligations and timeline? Who needs to be notified internally within the first hour?</p></li><li><p>What post-incident changes to remote access policy and network segmentation are non-negotiable?</p></li></ol><h1><strong>What to Tell Your Board</strong></h1><p><em>A <a href="https://stratsec.com/wp-content/uploads/2026/05/Stratsec_Board_Brief_Iranian_Cyber_Threat-30Apr2026.pptx">board-ready slide</a> summarising this briefing is available as a separate PPTX file for inclusion in your next risk committee deck.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://stratsec.com/wp-content/uploads/2026/05/Stratsec_Board_Brief_Iranian_Cyber_Threat-30Apr2026.pptx&quot;,&quot;text&quot;:&quot;Board-Ready Slide [.pptx]&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://stratsec.com/wp-content/uploads/2026/05/Stratsec_Board_Brief_Iranian_Cyber_Threat-30Apr2026.pptx"><span>Board-Ready Slide [.pptx]</span></a></p><p>Iranian-affiliated cyber actors are actively targeting internet-connected industrial control systems in Western critical infrastructure. US government agencies have confirmed that these attacks have caused operational disruptions and financial losses at water, energy, and government facilities. The attackers are exploiting basic security weaknesses in systems that should never have been connected to the public internet.</p><p>For our organisation, this means three things.</p><p>First, any remaining internet exposure of our operational technology must be eliminated or tightly controlled. We are conducting an emergency audit to verify that no industrial control systems are directly reachable from the internet, and are reviewing and hardening all remote-access pathways and PLC configurations this quarter.</p><p>Second, our OT environments require the same disciplined segmentation and monitoring long applied to corporate IT networks. We are accelerating existing segmentation projects and ensuring offline backups of critical control system logic are in place and tested.</p><p>Third, we are extending our supplier assurance and third-party risk processes to cover OT/ICS security controls explicitly. A breach through a supplier&#8217;s remote access now moves at machine speed.</p><p>This is not a new threat category, but it is a clear demonstration that previously tolerable exposures are now actively exploited. The appropriate response is to close long-standing gaps in our OT security posture. We recommend the board receive a detailed OT exposure assessment and remediation plan at the next risk committee meeting, with implementation targeted for completion before end of Q3 2026.</p><h2><strong>Indicator Watch</strong></h2><p>Stratsec is tracking the potential proliferation of PLC exploitation techniques beyond Rockwell and Unitronics devices to other common industrial control system vendors. The April 2026 advisory explicitly notes actors probing ports associated with Siemens S7 protocols, and the pattern of scanning activity suggests interest in any internet-exposed industrial controller regardless of manufacturer.[1]</p><p>We are also monitoring for increased targeting of European CNI operators via supply-chain pathways or indirect spillover effects, consistent with UK NCSC guidance on heightened regional risk.[6] If exploitation tools and techniques continue to spread across the 60-plus affiliated hacktivist groups identified by threat intelligence firms, the volume of opportunistic attacks against exposed industrial control systems will increase over the next two quarters.[3]</p><h2><strong>References</strong></h2><p>[1] FBI, CISA, NSA, EPA, DOE, and CNMF, &#8220;Iranian-Affiliated Cyber Actors Exploit Programmable Logic Controllers Across US Critical Infrastructure,&#8221; AA26-097A, 7 April 2026. <a href="https://www.cisa.gov/news-events/cybersecurity-advisories/aa26-097a">https://www.cisa.gov/news-events/cybersecurity-advisories/aa26-097a</a></p><p>[2] CISA, &#8220;IRGC-Affiliated Cyber Actors Exploit PLCs in Multiple Sectors, Including US Water and Wastewater Systems Facilities,&#8221; AA23-335A, updated December 2024. <a href="https://www.cisa.gov/news-events/cybersecurity-advisories/aa23-335a">https://www.cisa.gov/news-events/cybersecurity-advisories/aa23-335a</a></p><p>[3] Tenable, &#8220;CyberAv3ngers: FAQ About Iran-Linked Threat Group Targeting U.S. Critical Infrastructure,&#8221; April 2026. <a href="https://www.tenable.com/blog/what-to-know-about-cyberav3ngers-the-irgc-linked-group-targeting-critical-infrastructure">https://www.tenable.com/blog/what-to-know-about-cyberav3ngers-the-irgc-linked-group-targeting-critical-infrastructure</a></p><p>[4] House of Commons Library, &#8220;Israel/US-Iran conflict 2026: Background and UK response,&#8221; Research Briefing CBP-10521, updated April 2026. <a href="https://commonslibrary.parliament.uk/research-briefings/cbp-10521/">https://commonslibrary.parliament.uk/research-briefings/cbp-10521/</a></p><p>[5] The Record, &#8220;British organizations urged to be alert to threat of Iranian cyberattacks,&#8221; 2 March 2026. <a href="https://therecord.media/iran-britain-cyber-threats-warning">https://therecord.media/iran-britain-cyber-threats-warning</a></p><p>[6] UK NCSC, &#8220;NCSC advises UK organisations to take action following conflict in the Middle East,&#8221; 2 March 2026. <a href="https://www.ncsc.gov.uk/news/ncsc-advises-uk-organisations-take-action-following-conflict-in-middle-east">https://www.ncsc.gov.uk/news/ncsc-advises-uk-organisations-take-action-following-conflict-in-middle-east</a></p><p>[7] Sapphire, &#8220;What the 2026 Iran Conflict Means for UK Cyber Risk,&#8221; March 2026. <a href="https://www.sapphire.net/blogs-press-releases/what-the-2026-iran-conflict-means-for-uk-cyber-risk/">https://www.sapphire.net/blogs-press-releases/what-the-2026-iran-conflict-means-for-uk-cyber-risk/</a></p><p>[8] Center for Strategic and International Studies, &#8220;The Iranian Cyber Threat to U.S. Critical Infrastructure,&#8221; April 2026. <a href="https://www.csis.org/analysis/iranian-cyber-threat-us-critical-infrastructure">https://www.csis.org/analysis/iranian-cyber-threat-us-critical-infrastructure</a></p><p>[9] FINRA, &#8220;Cybersecurity Alert: Heightened Threats From Iranian Cyber Actors,&#8221; March 2026. <a href="https://www.finra.org/rules-guidance/guidance/cybersecurity-alert-heightened-threats-iranian-cyber-actors">https://www.finra.org/rules-guidance/guidance/cybersecurity-alert-heightened-threats-iranian-cyber-actors</a></p><p>[10] Office of the Director of National Intelligence, &#8220;2026 Annual Threat Assessment of the U.S. Intelligence Community,&#8221; 2026. Referenced in CSIS analysis.</p><p><em>Stratsec: Emerging technology threats, without the hype.</em></p>]]></content:encoded></item><item><title><![CDATA[What Is Stratsec, and Why We Built This]]></title><description><![CDATA[Practitioner-led intelligence on emerging technology threats: what's real, what's overblown, and what to do about it.]]></description><link>https://intelligence.stratsec.com/p/what-is-stratsec-and-why-we-built</link><guid isPermaLink="false">https://intelligence.stratsec.com/p/what-is-stratsec-and-why-we-built</guid><dc:creator><![CDATA[Stratsec]]></dc:creator><pubDate>Thu, 30 Apr 2026 10:51:24 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/e47874ea-c94b-4f6c-8df7-3d04acfdd487_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Every week, a new headline tells you the sky is falling. AI is coming for your defences. Quantum will break encryption by Tuesday. Autonomous drones are a threat you&#8217;re completely unprepared for.</p><p>Most of it is noise. Some of it isn&#8217;t. The problem is telling which is which.</p><p>That&#8217;s what Stratsec does.</p><h2>The problem we solve</h2><p>The information landscape for emerging technology threats is broken. Vendor-funded research exists to sell products. Media coverage optimises for clicks, not accuracy. Analyst reports arrive months after you needed them. And government advisories, the ones that used to be reliable, are getting thinner as agencies lose headcount and budgets.</p><p>Meanwhile, NIS2, DORA, and the EU AI Act are creating board-level accountability for emerging technology risk. Directors have a regulatory obligation to oversee threats they don&#8217;t yet have frameworks to assess.</p><p>CISOs and security leaders need a source they can trust. One that tells them what&#8217;s real, what&#8217;s overblown, and what to actually do about it.</p><h2>What we publish</h2><p>The <strong>Stratsec Emerging Threat Monitor</strong> is a regular intelligence briefing covering emerging technology developments with security and risk implications. Each issue covers developments across five threat domains:</p><ul><li><p><strong>AI Security &amp; Governance:</strong> what new AI capabilities actually mean for your security posture</p></li><li><p><strong>Quantum Security:</strong> real timelines versus media hype, PQC migration priorities</p></li><li><p><strong>Robotics, Drones &amp; Autonomous Systems:</strong> cyber-physical attack surfaces that are expanding faster than the frameworks to manage them</p></li><li><p><strong>Tech-Geopolitics:</strong> how semiconductor politics, export controls, and state-sponsored competition affect your technology decisions</p></li><li><p><strong>Regulatory Horizon:</strong> NIS2, DORA, EU AI Act, and what&#8217;s coming next</p></li></ul><p>Issues take one of two formats. Multi-item issues cover four to five developments across the domains. Single-topic deep dives give one major development the comprehensive treatment it warrants: full narrative, governance analysis, and a complete operational toolkit.</p><p>Every item follows a three-part structure:</p><p><strong>The Development:</strong> what actually happened, in plain language, without hype. <em>(Free for all subscribers.)</em></p><p><strong>The Reality Check:</strong> is this a genuine shift or incremental? Should you brief your board or file it for later? The honest assessment a trusted peer would give you over coffee. <em>(Paid subscribers.)</em></p><p><strong>The Action Brief:</strong> what to do, practically. Not &#8220;assess your risk posture,&#8221; but concrete steps for this week. <em>(Paid subscribers.)</em></p><p>Readable in ten minutes. Dense with insight, zero filler.</p><h2>Who we are</h2><p>Stratsec is not a vendor, a media outlet, or a think tank. We are practitioners.</p><p>Our team and advisory circle includes former Fortune 500 CISOs and CROs, global cybersecurity practice leaders from the Big Four and major systems integrators, founders of what was at the time the world&#8217;s largest offensive security firm, heads of national cybersecurity organisations, intelligence professionals from NATO-aligned nations, and current leaders in quantum security and AI risk.</p><p>When we say &#8220;double down on hygiene,&#8221; it&#8217;s because people who&#8217;ve run programmes at scale know that&#8217;s usually the right answer. When we say &#8220;this is genuinely new,&#8221; it&#8217;s because people who&#8217;ve seen decades of threats know when something is different.</p><h2>What you get</h2><p><strong>Free subscribers</strong> receive The Development section of every item: a properly contextualised, hype-free summary of what happened. Better than what most people get from tech media.</p><p><strong>Paid subscribers</strong> receive the full intelligence product:</p><p>The <strong>Reality Check</strong> and <strong>Action Brief</strong> for every item: whether it matters and what to do about it.</p><p>The <strong>CISO Governance Briefing</strong>: how to register this in your risk framework, what it means for your budget, which policies need revision, where your regulatory exposure has shifted, what skills your team needs, and what your second and third-line functions should be checking.</p><p><strong>What to Tell Your Board</strong>: board-ready language you can use verbatim or adapt for your next risk committee meeting, with a companion slide in Stratsec brand you can drop directly into your deck.</p><p>The <strong>CISO Toolkit</strong>: supplier assurance questions you can send to your critical vendors this week, a team readiness checklist to surface gaps before they become incidents, assurance questions for risk management and internal audit, and a tabletop exercise scenario your incident response team can run without preparation.</p><p><strong>Indicator Watch</strong>: one pre-threat signal Stratsec is monitoring before it becomes a headline.</p><p>Full archive access to every issue.</p><h2>Subscribe</h2><p>The first full issue is in production now. Subscribe to make sure you don&#8217;t miss it.</p><p><em>Emerging technology threats, without the hype.</em></p>]]></content:encoded></item></channel></rss>