Key Takeaways

  • AI pen testing produces false confidence because the system tested in week one is not the system running in week six. Models, system prompts, tool configurations, and non-human identities drift constantly, and the report is stale by the time it lands on the CISO’s desk.
  • A CISO who signs off on an AI deployment based on a point-in-time pen test is approving a system that no longer exists. That reality is worse than not testing at all because it underwrites a sign-off of the risk that has fundamentally changed. 
  • Validation is the closure layer of View, Protect, Validate. View shows what is running. Protect produces the controls. Validate produces the evidence the controls are actually closing risks at runtime.
  • The trust outcome of a working validation program is that the CISO converts from a deployment blocker into a deployment enabler. Evidence is what makes the YES possible.
  • The path forward involves five steps that move the program off “pen-test-and-pray” and onto continuous, runtime-grounded validation tied to the same telemetry plane the detection layer runs on.

The CISO bottleneck in 2026

If you’re a CISO in 2026, you’re probably the slowest part of your company’s AI strategy. Not because you’re slow, but because the rest of the org can stand up new AI features in days, while the security review cycle is still measured in weeks. Product is shipping AI features the platform team has barely had a chance to look at, engineering is wiring up agents to internal systems on a sprint cadence, and customer support is calling models from places nobody documented six months ago. Not to mention, the board asked for an AI plan in February, the platform was in production by April, and now there is a queue of AI deployments waiting on a sign-off and the queue is growing every week. That’s a lot to keep pace with.

This is the bottleneck dynamic, and it is the most uncomfortable position I have been in since I started running security programs more than two decades ago. The job has always required keeping up with the platform, but the platform used to ship in weeks or months, and now it ships in days… or hours. The tooling that gave us our last clean catch-up cycle around 2022 was built for a slower system. We have been trying for the last few years to use it on a faster reality, and the gap shows up most clearly in the way we test.

The way most of us are testing AI deployments today is the same way we tested everything else… we schedule a pen test. The testers get a scope, two weeks, and a kickoff call. They produce a report. We read the report, we sign off, and the deployment goes live (after a retest if necessary). Six months later we do it again.

That sequence used to work reasonably well, but against AI workloads it produces sign-offs that do not reflect the running system. The pen test did not miss something so much as produce confidence that was not warranted, and we built a sign-off chain on top of that confidence.

Why the pen test produced false confidence

I have signed off on AI deployments based on pen-test reports that were already out of date by the time I read them. More than once. The report came in describing the state of a system at a moment two or three weeks earlier, the deployment was waiting on me, and the system in production had moved. The model had been swapped to a newer revision, the system prompt had been revised twice, a new tool had been added to the agent, and the non-human identity backing the workload had picked up a permission for an integration nobody had flagged. None of it was malicious; all of it was the normal velocity of an AI program. And the document I was using to make a yes-or-no decision was a snapshot of a system that did not exist anymore.

That is the structure of the false confidence. The pen test was honest, the methodology was honest, and the testers found what they could find in the scope they were given in the time they were given. What they could not do, because the methodology doesn’t support it, was tell me what the system was doing right now.

For traditional infrastructure, the gap between “what we tested” and “what is running” is small enough that the pen-test snapshot is a reasonable proxy. Stable systems, slow change windows, predictable scopes. AI workloads break that proxy in three places at once. Models are swapped, system prompts are revised, tools are added, identities are re-permissioned, and any one of those changes can open a new attack path the testers never had a chance to see. Rinki walked through the cadence problem and the drift mechanisms in detail in Why Testing AI Like Software Fails, so I won’t re-litigate them here. What I want to bring to your attention is what the methodology does to the sign-off chain when the cadence problem hits an organization that is moving fast.

The sign-off is the part most CISOs don’t openly talk about, but it’s the part that has been keeping me up at night. When I sign off on an AI deployment, I am attesting to my organization that the deployment is reasonably safe to run in production, given the data I have. If the data I have is a snapshot from three weeks ago of a system that has changed materially since, my attestation is about a system I read about, not about the one running in production today. The deployment goes live, the running system drifts further from the snapshot, and the next time I look at a status page I am six months out from when I last had any view of what is actually running.

This is the false-confidence problem put on CISOs, rather than identifying it as a methodology problem. The methodology produced a report that produced a sign-off that produced cover, and the running system was, in some real sense, operating without the security review I had thought I gave it.

That is what I mean by the title. The pen test did not lie in any technical sense; it described a system honestly. By the time the report landed, the system it described had moved on, and we kept making decisions as if it had not.

Validation is the closure layer of View, Protect, Validate

The instinct, when the pen test produces false confidence, is to ask for a better pen test. Tighter scope, better testers, shorter cycle. I have made this argument myself, and I have seen versions of it across four different security organizations. It doesn’t work, because the failure is not in the pen-test execution. The failure is in the methodology being asked to do a job it was never designed to do.

Think of the AI security program in three layers: View, the continuously refreshed inventory of what is running across every model, agent, MCP server, and non-human identity; Protect, the operational layer that catches things in motion through runtime detection, sensitive-data classification, and behavioral analysis of agent actions; and Validate, the evidence layer that runs continuous adversarial testing against the same running system the protection layer is watching, with results that prove which of the controls are actually closing risks and which ones are not.

The validation layer closes that loop. View tells you the surface area, Protect tells you what you have built to defend it, and Validate tells you whether the defenses are working against the system as it is right now rather than against a snapshot of the system as it was last quarter. Without validation, View and Protect amount to an inventory and a set of intentions. With validation, they become a posture you can attest to.

What I want to add to the conversation is the framing move that the Validate layer enables for the CISO role specifically. The pen-test methodology was forcing the CISO to attest to a system using data that did not match it. The validation methodology produces evidence that matches the running system, and lets the CISO attest to what is actually happening rather than to what was once captured.

This is what I mean by the closure layer. View and Protect are the upstream work; Validate is what proves the upstream work is doing what we said it would do, and the proof is what closes the loop on a sign-off that means something.

What counts as validation evidence

Three properties make the validation result trustworthy at the moment of decision: that it runs against the production environment with the production telemetry rather than against a staging clone with a frozen prompt, that the cadence is continuous and defender-set rather than testers’-calendar-set, and that the result lands on the same telemetry plane the detection layer runs on so the validation and detection are speaking the same language about the same system.

Inside the Upwind platform, the capability we built for this is the Upwind Red Agent. The Red Agent runs Exposure Analysis to find the entry points an attacker would use, Attack Path Validation to chain those entry points into reachable paths through the cloud, and Exploitation Scenarios to prove which of those paths actually land. The whole thing runs continuously against production, not against a rehearsal environment, and the results land on the same plane as the AI-DR telemetry the SOC is already watching (the runtime detection mechanics are the subject of Stop Prompt Injection at Runtime: Inside the Multi-Step AI Attack Chain). Naming the capability matters less for product reasons than for making concrete what continuous, runtime-grounded validation actually is when an organization builds it.

What this gives the CISO is evidence that updates as the system updates. When a model is swapped, the next validation pass reflects the swap. When an agent picks up a new tool, the next validation pass tests the path that tool opens. When a non-human identity gets re-permissioned, the next validation pass checks whether the new permission creates a reachable attack path. The evidence is not a snapshot, it is a feed.

That is the difference between a pen-test report and a validation result. The pen-test report describes a moment, and the validation result describes the running system.

From deployment blocker to deployment enabler

Here is what changes for the CISO when the validation result is doing the work the pen-test report used to do. The default answer to a deployment request shifts.

Today, the default is no, or yes-with-conditions, or yes-with-a-caveat-document. The conditions and caveats are how we handle the residual uncertainty the pen-test methodology was leaving on the table. We knew the report was a snapshot, we knew the system was probably moving, so the sign-off ended up qualified. That qualified sign-off is what made the CISO the slowest part of the AI program, the bottleneck I named at the top of this post. Meaning, every deployment becomes a negotiation, every negotiation slows the program, and the security team becomes the team that says no most often.

Validation evidence changes the math. If the running system is producing a continuous stream of evidence that the controls are working, the CISO is not deciding on the basis of a snapshot. The decision becomes “do the controls cover the risks I see right now, given the evidence” rather than “do I trust the methodology that produced the report.” The first question can be answered with confidence. The second question can really only be answered with experience and judgment, which is why pen-test-driven sign-offs always carried a fingerprint of the CISO’s risk tolerance rather than the system’s actual state.

This is the trust outcome that is the real point of the Validate layer. The evidence makes the yes possible. The CISO converts from a deployment blocker into a deployment enabler. Same standards as before, just better data underneath the sign-off, and the yes is built on something. The conversation with the platform team changes from “prove to me this is safe” to “here is what the evidence is showing, and here is what would have to change for me to need to revisit it.”

That conversation is the one we have all wanted to have, and it is the one the pen-test methodology was structurally preventing. Validation evidence is what unblocks it.

Where to start

If you’re a CISO sitting on a queue of AI deployment sign-offs and a pen-test methodology that is producing false confidence, five steps will move you toward a validation-and-evidence model.

  1. Audit your most recent AI pen-test report against the running system. Pull the report for an AI deployment that is already in production. Walk through it section by section and compare each finding to the current state of the system. The goal is to make concrete how much of the report still matches what is running, and to see for yourself the size of the gap the methodology has been producing, not to find new risks.
  2. Name the validation pillar your current testing program is not covering. Three pillars define the AI validation surface: Offensive Testing, Robustness, and Vulnerability Validation. Walk through each one against your existing program. The pillar nobody has been asked to own is the one that is drifting, and it is most often Vulnerability Validation, because the pen-test methodology covers that surface most superficially exactly where the AI-program risk lives.
  3. Run a baseline Vulnerability Validation pass against your current AI estate. A one-time pass against the production system gives you a reachability map of the existing AI risk queue. Most CISOs I have walked through this exercise find that a meaningful fraction of their critical AI risks are not exploitable from any real entry point, and that the ones that are had been ranked below the noise. The baseline lets you re-rank by reachability before you commit to a continuous program.
  4. Tie validation to the same telemetry plane your detection layer is on. Two separate pipelines for validation and detection mean two separate triage queues, two separate sets of context, and two separate places for an attack chain to disappear in transit. One plane means a validated reachable chain shows up in the same SOC queue as a detected in-progress one, with the context to tell them apart.
  5. Reframe the next AI sign-off conversation around evidence, not around report acceptance. When the next deployment comes up, walk into the conversation with a validation result that matches the running system, and make the decision on the basis of what the evidence is showing rather than on the basis of whether the report was credible. The first time you do this is the hardest. The fifth time is the new normal.

Five steps won’t catch every reachable risk on the first pass. They will move the program off pen-test-and-pray and onto validation-and-evidence, and they will start changing the shape of the sign-off conversation in the direction the CISO role actually needs to go in 2026.

The complete framework is in AI Security in 2026: A Field Guide to View, Protect, Validate, dropping this summer 2026.

Get the Field Guide →


Read more from the launch series

Explore the Upwind AI Security Platform →


Jake Martens is Field CISO at Upwind, where he works directly with security leaders on how runtime visibility and continuous validation reshape cloud and AI security programs. He has spent more than three decades in technology and security leadership.