Upwind raises $250M Series B to secure the cloud for the world →
Get a Demo

We know artificial intelligence (AI) services from their models. But AI services also include pipelines, APIs, and cloud-native apps. And those systems need to be secured, even while allowing models to do what they do best — giving and taking data from users. Elsewhere, we’ve looked into Dark AI (using AI to break down systems’ security controls) and AI threat detection (using AI to detect cyberthreats).

Today, we’re going deeper into securing services that run on artificial intelligence, no matter how teams go about securing them. Does it take AI to protect AI? What does it take? Let’s find out.

What are AI Services and What are Their Risks?

First, AI services include cloud-based and on-premises systems that deliver AI capabilities: apps, APIs, or platforms. That could consist of AI-powered search, enterprise tools with embedded AI, or those that manage AI workflows, or public model APIs for natural language generation.

AI services all stem from data, not only for training, but to automate inference, fine-tuning, context injection, and embeddings. Securing the entire data pipeline is therefore the basis for AI security operations and a core component of any data protection strategy. And it needs attention from multiple angles:

So, How Do You Secure AI Services?

The large attack surface represented by all those data streams is a key issue in securing AI. And AI services represent everything from project management to communications, data modeling, reporting, and customer service — use cases that embed AI technology across organizational workflows and ecosystems. Both issues make security a unique and massive undertaking.

According to one global business consultancy in 2024, 78% of companies used AI in some way, up 23% from 2023, with IT representing the largest jump, from 27 to 36%.

It’s also a new undertaking. But while AI is suddenly everywhere, securing AI is still an under-the-radar topic.

Today, protecting AI assets is often just a matter of hoping AI models stay secure under the shelter of existing company cloud security measures. But AI comes with its own risks. So let’s break down each of the risks for artificial intelligence assets. We’ll address AI in general, but specifically GenAI system protection for the specifics of LLMs and other generative models.

Runtime and Container Scanning with Upwind

Upwind offers runtime-powered container scanning features so you get real-time threat detection, contextualized analysis, remediation, and root cause analysis that’s 10X faster than traditional methods.

Get a Demo

Ingestion Security

AI services depend on ingestion pipelines that accept input from APIs, user queries, or third-party sources. 

That means multiple high-risk entry points where attackers can submit malicious inputs that leak sensitive information or manipulate downstream behavior. For instance, payloads might inadvertently include internal fields like email addresses or API keys, which get passed to inference engines. This risk is especially pronounced in multi-tenant systems where shared infrastructure increases the risk of data exposure and unauthorized access.

How can teams secure it? 

Tooling that can help includes API gateways with inspection and throttling, data validation libraries or schema-enforcement middleware, and open-source data loss prevention filters.

How to secure AI Services? In this case, a cybersecurity dashboard displays an alert for a Kubernetes deployment spawning a reverse shell, with a network flow diagram and risk analysis details highlighted on the right side.
CNAPPs can implement behavioral analysis capabilities that highlight abnormal traffic patterns targeting the AI service’s ingestion path.

Embedding and Context Security

AI services often retrieve and embed documents or data into runtime context, including client records, help center articles, or internal PDFs. Without inspection and filtering, sensitive data can be passed into the model. It can also be exposed in outputs. In multi-user systems, embeddings may also leak data across tenants.

Secure it by:

Use vector databases with access control and query audit logs, document pre-processing pipelines well, and employ prompt sanitation or context filtering frameworks. 

Runtime file access monitoring can detect issues like context injection, noting processes like an inference service reading from an unexpected volume.
Runtime file access monitoring can detect issues like context injection, noting processes like an inference service reading from an unexpected volume.

Dataset Integrity

Training and fine-tuning often rely on collected logs, user feedback, third-party files, and developer-uploaded files. Any one of those datasets may not be adequately verified, scanned, and versioned, which means bias, corrupted data, or even data poisoning through adversarial payloads. And model poisoning doesn’t always look like an attack; it often enters through behaviors that appear normal. 

Secure datasets by:

Teams can also implement dataset version control systems, pipeline-integrated dataset scanners, and build-time data provenance validation tools.

[Screenshot] Even though CNAPPs don’t deeply understand datasets, they can protect parts of the environment and the process around training, monitoring the runtime environment where training happens, and spotting misconfigurations, lateral movement risks, and secrets exposure. Is a training container running with root access or talking to an unapproved data store? A CNAPP can help.

Auditing and Lineage

Many AI pipelines operate without end-to-end traceability. That’s a liability for teams that need to prove compliance or respond to incidents. Lineage visibility also improves trust, both internally and externally, by showing the data that guides model decisions.

Protect it by:

Audit logging and SIEM tools can be valuable here, along with model metadata tracking and tamper-resistant artifact storage.

Where GenAI Adds to the Risk

It’s worth noting that generative AI brings new types of challenges, like prompt injection and output misuse, that don’t apply to traditional classifiers or regressors. But it builds on the same foundational risks:

Though protecting AI services involves some care about what models can leak, security best practices are primarily concentrated on protecting what goes into the model. On the other hand, GenAI security adds another attack surface and must contend more concretely with what comes out. That’s not an issue for many traditional resources that need protection.

Runtime Security for AI Services: Observability, Access, and Containment

What comes out? Sometimes it’s private data used in training but not meant to be identified. That’s why securing AI services isn’t solely about inputs or training pipelines, but about what happens after deployment, too. 

Once an LLM or generative model is integrated into customer-facing systems, it acts like other production workloads: it has access to cloud resources, live data, and external systems. And whether running in containers or managed endpoints, it will need observability, identity controls, and network containment.

Why Does Runtime Security for AI Services Matter?

Runtime security is the best lens for seeing into misuse, especially when traditional model monitoring can’t catch complicated behavioral exploits. Here’s why AI and runtime security mesh so well:

The bottom line? AI requires visibility at runtime. Because by the time a prompt triggers suspicious behavior, it’s already left the guardrails of training and entered an organization’s live infrastructure. Here are the critical areas to watch during runtime and the mitigations that can help reduce risks.

CategoryExample: What Can Go WrongWhat to Watch ForHow to Secure It
Identity and AccessThe model’s API is running with admin-level cloud permissions.Who’s using which roles, and any unexpected elevationEnforce least privilege. Use scoped service tokens.
Network BehaviorThe model starts making outbound calls to domains it’s never contacted before.DNS activity and egress traffic from the serviceBlock unknown destinations. Alert on first-time connections.
File AccessThe service reads or writes to sensitive volumes it shouldn’t be touching.File access logs and unexpected volume usageMount only what’s needed. Restrict with read-only access.
Package IntegrityA rogue dependency is loaded at runtime, like a malicious Python library.Libraries loaded at runtime and unverified downloadsUse signed packages. Prevent runtime installs
Execution BehaviorA prompt causes the model to spawn a shell command or trigger a subprocess.System calls, child processes, and abnormal activityUse eBPF monitoring. Sandbox or restrict executable paths.

A runtime-powered CNAPP with an eBPF sensor can handle most of these issues. It maps service behavior, like identity, egress, and file access, in real-time. It captures low-level execution behavior like syscalls, file reads, and spawned processes. It can also enforce image integrity and runtime immutability for better container security. Identity features can audit identity usage, over-privilege, and role assignment. And finally, look for a CNAPP with network security tools that discern which resources have the most traffic — and where it’s coming from.

What’s left? SIEM or XDR tools are complementary, correlating logs, alerts, and runtime signals across environments. And EDR tools focus on endpoints, while CNAPP tools are focused on cloud-native workloads.

The Future of Protecting AI Assets

Right now, most AI cybersecurity conversations revolve around model outputs like jailbreaks, prompt injections, and hallucinated leaks. But that’s just one slide of the risk surface. Organizations today are building decision engines, classifiers, recommendation models, and optimization tools that are deeply embedded into business logic, and while they don’t generate text, they still expose exploitable surfaces. Their protection needs will only get thornier as attackers harness AI and machine learning to penetrate these systems. 

So, what can teams look forward to?

For instance, a customer support bot can be prevented from making decisions that trigger unintended business outcomes, like issuing refunds or approving access.

Current implementations focus on controlling the volume of requests. For instance, companies can put a lid on thousands of applications flooding systems at once. In the future, they’ll also incorporate user behavioral context.

You can absolutely enforce model integrity today, including with CNAPPs to enforce runtime integrity. But AI model registries are still nascent, and tooling for CI/CD pipelines is fragmented. Further, unlike code, model signing isn’t yet standardized.

Instant insights into model transparency are at the heart of problems with AI models, which could be made to purposefully give incorrect answers or take on erroneous decision-making behaviors like incorporating patient zip codes into diagnoses.

In the future, expect AI models with built-in workflows, like a fraud detection model that flags suspicious financial transactions in real time, but instead of blocking or reversing transactions, the output is routed to a human risk analyst.

Upwind is the Foundation of Runtime AI Service Protection

Why? It’s watching more than just runtime behavior, with advanced behavioral analysis to detect abnormal patterns. It’s also watching network traffic by ports and protocols, watching what resources have the most traffic, and where it’s coming from. So if an AI resource starts reaching out to a new domain, assumes an unexpected identity, or triggers unexpected file access, the team will know it in real time.

With visibility into how AI services behave at runtime, teams will be able to catch misuse, misconfigurations, and lateral movement before it spreads. And that means you’ll get the real-time guardrails that traditional ML monitoring tools just don’t cover. To see how Upwind covers AI assets, schedule a demo.

FAQ

What are AI security risks in production?

Once AI models are deployed, they behave like other services, consuming input, processing logic, and interacting with organizational infrastructure. The attack surface of AI shifts from protecting training data to securing the runtime environment, looking at how the model actually behaves with users. The primary risks during this phase are:

Can AI be monitored?

Yes. Monitoring AI  means tracking how the system behaves, what it accesses, and whether it deviates from expected patterns. Monitoring includes:

These aren’t always handled by a single tool today. A full picture usually requires a combination of runtime and security tools (like CNAPPs), but also API gateways, model performance monitors, and logging tools.

How can I secure the entire lifecycle of AI applications?

There’s data ingestion, deployed models, and everything in between. Let’s look at what teams need to do at each step:

How is securing AI different from securing an API?

An AI model can look like any other service endpoint. But its behavior can be much more dynamic and unpredictable than a traditional API, so it requires deeper inspection into logic, execution, and the environment. Here’s why:

Does non-GenAI still need protection?

Yes, even non-generative AI needs protection. Models like classifiers, forecasters, and recommenders can still be manipulated or misconfigured. They also have the power to influence real business decisions, so they’re high-value targets for cyberattackers through decision manipulation, data leakage, shadow model deployments, inference abuse, and workflow triggering. 

Non-generative AI models need protection at the infrastructure level for better security posture. But they also need protection across other layers, like identity and logic, to prevent them from becoming security blind spots.