Upwind x NVIDIA: Securing the Next Generation of AI Infrastructure at Runtime
Artificial intelligence is rapidly transforming the enterprise landscape, powering everything from autonomous agents to large-scale LLM applications. However, as organizations adopt AI infrastructure at scale, they face an urgent challenge: ensuring the integrity, safety, and trustworthiness of their AI operations in the face of increasingly sophisticated cyber threats. Moreover, a new set of threats comes to mind as new attack surfaces are created.
As AI becomes the operational backbone of modern enterprises, CISOs and security engineering leaders are facing a dual challenge:
- AI workloads demand exceptional performance and trust.
- They also introduce entirely new attack surfaces, failure modes, and dependency chains.
For this reason, we’re excited to announce a deeper partnership with NVIDIA that directly addresses both problems, bringing runtime-first security to the heart of accelerated AI computing.
Why This Partnership Matters for Security Leaders
AI infrastructure is an ecosystem of GPU accelerators, inference services, orchestration layers, and model-specific pipelines, each introducing unique risks.
Through this partnership, Upwind and NVIDIA are unifying performance, visibility, and protection across AI environments. Together, we provide runtime-first security engineered directly into the systems that enterprises rely on to run AI at scale.
The partnership centers on two capabilities from NVIDIA:
- NVIDIA NIMs (NVIDIA Inference Microservices): powering Upwind’s internal AI-driven security operations, from runtime analytics to large-scale vulnerability correlation and threat modeling.
- NVIDIA Garak: integrated into Upwind’s LLM security validation layer for adversarial testing, jailbreak simulation, and data exfiltration detection.
Together, these capabilities enable continuous validation of AI applications, backed by real runtime context, workload behavior, and API observability from the Upwind platform. In addition to leveraging NVIDIA technology for AI-driven security operations, Upwind also now provides robust security for NVIDIA AI workloads – demonstrating a deep, bidirectional partnership to power the future of AI.
“As AI takes on a central role in business and infrastructure, developers must design systems that are secure from the start. By incorporating NVIDIA’s accelerated computing, advanced AI frameworks, and security-ready infrastructure, Upwind is changing how organizations understand and defend modern cloud environments.”
-Ariel Levanon, Vice President of Cybersecurity, NVIDIA
Securing GPU-Powered AI Workloads
Security leaders are already preparing for a world where AI systems are targeted with model poisoning, inference manipulation, supply chain compromise, or runtime exploitation. This collaboration directly addresses these high-impact scenarios.
Upwind now provides dedicated protection for AI workloads running on NVIDIA GPU-based systems, including the NVIDIA DGX™ and Blackwell™ platforms. That includes:
- Continuous runtime visibility across GPU-powered environments
- Real-time risk prioritization tailored to AI infrastructure
- Zero performance impact on inference or training pipelines
This is powered by a co-engineered framework between Upwind and NVIDIA that delivers five core advantages:
- Enhanced performance through GPU acceleration
- Deployment flexibility across sovereign and private clouds
- Cost-efficient scalability for inference and analytics
- Strict data privacy and locality enforcement
- Tailored engineering aligned to customer-specific AI architectures
As a result, security teams gain a real-time understanding of how AI workloads behave, not just the configuration state. And when an abnormal API call, inference pattern, or process execution emerges, Upwind can surface it instantly with the full runtime context needed to take action.
AI-Driven Security Powered by NVIDIA NIM
NVIDIA NIM is now integrated directly into both Upwind’s platform architecture and our internal engineering workflows, serving as the foundation for deeper runtime context and AI operations.
NIM provides a standardized, containerized framework for deploying high-performance NVIDIA models across any environment: public cloud, private cloud, on-prem GPU clusters, or sovereign regions. By adopting NIM, Upwind can deliver advanced AI-driven security capabilities while remaining cloud-agnostic, operationally flexible, and aligned with the strict compliance requirements of global enterprises.
Within Upwind’s runtime security engine, NIM accelerates the AI components that matter most for real-time protection, such as contextual inference, anomaly analysis, automated reasoning, and correlation across cloud assets, data flows, workloads, and GPU-powered AI systems. This gives security teams the ability to cut noise, detect meaningful signals, and surface advanced threats.
NIM also powers Upwind’s internal AI operations ranging from product research and data processing pipelines to secure model evaluation. By building on standardized NVIDIA components, Upwind can iterate faster, maintain consistent performance across environments, and support stronger data sovereignty controls for customers who require local model execution.
The result is a unified, AI-driven security layer that brings NVIDIA-grade inference and intelligence directly into cloud and AI infrastructure. With NIM, customers gain:
- High-performance AI inference at scale for cloud and GPU environments
- Consistent, portable model deployment across sovereign and regulated workloads
- Enhanced runtime context and automated reasoning for precise risk reduction
- Cloud-agnostic architecture with no vendor lock-in
- Stronger data protection and governance through local model execution
By embedding NVIDIA NIM into the Upwind platform, we’re not just accelerating AI, we’re redefining how enterprises secure cloud and AI systems in real time.
LLM Safety and Security Validated in Real Time
The integration of NVIDIA Garak strengthens one of the most important gaps emerging in enterprise AI programs: validation of LLM robustness against adversarial behavior.
With Garak, Upwind continuously exercises models against attacks such as:
- Prompt injection
- Jailbreak attempts
- Data extraction or exfiltration
- Manipulation of model behavior
Combined with Upwind’s runtime telemetry and API observability, security teams gain a feedback loop for model integrity and “under stress” activities that reflect real-world behavior.
“NVIDIA is setting the foundation for enterprise AI, and we’re proud to both leverage it and secure it. By combining NVIDIA’s accelerated AI infrastructure with Upwind’s runtime visibility and protection, we’re helping organizations deploy AI at scale: safely, efficiently, and with full confidence. Leveraging NVIDIA’s core technologies such as NIM and Garak, Upwind uses AI to deliver superior security outcomes for our customers.”
-Dan Yahav, SVP Platforms, Upwind
Raising the Standard for Trusted AI
This partnership is part of Upwind’s broader AI security strategy, which includes:
- AI workload runtime protection
- AI vulnerability management
- LLM-aware API security
- AI-SPM
Upwind and NVIDIA are moving security closer to where AI actually run on real workloads and GPU systems, in real time. To learn more about the partnership and how Upwind both leverages AI within its platform and also secures AI infrastructure, schedule a demo today.
