NVIDIA AI Enterprise

NVIDIA AI Enterprise is an end-to-end software platform for developing, deploying, and managing AI applications across cloud, data center, and edge environments. It delivers AI frameworks, NIM microservices, and SDKs through its Application Layer, and GPU drivers, Kubernetes operators, and cluster management tools through its Infrastructure Layer — each with independent release branches and lifecycle — plus enterprise support backed by SLAs.

NVIDIA AI Enterprise consists of two software layers — each with its own release branches and lifecycle — plus built-in enterprise support:

Application Software

Tools for building and deploying AI solutions—generative AI, AI agents, physical AI, and domain-specific applications.

  • NVIDIA NIM and NeMo microservices — Optimized model serving and generative AI deployment
  • Inference and training runtimes — NVIDIA Triton Inference Server with multiple optimized backends
  • Domain-specific SDKs — Speech AI, computer vision, cybersecurity, and more

Application software is distributed through three release branch types: Feature Branch (FB) for latest innovation, Production Branch (PB) for production stability, and Long-Term Support Branch (LTSB) for regulated environments requiring 36 months of API stability.

Infrastructure Software

Tools for deploying, managing, and scaling AI compute resources across bare-metal, virtualized, and containerized environments.

  • Kubernetes operators — GPU Operator, Network Operator, DPU Operator (DPF), and NIM Operator for deployment and lifecycle management
  • GPU and network drivers — Support for bare-metal and virtualized configurations
  • Base Command Manager — Cluster provisioning and infrastructure monitoring

Infrastructure software is distributed through the Infrastructure Branch, which provides regular updates with 1-year support windows and can be designated as a Long-Term Support Branch (LTSB) for extended 3-year support.

AI Enterprise Overview

Enterprise Support

All NVIDIA AI Enterprise subscriptions include enterprise support:

  • Business Standard — Access to NVIDIA AI experts during local business hours, backed by service-level agreements
  • Business Critical (add-on) — 24/7 support coverage
  • Technical Account Manager (add-on) — Dedicated NVIDIA expert for strategic guidance

Getting Started with NVIDIA AI Enterprise

Explains the NVIDIA AI Enterprise platform architecture, covering the infrastructure and application layers, software components, branch types, and support resources.
Guides you through deploying NVIDIA AI Enterprise in 30-60 minutes, from activating your enterprise account to deploying GPU-accelerated containers and running your first AI workload.
Explains the NVIDIA AI Enterprise software lifecycle, including release branch types (Feature, Production, LTS, Infrastructure), their support periods, and how to select the right branch for your deployment needs.

Application Software PB and LTSB Release Documentation

Active Production Branches (PB) and Long-Term Support Branches (LTSB)

Software BranchCompatible Infra ReleaseFirst Planned ReleaseLast Planned ReleasePlanned EOL
Production Branch - October 2025 (PB 25h2)Infra Release 7.xOctober 2025June 2026July 2026
Long-Term Support Branch 2 (LTSB 2)Infra Release 4.x and 7.xNovember 2024August 2027October 2027

All Production Branches (PB) and Long-Term Support Branches (LTSB)

Software BranchCompatible Infra ReleaseFirst Planned ReleaseLast Planned ReleasePlanned EOL
Production Branch - May 2025 (PB 25h1)Infra Release 6.x and 7.xMay 2025December 2025January 2026
Production Branch - October 2024 (PB 24h2)Infra Release 5.x and 6.xOctober 2024June 2025July 2025
Production Branch - May 2024 (PB 24h1)Infra Release 4.x and 5.xMay 2024December 2024January 2025
Production Branch - October 2023 (PB 23h2)Infra Release 3.xOctober 2023June 2024July 2024
Long-Term Support Branch 1 (LTSB 1)Infra Release 1.xAugust 2021February 2024June 2024

Application Layer Software Release Branches

Describes the latest versions of NVIDIA-built and NVIDIA-optimized AI frameworks, NIM microservices, pre-trained models, and SDKs, including monthly release cadence, support timelines, and compatible infrastructure releases.
Lists production-ready AI frameworks and SDKs included in each release, including product versions, release timelines, EOL dates, Government Ready availability, and links to product documentation.
Lists long-term supported AI frameworks and SDKs included in each release, providing 36 months of API stability for highly regulated industries, including product versions, release timelines, EOL dates, and documentation links.

Application Layer Software on NGC

Easy-to-use microservices accelerate generative AI implementation in enterprises. Available exclusively to NVIDIA AI Enterprise Essentials subscribers, backed by NVIDIA's Enterprise support team.

Application Layer Software Documentation

Explains NVIDIA NIM microservices for accelerating foundation model deployment on any cloud or data center, including production-grade runtimes, security updates, deployment guides, API references, and integration options.

Infra Software Release Documentation

The following products are part of NVIDIA AI Enterprise Infrastructure software.

  • Core Infrastructure Drivers
    • NVIDIA GPU Data Center Driver Container
    • NVIDIA Fabric Manager
    • NVIDIA DOCA-OFED Driver for Networking
  • Virtualization
    • NVIDIA vGPU for Compute (Virtual GPU Manager and Guest Drivers)
  • Container Platform
    • NVIDIA Container Toolkit
  • Kubernetes Operators
    • NVIDIA GPU Operator
    • NVIDIA Network Operator
    • NVIDIA DPU Operator (DPF)
    • NVIDIA NIM Operator
  • Data Center Services
    • NVIDIA DOCA Microservices
  • Cluster Management and Orchestration
    • NVIDIA Base Command Manager (BCM)

Infra Layer Release Branches

Active Infra Release Branches

Software BranchDriver BranchLatest Driver in BranchBranchLatest Release in BranchLatest Release Date in BranchBranch EOL
NVIDIA AI Enterprise Infra 7.xR580580.126.09Long-Term Support7.4January 2026July 2028
NVIDIA AI Enterprise Infra 6.xR570570.211.01Feature and Production6.7January 2026March 2026
NVIDIA AI Enterprise Infra 4.xR535535.288.01Long-Term Support4.9January 2026July 2026

All Infra Release Branches

Software BranchDriver BranchLatest Driver in BranchBranchLatest Release in BranchLatest Release Date in BranchBranch EOL
NVIDIA AI Enterprise Infra 5.xR550550.144.02Feature and Production5.3January 2025April 2025
NVIDIA AI Enterprise Infra 3.xR525525.147.05Feature and Production3.3November 2023December 2023
NVIDIA AI Enterprise Infra 2.xR520520.61.05Feature2.3October 2022November 2022
NVIDIA AI Enterprise Infra 1.xR470470.256.02Long-Term Support1.9July 2024September 2024

Infra Layer Software on NGC

Lists all NVIDIA AI Enterprise supported software available on NGC, including AI frameworks, microservices, pre-trained models, SDKs, and tools with enterprise support.
This catalog page lists NVIDIA AI Enterprise Infrastructure collections available on NGC, including infrastructure software components such as drivers, operators, and management tools for deploying and managing AI workloads.

Infra Layer Softare Documentation

Version 7.4 is the latest release.
Version 6.7 is the latest release.
Version 4.9 is the latest release.

Deployment Guides

This document provides guidance for deploying and running NVIDIA AI Enterprise in the Cloud. This resource can be used for understand system pre-requisites, installation and configuration.
This document provides insights into deploying NVIDIA AI Enterprise for VMware vSphere and serves as a technical resource for understanding system pre-requisites, installation, and configuration.
This document provides insights into deploying NVIDIA AI Enterprise on Bare Metal Servers and serves as a technical resource for understanding system pre-requisites, installation, and configuration.
This document provides insights into deploying NVIDIA AI Enterprise with Red Hat OpenShift on bare metal servers. This technical resource can be used for understanding system pre-requisites, installation, and configuration.
This document provides insights into deploying NVIDIA AI Enterprise with Red Hat OpenShift on VMware vSphere. This technical resource can be used for understanding system pre-requisites, installation, and configuration.
This document provides insights into deploying NVIDIA AI Enterprise on Red Hat Enterprise Linux with KVM Virtualization and serves as a technical resource for understanding system prerequisites, installation, and configuration.
NVIDIA RAPIDS Accelerator for Apache Spark enables data engineers to speed up Apache Spark 3 data science pipelines and AI model training while lowering infrastructure costs.
This guide aims to provide guidance on how to set up a high-performance multi-node cluster as virtual machines. Within this guide, you will become familiar with GPUDirect RDMA and ATS while using Docker as the platform for running high-performance multi-node Deep Learning Training. ATS is a VMware PCIe support enhancement in vSphere 7 Update 2. GPUDirect RDMA benefits from ATS and is certified and supported by NVIDIA AI Enterprise.
This solution guide outlines the creation of an AI pipeline on NVIDIA AI Enterprise by leveraging a Natural Language Processing use case example.

Reference Architecture

This reference architecture provides an example deployment of NVIDIA AI Enterprise software suite. It showcases a deployment with VMWare vShpere, and provides example workloads to showcase the platform’s capabilities. Topics such as hardware, network, and workload topologies will be discussed.
This sizing guide is intended to guide customers who want to implement NVIDIA AI Enterprise with NVIDIA-Certified Systems at scale.
This document outlines a reference architecture for a cost-effective, performant Kubernetes-as-a-Service to build out an infrastructure stack for large-scale AI training and inference workload in virtualized environments.

White Papers

The NVIDIA AI Factory for Government Reference Design is a purpose-built, full-stack architecture that enables federal agencies to deploy secure, scalable AI in mission-critical environments. Designed to meet stringent government requirements, it integrates NVIDIA accelerated computing, high-performance networking, NVIDIA-Certified Systems and Storage, Nemotron models, and NVIDIA AI Enterprise government-ready software with a broad partner ecosystem.
This white paper describes how NVIDIA is helping to accelerate the deployment of AI into government and regulated industries by providing a new secure baseline for NVIDIA AI Enterprise software.
This white paper details NVIDIA's commitment to securing the NVIDIA AI Enterprise software stack. It outlines the processes and measures NVIDIA takes to ensure container security.
This whitepaper outlines the essential components, integration strategies, and tools needed for enterprises to deploy robust, single-tenant AI solutions efficiently and securely with the help of ecosystem partners.
This white paper provides detailed guidance on configuring virtual machines (VMs) to support AI/ML workloads when a hypervisor layer is deployed on top of HGX systems.

NVIDIA AI Enterprise Licensing and Support Policies

Covers entitlement, packaging, and licensing for NVIDIA Enterprise. Provides a quick reference to understand the product at a high level, including corresponding SKU information.
Covers support and services for potential and existing enterprise customers. Informational and non-binding reference.
Enterprise Services designed to increase uptime and improve ROI for NVIDIA AI Enterprise, including Support Services, Infrastructure Services, and Education Services.