Updated
25 Apr 2026Form Number
LP2311PDF size
24 pages, 1.4 MBAbstract
To generate maximum performance from our Hybrid AI Platforms, Lenovo presents this Software Platform which outlines the recommended software stack for AI workloads. Utilizing a combination of Lenovo’s XClarity One software, NVIDIA’s AI Enterprise Stack, a Red Hat OpenShift-based stack or a fully open-source stack will give enterprises a fast pipeline to standing up a functioning AI Factory with cutting edge performance.
Change History
Changes in the April 25, 2026 update:
- Addition and updates to the following subsections
- Updated Vanilla Kubernetes software stack
- Addition of Red Hat software stack
- Addition of Nutanix software stack for Hybrid AI 221 Platform
Introduction
Lenovo’s solutions encompass the entire technology stack, from infrastructure to tooling, needed to stand up a production-ready AI Factory. This software platform provides an overview of every software and firmware component that is recommended to take full advantage of Lenovo’s Hybrid AI Platforms.
This document does not intend to provide a deep dive into the exact hardware, software, and network configurations recommended to set up an AI cluster. Instead, this document should serve as an overview of the recommended combination of software for AI workloads on a Lenovo Hybrid AI platform. Furthermore, a set of recommendations for how to size and scale Large Language Models (LLMs) on the Hybrid AI platform is provided for once the software stack is applied. For expert guidance on deploying this software platform, see the AI Services section.
For expert guidance on setting up and maintaining this software platform this document provides a description of Lenovo’s deployment and setup services to help organizations deploy their Hybrid AI Factory quickly, bridging expertise gaps where necessary. Different deployment options are available to provide services in a flexible manner.
The Lenovo Hybrid AI Software platform is meant to be used alongside the Lenovo Hybrid AI 221, 285 and 289 platforms. Each of these platforms provides different use cases depending on the organization’s needs. This document lays out a standardized software stack across all of the hardware focused platforms, however depending on the deployment scenario and requirements the configuration can be changed.
The Hybrid AI 221 platform provides a lower cost starting point for enterprises looking to deploy AI infrastructure. Intended for inference-focused workloads on a single node or smaller deployments, the 221 platform provides flexibility to scale up and out into a 285 configuration over time if needed.
The Hybrid AI 285 platform is ideal for organizations that wish to scale their AI platforms as needed, beginning with a minimal starter kit for a small-scale operation with the capability to expand to a full-fledged AI Factory without needing to create a whole new solution. This platform is best suited for organizations wanting to run mostly AI inference workloads that may need the capacity to scale in the future.
The Hybrid AI 289 platform is meant for the heavy-duty AI workloads of tomorrow – scaling, fine-tuning, and serving AI models at large scales. With the 289 platform, enterprises will get maximum usage of the compute power available to them with ultra-high bandwidth networking and top tier GPU performance. This platform is best suited for organizations that need the most performance for both training and inference.123
Requirements
Please refer to the following Software Platform for more details however the stack has been listed below.
- Software Requirements
- Service & Support Offerings
- Hardware Requirements for Full Deployment (Infrastructure Nodes)
Software Requirements
Service & Support Offerings
The following table lists the Service and support offerings
Hardware Requirements for Full Deployment (Infrastructure Nodes)
The following table lists the Scalable Unit deployment
Proxmox or VMWare ESXi can be used as the hypervisor for the VM Nodes above. The VMWare license is outside the scope of this document.
Note: For a Starter Kit deployment, the infrastructure nodes and portions of the software stack are not recommended. Table 1 shows which software components are needed for each deployment type, and the Deployment Sizing Differences section section contains additional details.
AI Software Stack
Deploying AI to production involves implementing multiple layers of software. The process begins with the system management and operating system of the compute nodes, progresses through a workload or container scheduling and cluster management environment, and culminates in the AI software stack that enables delivering AI agents to users.
Vanilla Kubernetes with BCM
NVIDIA Base Command Manager (BCM) streamlines the deployment of Kubernetes clusters by automating the provisioning, configuration, and lifecycle management of GPU-enabled infrastructure, allowing organizations to move quickly from bare metal to production-ready AI platforms. BCM provides integrated support for deploying Kubernetes alongside essential NVIDIA components - such as GPU drivers, the NVIDIA Container Toolkit, and Kubernetes integrations, creating a consistent and validated foundation for containerized AI workloads. It is recommended to use Run:ai with this stack for AI workload scheduling and management. This approach allows data scientists and developers to focus on running and scaling AI workloads rather than assembling and maintaining the underlying infrastructure.
One extra service node is required to run NVIDIA Base Command Manager (BCM).
Red Hat Stack with Validated Patterns
Reduce your time-to-first-token using Lenovo AI Servers and OpenShift with Validated Patterns.
The combination of Lenovo's AI ready servers and Red Hat's Validated Patterns provides a simple, two-step installation to get from bare metal to a running AI application, reducing time and complexity. Step one is to install Red Hat OpenShift using Red Hat Terraform and Redfish to deploy your whole OpenShift HA cluster in a single click on Lenovo AI servers. Step two is to use Red Hat's GitOps based Validated Patterns to deploy OpenShift AI, its dependencies e.g. NVIDIA's GPU Operators, local storage, etc. together with a ready-to-run AI application. Everything you need to start running production ready AI workloads is achieved in a fully automated framework using GitOps.
Nutanix Enterprise AI Stack (221 Only)
Nutanix Enterprise AI (NAI) is a Kubernetes-based application that forms the AI platform layer of the Enterprise AI stack, giving IT teams the ability to deploy, manage, and monitor large language models (LLMs) and inference endpoints. Leveraging the capabilities of NKP, NAI provides the higher-level services required to operationalize generative AI, acting as a centralized inferencing control plane. With support for endpoint APIs from leading LLM providers, including NVIDIA NIM and Hugging Face, organizations can securely run a wide range of generative AI models on-premises or in the public cloud.
NAI includes a streamlined, UI-driven interface, role-based access controls (RBAC), and untethered deployment options for dark-site or air-gapped environments, simplifying Day 2 operations, monitoring, and adaptation of AI models with enterprise-grade resilience and compliance. This approach allows teams to quickly deploy, monitor, and manage AI models and secure endpoints, providing flexibility in model selection and making AI tools accessible across the enterprise, empowering every team to leverage AI effectively.
Visit Hybrid AI with ThinkAgile HX Solution Brief for more info.
Component List
Note that not all software components are required to create a functioning AI solution; however, this is the recommended stack. Please see Table 1 and the Deployment Sizing Differences section for details on the recommended stack for smaller deployments.
In the following sections, we take a deeper dive into the components:
Commonalities
In the following sections, we take a deeper dive into the software elements:
Linux Operating System
For the open source stack, the AI Compute nodes are typically deployed with Ubuntu Server LTS Edition which is a Linux distribution that is maintained for a minimum of 5 years by Canonical as standard, thereby reducing the need for major upgrades. Customers should also consider purchasing additional support for Ubuntu Server LTS using the Ubuntu Pro Edition upgrade. Lenovo Hybrid AI platforms also support Red Hat Enterprise Linux (RHEL) which is distributed as part of a paid licensed model that automatically comes with support. Ultimately, the choice of Linux distributions is one the customer needs to make based on their familiarity with Linux and their ability to support an unlicensed distribution v.s. a licensed distribution with contractual support.
Lenovo XClarity One
Lenovo XClarity One is a management-as-a-service offering for hybrid-cloud management of on-premises data-center assets from Lenovo. Local management hubs can be installed across multiple sites to collect inventory, incidents, and service data, and to provision resources, creating a bridge between devices and the XClarity One portal. The XClarity One portal provides a modern, intuitive interface that centralizes IT orchestration, deployment, automation, and support from edge to cloud, with enhanced visibility into infrastructure performance, usage metering, and analytics.
The following functions are supported by XClarity One:
XClarity One can be installed flexibly, either hosted in the Lenovo cloud with on-premises management hubs or as a fully on-premises solution.
For more information on XClarity One functionality https://pubs.lenovo.com/lxc1/
UEFI Operating Modes for AI Clusters
As a rule, servers being used for cluster management and administration workloads should have their UEFI Operating mode set to one of the ‘Power Efficiency' modes and GPU Servers should have their UEFI Operating Mode set to "Max Performance”.
The UEFI Operating Mode on Lenovo servers can be configured in a variety of ways
- Specified during the Lenovo ordering process and configured at the factory
- Modified on-site using the XCC LXPM tool that is accessed using F1 at boot time
- Using Redfish API for AMD Systems or Intel Systems
Note. The performance of x86 servers can be influenced by the UEFI configuration and recommendations will vary depending on the GPUs, NICs and intended workloads.
Operators
- GPU Operator – Container scheduling on GPUs across server nodes
- Network Operator – Container overlay networking switching and routing with network infrastructure
- NIM Operator – Oversight of NIM to ensure that correct GPU profile and settings are applied for optimal model performance
- Prometheus Operator – Collection of system metrics for monitoring and performance analysis
DOCA-Host - Networking
DOCA-Host is a package in NVIDIA’s DOCA software framework that contains the host drivers and tools necessary to operate BlueField and ConnectX devices. It is available on Linux systems as a standalone package.
DOCA offers 4 installation profiles supporting different deployment scenarios and workload types:
- doca-all
- doca-networking
- doca-ofed
- doca-roce
For BlueField devices it is generally recommended to use doca-all, and for ConnectX it is generally recommended to use doca-networking.
NVIDIA AI Enterprise
NVIDIA AI Enterprise is a comprehensive suite of artificial intelligence and data analytics software designed for optimized development and deployment in enterprise settings. This section outlines some of the remaining tools present in NVIDIA AI Enterprise that are not already discussed in previous sections.

Figure 1. NVIDIA AI Enterprise
Additionally, NVIDIA AI Enterprise provides access to ready-to-use open-sourced containers and frameworks from NVIDIA like NVIDIA NeMo, NVIDIA RAPIDS, NVIDIA TAO Toolkit, NVIDIA TensorRT and NVIDIA Triton Inference Server.
It also provides full access to the NVIDIA NGC catalogue, a collection of tested enterprise software, services and tools supporting end-to-end AI and digital twin workflows and can be integrated with MLOps platforms such as ClearML, Domino Data Lab, Run:ai, UbiOps, and Weights & Biases. An NVIDIA AI Enterprise License provides access to the fully secured, vetted, and tested software artifacts that are supported by NVIDIA.
NVIDIA Inference Microservice
Finally, NVIDIA AI Enterprise introduced NVIDIA Inference Microservices (NIM), a set of performance-optimized, portable microservices designed to accelerate and simplify the deployment of AI models. Those containerized GPU-accelerated pretrained, fine-tuned, and customized models are ideally suited to be self-hosted and deployed on the Lenovo Hybrid AI platforms.
The ever-growing catalog of NIM microservices contains models for a wide range of AI use cases, from chatbot assistants to computer vision models for video processing. The image below shows some of the NIM microservices, organized by use case.

Figure 2. Examples of NIM Catalogue
For an in-depth guide to the NVIDIA Software portfolio with Lenovo Part Numbers, please reference the NVIDIA Software Product Guide on Lenovo Press.
Open Source NVIDIA Stack
In the following sections, we take a deeper dive into the software elements:
NVIDIA Base Command Manager - Provisioning
Base Command Manager (BCM) provisions the AI environment, incorporating components such as the Operating System, Vanilla Kubernetes (K8S), GPU Operator, and Network Operator to manage the AI workloads. BCM Supports 3 types of network topologies depending on how the user wants nodes to be accessed.
- Type 1: All communication is centralized through the head node, providing a controlled and secured gateway
- Type 2: Worker nodes can be accessed directly via a router, so that traffic does not need to go through the head node.
- Type 3: A routed public network is used, where regular nodes are on Internalnet and the head node is on Managementnet.
The following table displays the recommended BCM Networks for a Type 2 network.
Kubernetes Container Orchestration
Kubernetes is the open-source platform for automating container deployment, scaling, and management. In on-prem computing environments, Kubernetes offers a flexible and robust foundation for organizing application workloads, allowing organizations to deploy, scale, and manage containers across their own infrastructure. With built-in support for High Availability, Kubernetes enables resilient deployments that minimize downtime. Its self-healing capabilities automatically detect and respond to failures, replacing or restarting containers as needed to ensure applications remain available and performant. This approach empowers IT teams with full control over cluster configuration and operations, enabling customization to meet specific operational or security needs while reducing reliance on vendor-specific solutions.
Run: ai - Orchestration
NVIDIA Run: ai is a Kubernetes-native orchestration platform that provides GPU allocation, resource management, and AI Lifecycle Integration. Run:ai will allow users to run more workloads by increasing GPU utilization and replaces the default Kubernetes scheduler with a more purpose-built AI scheduler.
Three modes of operation are supported (Saas, Self-Hosted, and Air-Gapped), and for this Software ERA the Saas mode is recommended.
Run: ai must be installed after Containerd and Kubernetes, as well as requiring some pre-configuration of said Kubernetes cluster. The exact requirements can be found on NVIDIAs documentation page.
Deployment Sizing Differences
For an AI Starter Kit deployment, the Kubernetes control plane operates directly on the AI Compute nodes, negating the requirement for dedicated service nodes to run Kubernetes control plane. The AI Compute nodes will function as master-worker nodes. A minimum of one service node is still required to run NVIDIA Base Command Manager (BCM). The software stack for AI Starter Kit or Single Node deployments is similar to the recommended full AI Software Stack, though some components may be considered optional or less practical for these smaller configurations.
For a Single Node deployment, there will be no networking. Therefore, the following components of the Software Stack will not be needed:
- Network Operator
- NVIDIA Network Congestion Control
- NetQ
- Cumulus Linux (No Switch in Single Node Deployment)
- Run:ai
- Lenovo XClarity One
For a deployment of less than 1 SU the following components are not needed but may be added depending on customer needs:
- NetQ
- Run: ai
Red Hat Stack
In the following sections, we take a deeper dive into the software elements:
Red Hat Core OS
Red Hat Enterprise Linux (RHEL) is an enterprise-grade operating system designed to provide the stability, security, and long-term support required for mission-critical applications. OpenShift uses Red Hat CoreOS (RHCOS)—a container-optimized version of RHEL—as its underlying host operating system, enabling automated updates and consistent configuration across the cluster. This tight integration helps simplify lifecycle management while maintaining a secure and reliable platform for modern, cloud-native workloads. Together, RHCOS and OpenShift provide a trusted foundation for running enterprise and AI-driven applications in hybrid and multi-cloud environments.
Red Hat OpenShift
Red Hat OpenShift provides a scalable platform for running GPU-accelerated workloads by enabling organizations to deploy, manage, and scale containerized AI and high-performance computing applications across hybrid and multi-cloud environments. With built-in support for Kubernetes device plugins and certified NVIDIA GPU Operators, OpenShift helps streamline GPU provisioning, driver lifecycle management, and runtime configuration across cluster nodes. This allows data scientists and developers to efficiently run AI training, inference, and data processing workloads while IT teams maintain centralized control over resource allocation and infrastructure consistency. By combining enterprise Kubernetes orchestration with validated GPU enablement, OpenShift helps organizations operationalize AI workloads in a secure and production-ready environment.
OpenShift AI
Red Hat OpenShift AI extends the OpenShift platform with integrated tools and frameworks for building, training, deploying, and monitoring AI/ML models at scale. It provides a curated set of open-source components such as Jupyter notebooks, model serving frameworks, and pipelines within a secure, containerized environment. With native integration into OpenShift’s orchestration and lifecycle management, OpenShift AI enables data scientists and MLOps teams to collaborate efficiently while maintaining enterprise-grade governance and reproducibility. For GPU-accelerated use cases, it supports seamless access to NVIDIA GPUs and leverages certified operators to simplify driver and runtime management. This makes it an ideal foundation for organizations looking to operationalize AI across hybrid cloud environments with consistency and control.
Nutanix Stack
In the following sections, we take a deeper dive into the software elements:
Nutanix Cloud Infrastructure with AHV
Nutanix Cloud Infrastructure (NCI) with AHV provides a scalable, software-defined foundation for running AI and machine learning workloads across on-premises and hybrid cloud environments. By integrating compute, storage, and networking into a unified platform, NCI enables organizations to efficiently provision the infrastructure required to support GPU-accelerated training and inference workloads. The built-in AHV hypervisor allows IT teams to deploy and manage virtualized GPU-enabled environments without the need for third-party virtualization software. This simplifies infrastructure operations while maintaining performance and flexibility for data science and AI application teams. As a result, organizations can support AI initiatives on a secure and resilient platform designed for enterprise-scale deployments.
Nutanix Kubernetes Platform
Nutanix Kubernetes Platform (NKP) enables organizations to deploy and manage containerized AI and data-driven applications across distributed environments. With integrated lifecycle management for Kubernetes clusters, NKP helps streamline the deployment of GPU-accelerated workloads used for AI model training, inference, and data processing. By combining Kubernetes orchestration with Nutanix’s software-defined infrastructure, NKP supports scalable and portable AI workload execution across hybrid cloud environments. Together, NCI and NKP provide a flexible platform for operationalizing AI applications from development through production.
AI Services
The services offered with the Lenovo Hybrid AI platforms are specifically designed to enable broad adoption of AI in the Enterprise. This enables both Lenovo AI Partners and Lenovo Professional Services to accelerate deployment and provide enterprises with the fastest time to production.
The Lenovo AI services offered alongside the Lenovo Hybrid AI platforms enable customers to overcome the barriers they face in realizing ROI from AI investments by providing critical expertise needed to accelerate business outcomes and maximum efficiency. Leveraging Lenovo AI expertise, Lenovo’s advanced partner ecosystem, and industry leading technology we help customers realize the benefits of AI faster. Unlike providers that tie GPU services to proprietary stacks, Lenovo takes a services-first approach, helping enterprises maximize existing investments and scale AI on their own terms.
The two current services offerings broken down to the right of the AI Factory and AI Foundation layers found in the figure below. AI Fast Start Services provide use case development and validation for agentic AI and GenAI applications. The GPU Advance Services provide the foundation needed for AI use case development, including AI factory design, deployment of the software and firmware stack, and setup of orchestration software. Optionally, TruScale RedHat OpenShift service can be added for those wanting to use OpenShift on RHEL. All services are flexible to meet the unique needs of different organizations while adhering to Lenovo's Reference Architectures and Platform Guides. Figure that follows shows some of the possible combinations of software and orchestration that are found in this Reference Architecture.
Customer can choose from these three modular services:
GPU Plan & Design Services
Lenovo offers advisory services to support organizations in planning and optimizing high-performance GPU workloads, including assessment of current infrastructure and identifying intended use cases.
This service helps customers with:
- Planning for optimal GPU utilization
- Aligning business and tech strategy
- Workload assessment
- deployment strategy
- Architecture and design
- Solution sizing and technology selection
- High-level architecture
The outcomes of the GPU Plan & Design Services are reduced risk and access to proven best practices. Improved performance and an optimized infrastructure at the outset build a solid base that is future-proofed to accommodate growth. The following table lists the Service and support offerings.
| Service | Description |
|---|---|
| 5MS7C33028 |
GPU Advanced Services - Plan & Design with Kubernetes |
GPU Configuration and Deployment Services
Configuration and deployment services help organizations accelerate their timeline. By providing installation and setup for the complete Lenovo recommended software stack for AI, this service acts as the engine of the solution, significantly accelerating the time to value.
Lenovo deployment services help provide the fastest time to first token for an enterprise building their Hybrid AI factory. The GPU advanced configuration and deployment services provide expert guidance on software and hardware components to get your AI factory up and running, including:
- Operating system
- Kubernetes
- GPU configuration
- DDN storage configuration
With this service Lenovo enables deployment of the Lenovo Hybrid AI configurations, from a single node to multi-node, with customizable AI software stack and services. Leveraging our deep relationship with NVIDIA, we can fine-tune the GPU performance to precisely match the customer’s workload requirements.
Lenovo enables customers to overcome skills gaps to fully utilize their GPU configurations and boost the performance of their most challenging workloads. Lenovo also helps upskill a diverse customer team with knowledge transfer from Lenovo experts, working within the framework of our scalable, customizable Lenovo Hybrid AI architectures deliver end-to-end solutions designed to accelerate enterprise AI adoption.
Customers will receive a low-level design of the configuration as well as knowledge transfer from Lenovo’s experts.
| Service | Description |
|---|---|
|
5MS7C33029 |
GPU Advanced Services - Configuration with Kubernetes |
GPU Managed Services
Lenovo provides managed NVIDIA-based GPU systems that can address customer needs on an ongoing basis. This includes support, security & compliance, and business functions.
Customers can consistently maintain peak performance of their GPU infrastructure through the following support:
- L1 support for the GPU and NVIDIA AI Enterprise software (if applicable)
- Security and compliance verification
- Ongoing GPU performance monitoring and tuning with logging and alerts
- Backup and restore of the management components and configuration
GPU Managed Services seamlessly scales with the organization and giving greater visibility and monitoring of performance. Additionally these services provide vulnerability patching to stay ahead of risks which helps free up developers and data scientists to focus on innovation. The following table lists the Service and support offerings.
Licensing
Lenovo XClarity One is a for-fee cloud or on-premises application. There is a free trial that allows management of 50 devices for up to 30 days, after which a license must be purchased. The following types of licenses are supported:
- Managed device, per endpoint licenses
- Required for every managed device to use basic monitoring and management in the XClarity One portal. The license is determined based on the total number of managed devices in the organization
- Premium Licenses
Memory Predictive Failure Analytics License: Monitor and analyze memory errors and failure predictions to ensure that your devices are operating at peak performance.
The following table lists XClarity One offerings.
Each NVIDIA H100 (PCIe or NVL) and H200 NVL GPU includes a five-year NVIDIA AI Enterprise Subscription. NVAIE Licensing provides fully secured and tested NVAIE suite of software, as opposed to the open-sourced versions which are not directly supported. The licensing is done on a per-GPU basis. All NVIDIA AI Enterprise subscriptions include NVIDIA Business Standard support, which includes the following:
- Technical Support – 24/7 availability for case filing, 8am-5pm local time support coverage
- Maintenance – Access to maintenance releases, defect resolutions, and security patches
- Direct Support – Access to NVIDIA Support engineering for timely resolution of issues
- Knowledgebase access, web support, email support, phone support
To receive an NVIDIA AI Enterprise license for GPU(s) that do not include it, please contact your Lenovo sales representative. Lenovo provides both licensing and services for NVIDIA AI Enterprise.
The following table lists NVIDIA AI Enterprise offerings.
The following table lists Run:ai subscription offerings.
Ubuntu is an open-source Linux distribution. Version 22.04 is the LTS (Long Term Support) version of Ubuntu that will be maintained until April 2027. Ubuntu Pro can be optionally purchased to include additional weekday or 24/7 support, expanded security maintenance, kernel live patch, among other features.
The following table lists Canonical Ubuntu.
Sizing Guide for NIM LLM
This section aims to recommend the number of pods to deploy and/or GPUs to consume in order to achieve a certain user concurrency and throughput for LLM inference. Keep in mind that the data presented in this section is meant to be an example, and your exact results will heavily depend on the hardware, configuration, and NIM model used.
This guide assumes that an enterprise would aim to achieve a time-to-first-token of less than 1 second, as latency beyond that mark will be unappealing to users.
Concurrent user queries and tokens per second throughput both scale linearly with the number of pods allocated (1 GPU per pod). Based on the data for Meta-Llama 8B Instruct running summarization queries on 1 H100 NVL, the ratio is approximately 350 users per pod. In other words, if you know you need to account for x concurrent user queries at any given time, you should allocate the rounded-up value of pods. The scaling value of 350 is likely to be different depending on the model you are using; however, the linear scaling property is expected to hold across various models if the model can fit on the memory of 1 GPU. Running Gen-AI Perf with the specific model you would like to use will yield the number of concurrent users it can support per GPU, and then the scaling rule of thumb can be applied to approximate the desired number of nodes. Please see some additional data here to view results for other NVIDIA GPUs and NIM LLMs.
One data point to note is that decreasing model precision, for example from bf16 to fp8, does not substantially increase the number of concurrent users, but does slightly increase the throughput. Whether or not the tradeoff is worthwhile depends on the use case. However, if decreasing model precision allows the model to fit in the memory of one GPU, that is likely to lead to a much more significant increase in concurrent user support.
Lenovo LLM Sizing Guide
Although the results above are derived using a specific 8B parameter model, it is likely that AI workloads on Lenovo platforms will involve models of various types and sizes. The size of a model will affect how it can be loaded into the GPUs memory and therefore can change results such as throughput and concurrency. Lenovo has published an LLM Sizing Guide to outline how to size the LLM relative to the GPU, and provide a rule of thumb for the computational requirements of running an LLM.
For a higher-level method of calculating memory requirements automatically and simulating the inferencing process, the site ApX has a useful calculator.
Lenovo AI Center of Excellence
In addition to the choice of utilizing Lenovo EveryScale Infrastructure framework for the Enterprise AI platform to ensure tested and warranted interoperability, Lenovo operates an AI Lab and CoE at the headquarters in Morrisville, North Carolina, USA to test and enable AI applications and use cases on the Lenovo EveryScale AI platform.
The AI Lab environment provides customers and partners a means to execute proof of concepts for their use cases or test their AI middleware or applications. It is configured as a diverse AI platform with a range of systems and GPU options, including NVIDIA L40S and NVIDIA HGX8 H200.
The software environment utilizes Canonical Ubuntu Linux along with Canonical MicroK8s to offer a multi-tenant Kubernetes environment. This setup allows customers and partners to schedule their respective test containers effectively.
Lenovo AI Innovators
Lenovo Hybrid AI platforms offer the necessary infrastructure for a customer’s hybrid AI factory. To fully leverage the potential of AI integration within business processes and operations, software providers, both large and small, are developing specialized AI applications tailored to a wide array of use cases.
To support the adoption of those AI applications, Lenovo continues to invest in and extend its AI Innovators Program to help organizations gain access to enterprise AI by partnering with more than 50 of the industry’s leading software providers.
Partners of the Lenovo AI Innovators Program get access to our AI Discover Labs, where they validate their solutions and jointly support Proof of Concepts and Customer engagements.
LAII provides customers and channel partners with a range of validated solutions across various vertical use cases, such as for Retail or Public Security. These solutions are designed to facilitate the quick and safe deployment of AI solutions that optimally address the business requirements.
The following are a few examples of Lenovo customers implementing an AI solution:
- Kroeger (Retail) – Reducing Customer friction and loss prevention
- Peak (Logistics) – Streamlining supply chain ops for fast and efficient deliveries
- Bikal (AI at Scale) – Delivering shared AI platform for education
- VSAAS (Smart Cities) – Enabling accurate and effective public security
Lenovo Validated Designs
Lenovo Validated Designs (LVDs) are pre-tested, optimized solution designs enabling reliability, scalability, and efficiency in specific workloads or industries. These solutions integrate Lenovo hardware like ThinkSystem servers, storage, and networking with software and best practices to solve common IT challenges. Developed with technology partners such as VMware, Intel, and Red Hat, LVDs ensure performance, compatibility, and easy deployment through rigorous validation.
Lenovo Validated Designs are intended to simplify the planning, implementation, and management of complex IT infrastructures. They provide detailed guidance, including architectural overviews, component models, deployment considerations, and bills of materials, tailored to specific use cases such as artificial intelligence (AI), big data analytics, cloud computing, virtualization, retail, or smart manufacturing. By offering a pretested solution, LVDs aim to reduce risk, accelerate deployment, and assist organizations in achieving faster time-to-value for their IT investments.
Lenovo Hybrid AI platforms act as infrastructure frameworks for LVDs addressing data center-based AI solutions. They provide the hardware/software reference architecture, optionally Lenovo EveryScale integrated solution delivery method, and general sizing guidelines.
AI Discover Workshop
Lenovo AI Discover Workshops help customers visualize and map out their strategy and resources for AI adoption to rapidly unlock real business value. Lenovo’s experts assess the organization’s AI readiness across security, people, technology, and process – a proven methodology – with recommendations that put customers on a path to AI success. With a focus on real outcomes, AI Discover leverage proven frameworks, processes and policies to deliver a technology roadmap that charts the path to AI success.
AI Fast Start
With customers looking to unlock the transformative power of AI, Lenovo AI Fast Start empowers customers to rapidly build and deploy production-ready AI solutions tailored to their needs. Optimized for NVIDIA AI Enterprise and leveraging accelerators like NVIDIA NIMs, Lenovo AI Fast Start accelerates use case development and platform readiness for AI deployment at scale allowing customers to go from concept to production ready deployment in just weeks. Easy to use containerized and optimized inference engines for popular NVIDIA AI Foundation models empower developers to deliver results. AI Fast Start provides access to AI Experts, platforms and technologies supporting onsite and remote models to achieve business objectives.
Bill of Materials
The following table lists Bill of Materials.
Note – Ubuntu Pro and NVIDIA AI Enterprise provide support. The software stack can be deployed without support, however it is recommended to ensure consistent and stable deployment.
Seller training courses
The following sales training courses are offered for employees and partners (login required). Courses are listed in date order.
-
VTT AI: Introducing the Lenovo Hybrid AI 285 Platform with Cisco Networking
2025-11-04 | 36 minutes | Partners Only
DetailsVTT AI: Introducing the Lenovo Hybrid AI 285 Platform with Cisco Networking
The Lenovo Hybrid AI 285 Platform enables enterprises of all sizes to quickly deploy AI infrastructures supporting use cases as either new greenfield environments or as an extension to current infrastructures.
Published: 2025-11-04
This session will describe the hardware architecture changes required to leverage Cisco networking hardware and the Cisco Nexus Dashboard within the Hybrid AI 285 Platform.
Topics include:
- Value propositions for the Hybrid AI 285 platform
- Updates for the Hybrid AI 285 platform
- Leveraging Cisco networking with the 285 platform
- Future plans for the 285 platform
Length: 36 minutes
Course code: DVAI220_PStart the training:
Partner link: Lenovo 360 Learning Center
-
Partner Technical Webinar - Lenovo AI Hybrid Factory offerings
2025-10-13 | 50 minutes | Employees and Partners
DetailsPartner Technical Webinar - Lenovo AI Hybrid Factory offerings
In this 50-minute replay, Pierce Beary, Lenovo Senior AI Solution Manager, review the Lenovo AI Factory offerings for the data center. Pierce showed how Lenovo is simplifying the AI compute needs for the data center with the AI 281, AI 285 and AI 289 platforms based on the NVIDIA AI Enterprise Reference Architecture.
Published: 2025-10-13
Tags: Artificial Intelligence (AI)
Length: 50 minutes
Course code: OCT1025Start the training:
Employee link: Grow@Lenovo
Partner link: Lenovo 360 Learning Center
-
Lenovo VTT Cloud Architecture: Empowering AI Innovation with NVIDIA RTX Pro 6000 and Lenovo Hybrid AI Services
2025-09-18 | 68 minutes | Employees Only
DetailsLenovo VTT Cloud Architecture: Empowering AI Innovation with NVIDIA RTX Pro 6000 and Lenovo Hybrid AI Services
Join Dinesh Tripathi, Lenovo Technical Team Lead for GenAI and Jose Carlos Huescas, Lenovo HPC & AI Product Manager for an in-depth, interactive technical webinar. This session will explore how to effectively position the NVIDIA RTX PRO 6000 Blackwell Server Edition in AI and visualization workflows, with a focus on real-world applications and customer value.
Published: 2025-09-18
We’ll cover:
- NVIDIA RTX PRO 6000 Blackwell Overview: Key specs, performance benchmarks, and use cases in AI, rendering, and simulation.
- Positioning Strategy: How to align NVIDIA RTX PRO 6000 with customer needs across industries like healthcare, manufacturing, and media.
- Lenovo Hybrid AI 285 Services: Dive into Lenovo’s Hybrid AI 285 architecture and learn how it supports scalable AI deployments from edge to cloud.
Whether you're enabling AI solutions or guiding customers through infrastructure decisions, this session will equip you with the insights and tools to drive impactful conversations.
Tags: Industry solutions, SMB, Services, Technical Sales, Technology solutions
Length: 68 minutes
Course code: DVCLD227Start the training:
Employee link: Grow@Lenovo
-
VTT AI: Introducing The Lenovo Hybrid AI 285 Platform with Cisco Networking
2025-08-26 | 54 minutes | Employees Only
DetailsVTT AI: Introducing The Lenovo Hybrid AI 285 Platform with Cisco Networking
Please view this session as Pierce Beary, Sr. AI Solution Manager, ISG ESMB Segment and AI explains:
Published: 2025-08-26
- Value propositions for the Hybrid AI 285 platform
- Updates for the Hybrid AI 285 platform
- Leveraging Cisco networking with the 285 platform
- Future plans for the 285 platform
Tags: Artificial Intelligence (AI), Technical Sales, ThinkSystem
Length: 54 minutes
Course code: DVAI220Start the training:
Employee link: Grow@Lenovo
-
VTT AI: Introducing the Lenovo Hybrid AI 285 Platform April 2025
2025-04-30 | 60 minutes | Employees Only
DetailsVTT AI: Introducing the Lenovo Hybrid AI 285 Platform April 2025
The Lenovo Hybrid AI 285 Platform enables enterprises of all sizes to quickly deploy AI infrastructures supporting use cases as either new greenfield environments or as an extension to current infrastructures. The 285 Platform enables the use of the NVIDIA AI Enterprise software stack. The AI Hybrid 285 platform is the perfect foundation supporting Lenovo Validated Designs.
Published: 2025-04-30
• Technical overview of the Hybrid AI 285 platform
• AI Hybrid platforms as infrastructure frameworks for LVDs addressing data center-based AI solutions.
• Accelerate AI adoption and reduce deployment risks
Tags: Artificial Intelligence (AI), Nvidia, Technical Sales, Lenovo Hybrid AI 285
Length: 60 minutes
Course code: DVAI215Start the training:
Employee link: Grow@Lenovo
Related publications and links
For more information, see these resources:
- Lenovo Hybrid AI 221 Platform Guide:
https://lenovopress.lenovo.com/lp2313-lenovo-hybrid-ai-221-platform-guide - Lenovo Hybrid AI 285 Platform Guide:
https://lenovopress.lenovo.com/lp2181-lenovo-hybrid-ai-285-platform-guide - Lenovo Hybrid AI 289 Platform Guide:
https://lenovopress.lenovo.com/lp2286-lenovo-hybrid-ai-289-platform-guide - NVIDIA Software Product Guide
https://lenovopress.lenovo.com/lp2311-lenovo-hybrid-ai-software-platform - Implementing AI Workloads using NVIDIA GPUs on ThinkSystem Servers:
https://lenovopress.lenovo.com/lp1928-implementing-ai-workloads-using-nvidia-gpus-on-thinksystem-servers - Making LLMs Work for Enterprise Part 3: GPT Fine-Tuning for RAG:
https://lenovopress.lenovo.com/lp1955-making-llms-work-for-enterprise-part-3-gpt-fine-tuning-for-rag - Lenovo to Deliver Enterprise AI Compute for NetApp AIPod Through Collaboration with NetApp and NVIDIA
https://lenovopress.lenovo.com/lp1962-lenovo-to-deliver-enterprise-ai-compute-for-netapp-aipod-nvidia
Trademarks
Lenovo and the Lenovo logo are trademarks or registered trademarks of Lenovo in the United States, other countries, or both. A current list of Lenovo trademarks is available on the Web at https://www.lenovo.com/us/en/legal/copytrade/.
The following terms are trademarks of Lenovo in the United States, other countries, or both:
Lenovo®
ThinkAgile®
ThinkSystem®
XClarity®
The following terms are trademarks of other companies:
AMD is a trademark of Advanced Micro Devices, Inc.
Intel®, the Intel logo is a trademark of Intel Corporation or its subsidiaries.
Linux® is the trademark of Linus Torvalds in the U.S. and other countries.
Terraform® is a trademark of IBM in the United States, other countries, or both.
Other company, product, or service names may be trademarks or service marks of others.
Configure and Buy
Full Change History
Changes in the April 25, 2026 update:
- Addition and updates to the following subsections
- Updated Vanilla Kubernetes software stack
- Addition of Red Hat software stack
- Addition of Nutanix software stack for Hybrid AI 221 Platform
Changes in the November 5, 2025 update:
- Added the following under Requirements section
- XClarity Controller
- Updated figure 1 and figure 4
Changes in the October 15, 2025 update:
- Updated description for tables under AI Services section
First published: October 14, 2025
Course Detail
Employees Only Content
The content in this document with a is only visible to employees who are logged in. Logon using your Lenovo ITcode and password via Lenovo single-signon (SSO).
The author of the document has determined that this content is classified as Lenovo Internal and should not be normally be made available to people who are not employees or contractors. This includes partners, customers, and competitors. The reasons may vary and you should reach out to the authors of the document for clarification, if needed. Be cautious about sharing this content with others as it may contain sensitive information.
Any visitor to the Lenovo Press web site who is not logged on will not be able to see this employee-only content. This content is excluded from search engine indexes and will not appear in any search results.
For all users, including logged-in employees, this employee-only content does not appear in the PDF version of this document.
This functionality is cookie based. The web site will normally remember your login state between browser sessions, however, if you clear cookies at the end of a session or work in an Incognito/Private browser window, then you will need to log in each time.
If you have any questions about this feature of the Lenovo Press web, please email David Watts at dwatts@lenovo.com.
