LEVIATHAN SYSTEMS
Testing

GPU Infrastructure Commissioning: From Assembly to Production-Ready

Leviathan SystemsPublished 2026-02-159 min read
TL;DR

Complete guide to GPU infrastructure commissioning including POST verification, NVLink validation, thermal testing, and documentation handoff for NVIDIA platforms.

Building GPU infrastructure is only half the battle. Between the moment racks are physically assembled and the point where operations teams can confidently load production workloads lies a critical phase that separates functional hardware from production-ready systems: commissioning.

Commissioning is the systematic validation process that proves every component works as designed, every connection performs to specification, and every system can sustain production loads. It's where theoretical designs meet operational reality, where assembly errors are caught before they become production incidents, and where documentation is created that will support the infrastructure throughout its operational life.

This guide covers the complete commissioning process for GPU infrastructure, from initial power-on tests through final handoff to operations teams.

What Commissioning Validates

Commissioning validates five critical domains that determine whether infrastructure is truly production-ready. Each domain addresses specific failure modes that can't be detected through visual inspection alone.

POST Verification: Confirming Hardware Detection

Power-On Self-Test (POST) is the first validation checkpoint. When a server boots, the system firmware performs hardware enumeration and basic functionality checks. For GPU servers, POST verification confirms several critical elements:

  • All GPUs are detected by the system
  • All NVLink connections show as active
  • GPU firmware versions match specification
  • HBM memory is detected at correct capacity
  • BMC and IPMI interfaces are accessible

Basic POST checks catch obvious hardware failures—missing GPUs, unseated components, firmware mismatches—but they don't validate performance or sustained operation. For deeper validation, POST can be supplemented with NVIDIA Validation Suite (NVVS), which performs more comprehensive hardware checks including memory tests, stress tests, and bandwidth validation.

POST verification must be performed on every server. A single server with undetected GPUs or inactive NVLink connections represents a capacity loss and potential job failure point. The goal is 100% pass rate before proceeding to network validation.

NVLink Domain Validation: Verifying GPU-to-GPU Communication

For platforms like GB200 and GB300 NVL72, where 72 GPUs form a single NVLink domain spanning an entire rack, domain validation is critical. This rack-level check verifies that all 72 GPUs can communicate at the specified bandwidth.

NVLink domain validation catches a specific class of errors: cables that are physically connected but misrouted. A cable plugged into the wrong port may pass visual inspection and even show link-up status, but it breaks the intended topology. The result is degraded bandwidth, asymmetric communication patterns, or complete domain fragmentation.

These errors are detectable through NVVS or NVIDIA Data Center GPU Manager (DCGM). Both tools can perform topology discovery and bandwidth testing across the entire NVLink domain. The validation process typically involves:

  • Topology enumeration to confirm all expected connections are present
  • Bandwidth testing between all GPU pairs to verify performance
  • Error counter checks to identify marginal connections

For NVL72 platforms, this validation is non-negotiable. A single misrouted cable can fragment the domain, turning a 72-GPU system into multiple smaller domains with dramatically reduced collective communication performance. Catching these errors during commissioning prevents discovering them during production training runs.

Network Fabric Validation: Testing Every Connection

Network fabric validation verifies every switch-to-server connection in the infrastructure. This is not a sampling exercise—every single connection must be tested and documented. The validation confirms:

  • Link is up at the correct speed (typically 400GbE or 800GbE)
  • No CRC errors or packet drops are occurring
  • Correct VLAN and partition membership is configured
  • Routing and switching configuration matches design

For fiber optic connections, validation includes Optical Time-Domain Reflectometer (OTDR) testing. OTDR traces provide a permanent record of fiber quality, showing insertion loss, return loss, and any defects along the fiber path. These traces become part of the permanent documentation package.

Network validation also verifies logical configuration. It's not enough for links to be physically up—they must be configured correctly. A server connected to the wrong VLAN or partition can communicate but won't reach the intended resources. These logical errors are often harder to diagnose than physical failures, making commissioning-time validation essential.

The output of network validation is a complete connection map: every port on every switch documented with its connected endpoint, link speed, error counters, and OTDR trace. This documentation becomes the baseline for future troubleshooting.

Liquid Cooling Validation: Thermal Commissioning Under Load

For liquid-cooled platforms, thermal commissioning validates that the cooling system can sustain production workloads. This validation must be performed under load—idle temperatures prove nothing about cooling capacity.

Liquid cooling validation confirms:

  • Coolant flow rate meets specification for each cooling loop
  • Supply and return temperatures are within target range
  • GPU temperatures under sustained compute load stay within thermal envelope
  • No leaks are detected in any connections or components
  • Leak detection sensors are operational and properly configured

The thermal load test typically runs for several hours at maximum GPU utilization. This sustained load reveals issues that wouldn't appear during brief testing: inadequate flow rates, air pockets in cooling loops, marginal quick-disconnect fittings, or insufficient facility cooling capacity.

Temperature monitoring during load testing must cover all critical points: GPU die temperatures, coolant supply and return temperatures, and ambient temperatures around the rack. The goal is to verify that the entire thermal path—from GPU die through cold plate, through coolant loop, through facility heat rejection—can sustain maximum power dissipation.

Leak detection validation is equally critical. Every leak detection sensor must be tested to confirm it triggers alarms correctly. The consequences of undetected coolant leaks in a data center environment are severe enough that sensor validation cannot be skipped or assumed.

Documentation Package: Creating the Operational Record

The commissioning documentation package is not an afterthought—it's a primary deliverable. This documentation supports the infrastructure throughout its operational life, enabling troubleshooting, capacity planning, and future modifications.

A complete documentation package includes:

  • Per-connection test results with OTDR traces for every fiber
  • Cable maps showing port-to-port connectivity for all network connections
  • Rack elevation drawings showing as-built equipment placement
  • POST results for every server with GPU enumeration details
  • Thermal commissioning results for liquid-cooled systems
  • Photographs of each rack showing cable routing and labeling
  • Summary report with pass/fail status for all validation domains

This documentation must be organized for operational use, not just archival storage. When a network connection fails six months after deployment, operations teams need to quickly find the OTDR trace, cable map, and switch port assignment for that specific connection. Documentation structure matters as much as documentation completeness.

Operations Team Role During Commissioning

Commissioning is not something done to infrastructure and then handed over—it's a collaborative process that requires operations team participation. The team that will manage the infrastructure must be involved during commissioning, not just at handoff.

Be Present and Participating

Operations teams should be present during key commissioning activities. This presence serves multiple purposes: it builds familiarity with the infrastructure, enables real-time questions and clarifications, and ensures that operational concerns are addressed before handoff.

Participation doesn't mean operations teams must perform every test—that's the commissioning team's responsibility—but they should observe testing, understand what's being validated, and ask questions about anything unclear.

Review Documentation Before Signing Acceptance

The documentation package should be delivered before final acceptance, not after. Operations teams need time to review documentation, verify it matches their requirements, and identify any gaps or unclear sections.

This review is not a formality. Operations teams should verify that documentation is organized logically, that cable maps are readable and accurate, that test results are complete, and that any exceptions or deviations from design are clearly documented.

Participate in Walkthrough

A physical walkthrough of the infrastructure should be part of commissioning handoff. This walkthrough covers physical layout, cable routing, labeling conventions, cooling system components, and any site-specific considerations.

The walkthrough is where operations teams learn the physical reality of the infrastructure they'll manage. Documentation shows what should be there; the walkthrough shows what actually is there and how to navigate it.

Verify Management Tools Can See Every Device

Before accepting infrastructure, operations teams should verify that their management tools can discover and monitor every device. This includes servers, GPUs, switches, BMCs, and any other managed components.

If management tools can't see a device, that device is effectively invisible to operations. These visibility gaps must be resolved during commissioning, not discovered during production operations.

Flag Concerns Immediately

Operations teams should raise concerns as soon as they're identified, not wait until formal handoff. Early identification of issues—whether technical problems, documentation gaps, or operational concerns—allows resolution while commissioning resources are still on site.

The goal is collaborative problem-solving, not adversarial acceptance testing. Both commissioning and operations teams share the objective of production-ready infrastructure.

What Production-Ready Means

Production-ready is not a subjective assessment—it's a specific set of validated conditions. Infrastructure is production-ready when:

  • Every server passes POST with all GPUs and NVLink connections active
  • Every network connection is tested, documented, and operational
  • Every fiber optic connection has OTDR traces on file
  • Liquid cooling systems are validated under sustained thermal load
  • Complete documentation is delivered and reviewed by operations teams
  • Operations teams have walked through the infrastructure and confirmed they can manage it

These conditions are not negotiable. Infrastructure that meets some but not all of these criteria is not production-ready—it's partially commissioned. Partial commissioning creates operational risk because it's unclear which components are validated and which aren't.

Production-ready also means operations teams have confidence in the infrastructure. This confidence comes from understanding what was tested, seeing the test results, and knowing that any issues discovered during commissioning were resolved, not deferred.

The Cost of Skipping Commissioning

Organizations sometimes consider skipping or abbreviating commissioning to accelerate deployment timelines. This decision trades short-term schedule gains for long-term operational costs.

Without commissioning, assembly errors become production incidents. A misrouted NVLink cable discovered during commissioning is a 30-minute fix. The same cable discovered during a production training run means job failure, troubleshooting time, and rework—all while the infrastructure sits idle.

Without documentation, every troubleshooting session starts from zero. Operations teams must rediscover connection topology, trace cables manually, and guess at as-built configuration. This discovery work happens under pressure, during outages, when time matters most.

Without thermal validation, cooling problems appear under production load, when they're most expensive to address. A cooling system that can't sustain maximum power dissipation means either reduced performance or emergency maintenance—neither is acceptable for production infrastructure.

The cost of commissioning is measured in days or weeks. The cost of skipping commissioning is measured in reduced availability, extended troubleshooting, and operational uncertainty throughout the infrastructure's life.

Commissioning at Scale

Commissioning hundreds or thousands of GPU servers requires process discipline and tooling. Manual testing doesn't scale—automated validation frameworks are essential for large deployments.

Automated commissioning frameworks perform the same validation steps as manual processes but execute them in parallel across many systems. These frameworks collect results centrally, flag failures automatically, and generate documentation programmatically.

Even with automation, human oversight remains critical. Automated tools execute tests and collect data, but humans must interpret results, investigate anomalies, and make acceptance decisions. The goal is to automate execution while preserving human judgment.

Large-scale commissioning also requires clear process definition. When multiple teams are commissioning different sections of infrastructure simultaneously, consistent processes ensure consistent results. Process documentation should define test procedures, acceptance criteria, escalation paths, and documentation requirements.

Leviathan Systems follows this commissioning process on every deployment, regardless of scale. We deliver complete per-connection documentation including OTDR traces for every fiber. We perform POST verification on every server and validate NVLink domains on NVL72 platforms. For liquid-cooled systems, we thermally commission under sustained load to verify cooling capacity. When we hand off infrastructure, production-ready means exactly that—operations teams can load workloads immediately. We've delivered over 1,500 GPU racks to this standard at Meta, Oracle, and xAI facilities.

Ready to Deploy Your GPU Infrastructure?_

Tell us about your project. Book a call and we’ll discuss scope, timeline, and the best approach for your deployment.

Book a Call