H: 3.4″ (87.5 mm) x W: 17.6″ (448 mm) x D: 33.5″ (850 mm)
The Powerworks Matrix GZ2 is a purpose-built 2U GPU server designed to meet the demands of modern AI development, machine learning, and scientific computing. Powered by a single AMD EPYC 9555P processor and scalable up to eight NVIDIA L40S GPUs, it provides a flexible foundation for AI model training, inference, and high-throughput data processing in a dense, rack-friendly form factor.
Designed to scale with your models—from prototyping to production. Supports multi-GPU parallelism and AI training pipelines with top-tier throughput and power efficiency.
Ideal for NLP, computer vision, reinforcement learning, and transformer models. The L40S GPUs offer powerful tensor performance and memory capacity for cutting-edge model development.
A perfect fit for service providers offering AI model hosting, fine-tuning, or inference-on-demand across GPU-accelerated workloads.
64 high-efficiency cores designed for high-thread-count AI pipelines, virtualization, and low-latency data processing.
High-speed, error-correcting memory to support large model datasets, AI training batches, and memory-bound applications.
PCIe Gen5 connectivity enables massive compute parallelism. Perfect for training models in-house or offering AI-as-a-service.
Each GPU provides 96GB HBM3e with industry-leading memory bandwidth—ideal for large models and multimodal AI systems.
Redundant, high-efficiency power supplies ensure maximum uptime and stability, even under full GPU load.
Micro-Tower
Height: 14.5″ (368.3mm)
Width: 7.5″ (190.5mm)
Depth: 15.5″ (394mm)