Unparalleled AI Computing Performance on Your Desk
NVIDIA GB300 Grace Blackwell Ultra Desktop Superchip
Up to 775 GB of Coherent Memory
Up to 20 PFLOPS AI Performance
ConnectX-8 SuperNIC
Nvidia DGX OS
Model: FlexPrime V80B - DGX Workstation
The Most Powerful Deskside AI Performance Without Compromise
The Corsair PRO CS-GB300DGX1 Workstation based on NVIDIA DGX™ Station Station delivers unmatched deskside AI power, built on the groundbreaking NVIDIA GB300 Grace Blackwell Ultra Desktop Superchip and 775GB of coherent memory. Designed for large‑scale AI training and inference, it combines state‑of‑the‑art hardware with the NVIDIA AI Software Stack to give teams a turnkey, high‑performance platform for accelerated AI development—right at their deskside.
NVIDIA Grace Blackwell Ultra Desktop Superchip
Powered by the NVIDIA Blackwell Ultra GPU with next‑gen CUDA® and fifth‑gen Tensor Cores, connected to the NVIDIA Grace CPU via NVLink®‑C2C for maximum bandwidth and system performance.
ConnectX-8 SuperNIC
Delivers up to 800Gb/s of high‑efficiency network throughput, providing ultra‑fast connectivity optimized for hyperscale AI workloads and dramatically boosting performance for AI factory environments.
Fifth Generation Tensor Cores
NVIDIA DGX Stations leverage the latest Blackwell‑generation Tensor Cores to enable 4‑bit floating‑point (FP4) AI, allowing larger next‑gen models and higher performance while maintaining accuracy.
NVLink-C2C Interconnect
Extends NVLink to a high‑bandwidth, chip‑to‑chip interconnect between the CPU and GPU, enabling fast, coherent data transfer across processors and accelerators
Workload Optimized Power-Shifting
DGX Stations use AI‑driven system optimizations that dynamically allocate power based on active workloads, continually maximizing performance and efficiency.
Large Coherent Memory for AI
As AI models grow in size and complexity, the NVIDIA Grace Blackwell Ultra architecture enables them to train and run efficiently within a single large coherent memory pool. Its C2C superchip interconnect removes traditional CPU‑GPU bottlenecks, delivering fast, scalable performance for massive‑scale models
Nvidia AI Software Stack
Supports AI software stack supporting fine‑tuning, inference, and data science, enabling seamless development on the desktop and deployment to the cloud or data center using the same tools, libraries, frameworks, and pretrained models.
NVidia DGX OS
Enterprises running production workloads
Workstations with predictable upgrade cycles
AI/ML and data science environments
Cloud and container deployments
















AI Development
With NVIDIA CUDA‑optimized libraries accelerating deep learning and machine learning workloads—and combined with DGX Station’s massive memory and high‑throughput superchip architecture—NVIDIA’s accelerated computing platform delivers a powerful AI development environment for applications ranging from predictive maintenance to medical imaging and natural language processing.
Data Science
With NVIDIA AI software—including RAPIDS™ open‑source libraries—GPUs deliver superior performance and lower infrastructure costs across end‑to‑end data science workflows. DGX Station’s large coherent memory pool enables massive datasets to be loaded directly into memory, eliminating bottlenecks and accelerating overall throughput.
Supercomputing AI and Data Science comes to Desktop with The latest Corsair PRO CS-WS300DGX1
AI Inference
As AI models grow in size and complexity, Corsair PRO DGX Stations accelerate local inference, delivering exceptional performance for large language model (LLM) token generation, data analysis, content creation, AI chatbots, and more.
Personal Cloud
Can function as a high‑performance deskside system for a single user running advanced AI models on local data, or as a shared compute resource for teams fine‑tuning and deploying custom models. With support for NVIDIA Multi‑Instance GPU (MIG), it can be partitioned into up to seven fully isolated GPU instances—each with dedicated memory, cache, and compute cores—and seamlessly scaled to larger MIG environments in the cloud or data center. This allows administrators to deliver consistent QoS across all workloads and extend accelerated computing to every user.








Model: FlexPrime V80B SPECIFICATION
Form Factor Full Tower Workstation Desktop
Processor NVIDIA Grace™ CPU Superchip with 72 Arm® Neoverse V2 cores
GPU Single NVIDIA Blackwell Ultra GPU
Memory Up to 775 GB of Coherent Memory
- CPU Memory Up to 496GB LPDDR5X | Up to 396 GB/s
- GPU Memory Up to 279GB HBM3e | 8 TB/s
NVLink-C2C 900 GB/s
Networking | Peak Bandwidth NVIDIA ConnectX®-8 SuperNIC | Up to 800 Gb/s | Ethernet
Ethernet Ports 2x QSFP 112 Ports with NVIDIA ConnectX-8
1x (RJ45 10GbE (Marvel ACQ113)
1x (1000Base-T dedicated out of band management port (connected to BMC)
MIG 7
PCIe Slots 1x Double-Wide PCIe 5.0 x 16 slots (for display adapter graphics card)
2x PCIe Gen 5 x16 (x8 electrical)
Storage 2TB + 2TB NVMe SSD M.2 Total Storage installed
1x PCIe 5.0 x4 NVMe M.2 ports (from NV Grace CPU)
4x PCIe 5.0 x4 U.2 NVMe bays (Internal, requires NVME HBA)
Server Management BMC Management Module with BMC
Rear I/O 4x USB3.1 Type A ports
1x Mini DP Port
1x 2Ch audio port
1x Micro USB COM port
Cooling | Thermal Solution Double 360 Radiator Closed loop liquid cooling
Targeting < 40dBA sound pressure at maximum workload @ 25 ambient
Power Supply (1) 1600W ATX 80 PLUS Titanium
Decoders 7 NVDEC
7 nv JPEG
Operating System Ubuntu 24.04 LTS












Contact | info
U.S. Headquarter office
115 N. McCarthy Blvd.
Milpitas, CA 95135
Company
