An upcoming, rack-ready deskside supercomputer utilizing the GB300 chip for trillion-parameter autonomous AI agent deployment.
The HP ZGX Fury AI Station represents a significant shift in AI infrastructure, moving frontier-scale compute from the data center to a deskside, rack-ready form factor. Developed by HP and powered by the NVIDIA Blackwell Ultra (B300) architecture, this system is designed for organizations and researchers who need to deploy trillion-parameter models locally. It bridges the gap between high-end workstations and enterprise server clusters, offering a "deskside supercomputer" experience for teams building autonomous agentic workflows.
In the current market, the ZGX Fury competes in the ultra-high-end tier of AI PCs and Laptops, though its performance profile aligns more closely with specialized AI servers like the NVIDIA DGX series or custom-built Lambda Labs configurations. By utilizing the GB300 Grace Blackwell Ultra Desktop Superchip, HP has prioritized coherent memory and massive bandwidth, specifically targeting the high-token-consumption demands of local AI agents and large-scale inference.
For AI engineers, the most critical specification of the HP ZGX Fury is its 748 GB of unified VRAM. This is facilitated by the NVLink-C2C interconnect, which allows the CPU and GPU to share a high-speed, coherent memory pool. This architecture eliminates the traditional PCIe bottleneck, enabling the system to handle massive datasets and model weights with minimal latency.
The HP ZGX Fury AI Station is one of the few single-node deskside systems capable of running 1-trillion parameter models. This makes it the premier choice for practitioners working with frontier-level research models or massive internal ensembles.
The "sweet spot" for this hardware is running 400B+ parameter models at FP16 or BF16 precision. Unlike consumer hardware that requires 4-bit quantization (EXL2/GGUF) to fit large models, the ZGX Fury allows engineers to maintain maximum model weights for higher reasoning accuracy. Additionally, the massive VRAM allows for long-context tasks (128k+ tokens) without offloading to slower system RAM.
The HP ZGX Fury is not a consumer machine; it is a production-grade tool for specialized AI development.
When evaluating the HP ZGX Fury AI Station for AI development, it is helpful to compare it against the most common alternatives:
The HP ZGX Fury AI Station is the definitive choice for practitioners who need the highest possible VRAM capacity and memory bandwidth available in a deskside format for the next generation of autonomous AI agents.
Llama 4 MaverickMeta | 400B(17B active) | SS | 39.1 tok/s | 146.4 GB | |
| 70B | SS | 50.7 tok/s | 112.8 GB | ||
| 70B | SS | 50.7 tok/s | 112.8 GB | ||
Nvidia Nemotron 3 SuperNVIDIA | 120B(12B active) | SS | 55.2 tok/s | 103.5 GB | |
GLM-5Z.ai | 744B(40B active) | SS | 65.2 tok/s | 87.7 GB | |
GLM-5.1Z.ai | 744B(40B active) | SS | 65.2 tok/s | 87.7 GB | |
Kimi K2.6Moonshot AI | 1000B(32B active) | SS | 66.3 tok/s | 86.2 GB | |
Kimi K2 Instruct 0905Moonshot AI | 1000B(32B active) | SS | 67.6 tok/s | 84.6 GB | |
Kimi K2 ThinkingMoonshot AI | 1000B(32B active) | SS | 67.6 tok/s | 84.6 GB | |
Kimi K2.5Moonshot AI | 1000B(32B active) | SS | 67.6 tok/s | 84.6 GB | |
GLM-4.6Z.ai | 355B(32B active) | SS | 81.3 tok/s | 70.3 GB | |
Mistral Large 3 675BMistral AI | 675B(41B active) | SS | 86.3 tok/s | 66.3 GB | |
DeepSeek-V3DeepSeek | 671B(37B active) | SS | 95.5 tok/s | 59.8 GB | |
DeepSeek-R1DeepSeek | 671B(37B active) | SS | 95.5 tok/s | 59.8 GB | |
DeepSeek-V3.1DeepSeek | 671B(37B active) | SS | 95.5 tok/s | 59.8 GB | |
DeepSeek-V3.2DeepSeek | 685B(37B active) | SS | 95.5 tok/s | 59.8 GB | |
GLM-4.5Z.ai | 355B(32B active) | SS | 110.3 tok/s | 51.8 GB | |
GLM-4.7Z.ai | 358B(32B active) | SS | 108.6 tok/s | 52.6 GB | |
Kimi K2 InstructMoonshot AI | 1000B(32B active) | SS | 110.3 tok/s | 51.8 GB | |
| 70B | SS | 125.1 tok/s | 45.7 GB | ||
Qwen3.5-397B-A17BAlibaba Cloud (Qwen) | 397B(17B active) | SS | 124.2 tok/s | 46.0 GB | |
Llama 2 70B ChatMeta | 70B | SS | 131.7 tok/s | 43.4 GB | |
Mixtral 8x22B InstructMistral AI | 141B(39B active) | SS | 131.2 tok/s | 43.6 GB | |
Qwen 3.5 OmniAlibaba Cloud | 397B(17B active) | SS | 126.5 tok/s | 45.2 GB | |
Qwen3-235B-A22BAlibaba Cloud (Qwen) | 235B(22B active) | SS | 157.3 tok/s | 36.3 GB |