Gustech Ventures

GPU Servers

GPU Servers

Discover a kaleidoscope of cutting-edge GPU solutions meticulously crafted to elevate your deep learning endeavours within the dynamic landscape of modern data centres. Our array of optimised offerings ensures unparalleled performance and seamless integration, empowering you to unlock the full potential of your data-driven innovations.

Universal GPU Systems

Modular Building Block Design, Future Proof Open-Standards Based Platform in 4U, 5U, or 8U for Large Scale AI training and HPC Applications

GPU:

NVIDIA HGX H100/A100 4-GPU/8-GPU, AMD Instinct MI300X/MI250 OAM Accelerator, Intel Data Center GPU Max Series

CPU:

Intel® Xeon® or AMD EPYC™

Memory:

Up to 32 DIMMs, 8TB

Drives:

Up to 24 Hot-swap U.2 or 2.5" NVMe/SATA drives

Universal GPU Systems

Modular Building Block Design, Future Proof Open-Standards Based Platform in 4U, 5U, or 8U for Large Scale AI training and HPC Applications

GPU:

NVIDIA HGX H100/A100 4-GPU/8-GPU, AMD Instinct MI300X/MI250 OAM Accelerator, Intel Data Center GPU Max Series

CPU:

Intel® Xeon® or AMD EPYC™

Memory:

Up to 32 DIMMs, 8TB

Drives:

Up to 24 Hot-swap U.2 or 2.5" NVMe/SATA drives

Liquid-Cooled Universal GPU Systems

Direct-to-chip liquid-cooled systems for high-density AI infrastructure at scale.

GPU:

NVIDIA HGX H100/H200/B200 4-GPU/8-GPUs

CPU:

Intel® Xeon® or AMD EPYC™

Memory:

Up to 32 DIMMs, 9TB

Drives:

Up to 24 Hot-swap U.2 or 2.5" NVMe/SATA drives

Liquid-Cooled Universal GPU Systems

Direct-to-chip liquid-cooled systems for high-density AI infrastructure at scale.

GPU:

NVIDIA HGX H100/H200/B200 4-GPU/8-GPUs

CPU:

Intel® Xeon® or AMD EPYC™

Memory:

Up to 32 DIMMs, 9TB

Drives:

Up to 24 Hot-swap U.2 or 2.5" NVMe/SATA drives

4U/5U GPU Lines with PCIe 5.0

Maximum Acceleration and Flexibility for AI/Deep Learning and HPC Applications

GPU:

Up to 10 NVIDIA H100 PCIe GPUs, or up to 10 double-width PCIe GPUs

CPU:

Intel® Xeon® or AMD EPYC™

Memory:

Up to 32 DIMMs, 9TB

Drives:

Up to 24 Hot-swap U.2 or 2.5" NVMe/SATA drives

4U/5U GPU Lines with PCIe 5.0

Maximum Acceleration and Flexibility for AI/Deep Learning and HPC Applications

GPU:

Up to 10 NVIDIA H100 PCIe GPUs, or up to 10 double-width PCIe GPUs

CPU:

Intel® Xeon® or AMD EPYC™

Memory:

Up to 32 DIMMs, 9TB

Drives:

Up to 24 Hot-swap U.2 or 2.5" NVMe/SATA drives

X13 2U/3U Front Loading

Modular Building Block Platform Supporting Today's and Future GPUs, CPUs, and DPUs

GPU:

Up to 4 NVIDIA PCIe GPUs including H100, H100 NVL, and L40S

CPU:

NVIDIA GH200 Grace Hopper™ Superchip, Grace™ CPU Superchip, or Intel® Xeon®

Memory:

Up to 960GB ingegrated LPDDR5X memory (Grace Hopper or Grace CPU Superchip) or 16 DIMMs, 4TB DRAM (Intel)

Drives:

Up to 8 E1.S + 4 M.2 drives

X13 2U/3U Front Loading

Modular Building Block Platform Supporting Today's and Future GPUs, CPUs, and DPUs

GPU:

Up to 4 NVIDIA PCIe GPUs including H100, H100 NVL, and L40S

CPU:

NVIDIA GH200 Grace Hopper™ Superchip, Grace™ CPU Superchip, or Intel® Xeon®

Memory:

Up to 960GB ingegrated LPDDR5X memory (Grace Hopper or Grace CPU Superchip) or 16 DIMMs, 4TB DRAM (Intel)

Drives:

Up to 8 E1.S + 4 M.2 drives

AMD APU Systems

Multi-processor system combining CPU and GPU, Designed for the Convergence of AI and HPC

GPU:

4 AMD Instinct MI300A Accelerated Processing Unit (APU)

CPU:

AMD Instinct™ MI300A Accelerated Processing Unit (APU)

Memory:

Up to 512GB integrated HBM3 memory (4x 128GB)

Drives:

Up to 8 2.5" NVMe or Optional 24 2.5" SATA/SAS via storage add-on card + 2 M.2 drives

AMD APU Systems

Multi-processor system combining CPU and GPU, Designed for the Convergence of AI and HPC

GPU:

4 AMD Instinct MI300A Accelerated Processing Unit (APU)

CPU:

AMD Instinct™ MI300A Accelerated Processing Unit (APU)

Memory:

Up to 512GB integrated HBM3 memory (4x 128GB)

Drives:

Up to 8 2.5" NVMe or Optional 24 2.5" SATA/SAS via storage add-on card + 2 M.2 drives

4U GPU Lines with PCIe 4.0

Flexible Design for AI and Graphically Intensive Workloads, Supporting Up to 10 GPUs

GPU:

NVIDIA HGX A100 8-GPU with NVLink, or up to 10 double-width PCIe GPUs

CPU:

Intel® Xeon® or AMD EPYC™

Memory:

Up to 32 DIMMs, 8TB DRAM or 12TB DRAM + PMem

Drives:

Up to 24 Hot-swap 2.5" SATA/SAS/NVMe

4U GPU Lines with PCIe 4.0

Flexible Design for AI and Graphically Intensive Workloads, Supporting Up to 10 GPUs

GPU:

NVIDIA HGX A100 8-GPU with NVLink, or up to 10 double-width PCIe GPUs

CPU:

Intel® Xeon® or AMD EPYC™

Memory:

Up to 32 DIMMs, 8TB DRAM or 12TB DRAM + PMem

Drives:

Up to 24 Hot-swap 2.5" SATA/SAS/NVMe

2U 2-Node Multi-GPU with PCIe 4

Dense and Resource-saving Multi-GPU Architecture for Cloud-Scale Data Center Applications

GPU:

Up to 3 double-width PCIe GPUs per node

CPU:

Intel® Xeon® or AMD EPYC™

Memory:

Up to 8 DIMMs, 2TB per node

Drives:

Up to 2 front hot-swap 2.5” U.2 per node

2U 2-Node Multi-GPU with PCIe 4

Dense and Resource-saving Multi-GPU Architecture for Cloud-Scale Data Center Applications

GPU:

Up to 3 double-width PCIe GPUs per node

CPU:

Intel® Xeon® or AMD EPYC™

Memory:

Up to 8 DIMMs, 2TB per node

Drives:

Up to 2 front hot-swap 2.5” U.2 per node

GPU Workstations

Flexible Solution for AI/Deep Learning Practitioners and High-end Graphics Professionals

GPU:

Up to 4 double-width PCIe GPUs

CPU:

Intel® Xeon®

Memory:

Up to 16 DIMMs, 6TB

Drives:

Up to 8 hot-swap 2.5” SATA/NVMe

GPU Workstations

Flexible Solution for AI/Deep Learning Practitioners and High-end Graphics Professionals

GPU:

Up to 4 double-width PCIe GPUs

CPU:

Intel® Xeon®

Memory:

Up to 16 DIMMs, 6TB

Drives:

Up to 8 hot-swap 2.5” SATA/NVMe

Universal GPU Systems

Modular Building Block Design, Future Proof Open-Standards Based Platform in 4U, 5U, or 8U for Large Scale AI training and HPC Applications

GPU:

NVIDIA HGX H100/A100 4-GPU/8-GPU, AMD Instinct MI300X/MI250 OAM Accelerator, Intel Data Center GPU Max Series

CPU:

Intel® Xeon® or AMD EPYC™

Memory:

Up to 32 DIMMs, 8TB

Drives:

Up to 24 Hot-swap U.2 or 2.5" NVMe/SATA drives

Universal GPU Systems

Modular Building Block Design, Future Proof Open-Standards Based Platform in 4U, 5U, or 8U for Large Scale AI training and HPC Applications

GPU:

NVIDIA HGX H100/A100 4-GPU/8-GPU, AMD Instinct MI300X/MI250 OAM Accelerator, Intel Data Center GPU Max Series

CPU:

Intel® Xeon® or AMD EPYC™

Memory:

Up to 32 DIMMs, 8TB

Drives:

Up to 24 Hot-swap U.2 or 2.5" NVMe/SATA drives

Liquid-Cooled Universal GPU Systems

Direct-to-chip liquid-cooled systems for high-density AI infrastructure at scale.

GPU:

NVIDIA HGX H100/H200/B200 4-GPU/8-GPUs

CPU:

Intel® Xeon® or AMD EPYC™

Memory:

Up to 32 DIMMs, 9TB

Drives:

Up to 24 Hot-swap U.2 or 2.5" NVMe/SATA drives

Liquid-Cooled Universal GPU Systems

Direct-to-chip liquid-cooled systems for high-density AI infrastructure at scale.

GPU:

NVIDIA HGX H100/H200/B200 4-GPU/8-GPUs

CPU:

Intel® Xeon® or AMD EPYC™

Memory:

Up to 32 DIMMs, 9TB

Drives:

Up to 24 Hot-swap U.2 or 2.5" NVMe/SATA drives

4U/5U GPU Lines with PCIe 5.0

Maximum Acceleration and Flexibility for AI/Deep Learning and HPC Applications

GPU:

Up to 10 NVIDIA H100 PCIe GPUs, or up to 10 double-width PCIe GPUs

CPU:

Intel® Xeon® or AMD EPYC™

Memory:

Up to 32 DIMMs, 9TB

Drives:

Up to 24 Hot-swap U.2 or 2.5" NVMe/SATA drives

4U/5U GPU Lines with PCIe 5.0

Maximum Acceleration and Flexibility for AI/Deep Learning and HPC Applications

GPU:

Up to 10 NVIDIA H100 PCIe GPUs, or up to 10 double-width PCIe GPUs

CPU:

Intel® Xeon® or AMD EPYC™

Memory:

Up to 32 DIMMs, 9TB

Drives:

Up to 24 Hot-swap U.2 or 2.5" NVMe/SATA drives

X13 2U/3U Front Loading

Modular Building Block Platform Supporting Today's and Future GPUs, CPUs, and DPUs

GPU:

Up to 4 NVIDIA PCIe GPUs including H100, H100 NVL, and L40S

CPU:

NVIDIA GH200 Grace Hopper™ Superchip, Grace™ CPU Superchip, or Intel® Xeon®

Memory:

Up to 960GB ingegrated LPDDR5X memory (Grace Hopper or Grace CPU Superchip) or 16 DIMMs, 4TB DRAM (Intel)

Drives:

Up to 8 E1.S + 4 M.2 drives

X13 2U/3U Front Loading

Modular Building Block Platform Supporting Today's and Future GPUs, CPUs, and DPUs

GPU:

Up to 4 NVIDIA PCIe GPUs including H100, H100 NVL, and L40S

CPU:

NVIDIA GH200 Grace Hopper™ Superchip, Grace™ CPU Superchip, or Intel® Xeon®

Memory:

Up to 960GB ingegrated LPDDR5X memory (Grace Hopper or Grace CPU Superchip) or 16 DIMMs, 4TB DRAM (Intel)

Drives:

Up to 8 E1.S + 4 M.2 drives

AMD APU Systems

Multi-processor system combining CPU and GPU, Designed for the Convergence of AI and HPC

GPU:

4 AMD Instinct MI300A Accelerated Processing Unit (APU)

CPU:

AMD Instinct™ MI300A Accelerated Processing Unit (APU)

Memory:

Up to 512GB integrated HBM3 memory (4x 128GB)

Drives:

Up to 8 2.5" NVMe or Optional 24 2.5" SATA/SAS via storage add-on card + 2 M.2 drives

AMD APU Systems

Multi-processor system combining CPU and GPU, Designed for the Convergence of AI and HPC

GPU:

4 AMD Instinct MI300A Accelerated Processing Unit (APU)

CPU:

AMD Instinct™ MI300A Accelerated Processing Unit (APU)

Memory:

Up to 512GB integrated HBM3 memory (4x 128GB)

Drives:

Up to 8 2.5" NVMe or Optional 24 2.5" SATA/SAS via storage add-on card + 2 M.2 drives

4U GPU Lines with PCIe 4.0

Flexible Design for AI and Graphically Intensive Workloads, Supporting Up to 10 GPUs

GPU:

NVIDIA HGX A100 8-GPU with NVLink, or up to 10 double-width PCIe GPUs

CPU:

Intel® Xeon® or AMD EPYC™

Memory:

Up to 32 DIMMs, 8TB DRAM or 12TB DRAM + PMem

Drives:

Up to 24 Hot-swap 2.5" SATA/SAS/NVMe

4U GPU Lines with PCIe 4.0

Flexible Design for AI and Graphically Intensive Workloads, Supporting Up to 10 GPUs

GPU:

NVIDIA HGX A100 8-GPU with NVLink, or up to 10 double-width PCIe GPUs

CPU:

Intel® Xeon® or AMD EPYC™

Memory:

Up to 32 DIMMs, 8TB DRAM or 12TB DRAM + PMem

Drives:

Up to 24 Hot-swap 2.5" SATA/SAS/NVMe

2U 2-Node Multi-GPU with PCIe 4

Dense and Resource-saving Multi-GPU Architecture for Cloud-Scale Data Center Applications

GPU:

Up to 3 double-width PCIe GPUs per node

CPU:

Intel® Xeon® or AMD EPYC™

Memory:

Up to 8 DIMMs, 2TB per node

Drives:

Up to 2 front hot-swap 2.5” U.2 per node

2U 2-Node Multi-GPU with PCIe 4

Dense and Resource-saving Multi-GPU Architecture for Cloud-Scale Data Center Applications

GPU:

Up to 3 double-width PCIe GPUs per node

CPU:

Intel® Xeon® or AMD EPYC™

Memory:

Up to 8 DIMMs, 2TB per node

Drives:

Up to 2 front hot-swap 2.5” U.2 per node

GPU Workstations

Flexible Solution for AI/Deep Learning Practitioners and High-end Graphics Professionals

GPU:

Up to 4 double-width PCIe GPUs

CPU:

Intel® Xeon®

Memory:

Up to 16 DIMMs, 6TB

Drives:

Up to 8 hot-swap 2.5” SATA/NVMe

GPU Workstations

Flexible Solution for AI/Deep Learning Practitioners and High-end Graphics Professionals

GPU:

Up to 4 double-width PCIe GPUs

CPU:

Intel® Xeon®

Memory:

Up to 16 DIMMs, 6TB

Drives:

Up to 8 hot-swap 2.5” SATA/NVMe

Hero background image

Optimising All Connections, Streamlining Every Process

Hero background image

Optimising All Connections, Streamlining Every Process

What’s New At Gustech Ventures