This website uses cookies to anonymously analyze website traffic using Google Analytics.

Instant Clusters

Self-service GPUs for training and inference

Spin up NVIDIA GPU clusters in minutes—Kubernetes or Slurm, high-speed networking included, pay as you go or reserve.

Why Together Instant GPU Clusters?

  • Instant AI Compute

    No approvals, no wait times—deploy high-performance clusters in minutes.

  • Optimized for Distributed AI

    Together Instant GPU Clusters feature NVIDIA Quantum-2 InfiniBand and NVLink networking, ensuring ultra-low-latency, high-throughput performance for large-scale AI workloads.

  • Kubernetes, Slurm, SkyPilot

    Choose between Kubernetes or Slurm for orchestration. With SkyPilot, burst to Together Instant Clusters as a complement to reserved infrastructure.

  • No Limitations on Driver or CUDA Versions

    Users have full control over the software environment, ensuring compatibility with any required driver or CUDA version, without restrictions.

  • No Long-Term Commitments

    Unlike traditional cloud GPU offerings, Instant GPU Clusters enable teams to spin up and down compute resources for short-term projects, eliminating the need for lengthy, upfront commitments.

  • Enterprise-Grade Performance

    Each cluster is built using NVIDIA H100 (80GB SXM) GPUs, engineered for AI training, inference, and fine-tuning at scale.

How it Works

Bare-Metal Performance Built with the NVIDIA NCP reference architecture, Together Instant Clusters provide bare-metal performance for compute, network and storage resources, making them ideal for high performance multi-node for AI training and inference.

End-to-End Cluster Management
From creation to deployment, our software stack streamlines every step including acceptance testing, validation, and installation of Kubernetes or Slurm.

Ultra-Fast Provisioning Get your cluster in minutes with Together’s instant deployment, ensuring quick access to high-performance AI infrastructure.

Instant Clusters Pricing

Ready to use, self-service GPUs.

Price per hour per GPU

Hardware

Hourly

1 - 6 Days

1 Week - 3 Months

NVIDIA HGX H100 Inference

$2.39

$2.00

$1.76

NVIDIA HGX H100 SXM

$2.99

$2.50

$2.20

NVIDIA HGX H200

$3.79

NVIDIA HGX B200

$5.50

Forging the AI Frontier with NVIDIA Reference Architecture

As an NVIDIA Cloud Partner, we’re on the leading frontier of optimizing and operating the deployment of NVIDIA GB200 NVL72 GPU clusters.

Together.aiNvidia Preferred Partner
Learn more

AI Data Centers and Power across North America

Data Center Portfolio

2GW+ in the Portfolio with 600MW of near-term Capacity.

Expansion Capability in Europe and Beyond

Data Center Portfolio

150MW+ available in Europe: UK, Spain, France, Portugal, and Iceland also.

Next Frontiers – Asia and the Middle East

Data Center Portfolio

Options available based on the scale of the projects in Asia and the Middle East.