Andromeda Cluster: 10 exaflops* for startups
- 2,512 H100s on 314 nodes interlinked with 3.2Tbps infiniband
- Available for experiments, training runs, and inference
- You can queue training runs that use the entire cluster, or part of it, or just ssh in
- Store data locally on NAS, or stream it in; no ingress/egress costs
- No minimum duration and superb pricing
- Big enough to train llama 65B in ~10 days
- Total mass of 3,291 kg (GPUs only; not counting chassis, system, rack)
- For use by startup investments of Nat Friedman and Daniel Gross
- Reach out if you want access
* based on 3,958 teraFLOPS per H100 at FP8 with sparsity, your results may vary, consult your doctor before beginning a new training routine