Prime Intellect Docs home page
Search or ask...
Platform
Support
Platform
Search
Navigation
Multi-Node Clusters
Run DistillKit
Documentation
API Reference
Discord
Twitter
Blog
Get Started
Introduction
Quickstart - Deploy a Pod
FAQ
On-Demand Cloud
Run Jupyter Notebooks
Deploy Llama via VLLM
Deploy Custom Docker Image
Multi-Node Clusters
Deploy Multi-Node Cluster
Run Torch FSDP
Run Megatron-Deepspeed
Run Hugging Face Accelerate
Llama 405B Inference in BF16
Run DistillKit
Multi-Node Clusters
Run DistillKit
Distill Llama 405B into smaller models via our Multi-node offering.
Coming Soon…
Llama 405B Inference in BF16