LLaMA 4 is Meta's latest open-weight model family. Scout uses a 109B MoE architecture with 17B active parameters, 10M token context window, and native multimodal capabilities. LLaMA 3.3 70B remains a strong general-purpose option.
Deploy LLaMA 4 in minutes
Starting at $0.64/hr on dedicated GPU
| Model | GPU | VRAM | Price | Action |
|---|---|---|---|---|
LLaMA 4 Scout Scout (109B MoE) | A100 80GB PCIe | 80 GB | $1.81/hr | Deploy |
LLaMA 3.3 70B Large (70B) | RTX A6000 | 48 GB | $0.64/hr | Deploy |
Prices include 30% service fee. Billed per minute while running.
Pick your GPU and have it running in minutes. No infrastructure setup required.