Skip to main content

Documentation

Documentation/Model Catalog

Model Catalog

Browse our curated selection of AI models. All models come pre-configured with optimal settings and are ready to deploy in minutes.

Text Generation Models

Chat, code generation, and general language tasks. All text models include OpenWebUI for easy interaction.

Qwen 3 30B MoE

Best balance of speed and intelligence. Mixture of Experts architecture.

$0.55/hr

L4 GPU

Mistral 7B

Fast, efficient general-purpose model. Great for chat and code.

$0.55/hr

L4 GPU

DeepSeek R1 8B

Advanced reasoning model with step-by-step thinking.

$0.55/hr

L4 GPU

Gemma 3

Google's latest model. Available in 4B, 12B, and 27B sizes.

$0.55-2.49/hr

L4-A100 GPU

Llama 3.3 70B

Meta's flagship model. Exceptional for complex tasks.

$2.49/hr

A100 GPU

Image Generation Models

Create stunning images from text prompts. All image models use ComfyUI for powerful workflow editing.

Flux Dev

State-of-the-art image generation. Best quality results.

$0.79/hr

A6000 GPU

Flux Schnell

Fast image generation. Great for rapid iteration.

$0.55/hr

L4 GPU

Z Image Turbo

Ultra-fast generation (1-2 seconds). Perfect for demos.

$0.79/hr

A6000 GPU

SDXL

Stable Diffusion XL. Versatile and well-supported.

$0.55/hr

L4 GPU

Qwen Image Gen

Qwen 2.5 VL for image understanding and generation.

$0.79/hr

A6000 GPU

Video Generation Models

Generate videos from text or images. Video models require more GPU power and time per generation.

Wan 2.2 T2V 5B

Text-to-video generation. Good quality, faster processing.

$0.79/hr

A6000 GPU

Wan 2.2 T2V 14B

Higher quality video generation. Recommended for production.

$2.49/hr

A100 GPU

HunyuanVideo

Tencent's video generation model. Excellent quality.

$0.79/hr

A6000 GPU

GPU Reference

GPUVRAMPriceBest For
NVIDIA L424 GB$0.55/hrSmall-medium models, fast inference
NVIDIA RTX A600048 GB$0.79/hrLarge image models, Flux, video
NVIDIA A10080 GB$2.49/hr70B+ text models, high-quality video
NVIDIA H10080 GB$4.49/hrMaximum performance, fastest inference

Ready to deploy?

Check out our Quick Deploy Guide to get started, or learn about ComfyUI and OpenWebUI interfaces.

Deploy a Model