Page is under construction.
Check the github issues for ways to contribute! Or provide your feedback in this quick form
Check the github issues for ways to contribute! Or provide your feedback in this quick form
How it works
Your application sends inference requests to a Gateway. The gateway discovers available Orchestrators, routes your job based on capability, price, and latency, handles retries and auth, and returns results. You never communicate with an orchestrator directly — the gateway handles all of that.| Layer | What it does | Who runs it |
|---|
What you can build
Livepeer AI is designed for streaming and real-time workloads. Strong fits include:Real-time video effects
Style transfer, background replacement, depth overlays, and image-to-image pipelines running frame-by-frame on live video.
Live speech & captions
Live ASR, real-time translation, and caption generation from audio chunks ingested via WebRTC.
Vision pipelines
Object detection, pose estimation, face parsing, segmentation — per-frame GPU inference for live streams.
Custom AI pipelines
Composable multi-step inference workflows via ComfyStream or BYOC: vision → conditioning → generation in sequence.
AI on Livepeer vs other infrastructure
| Livepeer AI | Generic GPU cloud | Hosted AI APIs |
|---|
How models get on the network
Models run inside Orchestrator nodes. Orchestrators can use:- ComfyUI — the most common approach; load
.safetensorsweights, build inference DAGs, serve via ComfyStream - Custom inference servers — any Torch / TensorRT / ONNX model wrapped in a Docker container (BYOC)
image-to-image, depth, style-transfer — not model names. Gateways route by capability and performance, not by which specific model weights are loaded. This means models can be swapped or improved without breaking your application.
How orchestrators host models
Step-by-step guide to running AI models on an orchestrator node
Start here
Is my workload a good fit?
Before building: understand what runs well on Livepeer and what doesn’t.
AI Pipelines
ComfyStream, BYOC, and how to compose multi-step inference workflows.
Model support
Full model family compatibility matrix — which models run on Livepeer and why.