Page is under construction.
Check the github issues for ways to contribute! Or provide your feedback in this quick form
Check the github issues for ways to contribute! Or provide your feedback in this quick form
Decision tree
Summary: Livepeer works best for streaming, GPU-bound inference with low latency and incremental output. If your workload fails the first two gates, don’t build it on Livepeer.
Capability matrix
| Category | Example workloads | Fit | Why |
|---|
Gateway vs orchestrator responsibilities by workload
Understanding the split between gateway and orchestrator helps you know where to direct integration effort for each workload type.Audio workloads (ASR, translation, intent)
Gateway handles: audio ingestion via WebRTC, chunking and buffering, authentication and retries, output aggregation and fan-out. Orchestrator handles: GPU-resident ASR / translation models, streaming inference execution, incremental token emission, language or model specialisation.Vision workloads (depth, pose, segmentation)
Gateway handles: frame routing, capability selection, latency monitoring, cost-aware routing. Orchestrator handles: vision model execution, GPU memory optimisation, per-frame inference, optional batching.Video workloads (generation, effects, diffusion)
Gateway handles: stream orchestration, QoS and failover, output delivery, session lifecycle management. Orchestrator handles: persistent GPU pipelines, multi-model composition, frame-by-frame generation, real-time conditioning.Text workloads (real-time only)
Gateway handles: request multiplexing, rate limiting, stable API surface. Orchestrator handles: lightweight LLMs or classifiers, prompt routing and control logic, real-time response generation.ASR pipeline examples
These are some of the best-fit workloads on Livepeer today.Live captions for video streams
Multilingual live translation
Voice-driven avatars or agents
Live moderation and safety
What about batch and file-based workloads?
File-to-file and batch workloads: doable vs sensible
File-to-file and batch workloads: doable vs sensible
Livepeer will not block file-based or batch workloads. The protocol is general at the container level — anything that can run in a container can run on a Livepeer orchestrator. But Livepeer’s economics, routing, and reliability are tuned for streaming inference, not batch conversion.The precise rule:
The real constraint:Livepeer’s bottleneck is GPU opportunity cost, not capability. If a job occupies a GPU for a long time without benefiting from low latency, it will lose out to workloads that do. This is by design — gateways will naturally route away from poor-fit workloads.Safe summary:Many batch and file-based AI workloads are technically runnable on Livepeer. However, Livepeer is economically and operationally optimised for streaming, low-latency inference, and such workloads will be routed and priced accordingly.
File-to-file is usually a bad fit — unless the conversion is actually streaming inference in disguise. Livepeer cares about execution shape, not inputs.Your examples, explicitly:YouTube video → MP3 Doable. Bad idea. CPU-bound, no inference, long-running batch job, wastes GPU slots, will be deprioritised by gateways. Technically works. Economically irrational.English → other language (translation)
- File-to-file (text in → text out): batch job, latency-tolerant — weak Livepeer fit.
- Live translation (speech or captions): audio arrives incrementally, translation emitted incrementally, latency matters — excellent Livepeer fit. Same model, different execution shape.
- Upload MP3 → wait → download transcript: marginal. Works, but batch infra is cheaper and gateways gain little from routing it.
- Streamed transcription (even from an MP3): chunk audio, emit tokens continuously, treat it like live audio — strong fit.
| Task | Batch | Streaming |
|---|---|---|
| MP3 → text | ⚠ weak | ✓ strong |
| Translation | ⚠ weak | ✓ strong |
| Video → audio | ✗ weak | ✗ still weak |
| ASR | ⚠ okay | ✓ excellent |