NVIDIA GPU compatibility, NVENC/NVDEC session limits, driver requirements, and CUDA versions for Livepeer orchestrators.
Livepeer orchestrators use NVIDIA GPUs for video transcoding (NVENC/NVDEC hardware encoders) and AI inference (CUDA cores / Tensor cores). This page covers GPU compatibility, session limits, and driver requirements.
Consumer NVIDIA GPUs enforce a hard limit on concurrent NVENC encoding sessions. This directly limits how many simultaneous transcoding streams your orchestrator can handle per GPU.
The community-maintained nvidia-patch removes the NVENC session limit on consumer GPUs. This is widely used by Livepeer orchestrators and pool operators (Titan Node uses this in their worker setup).
Copy
Ask AI
# Example (Linux) — always check the repo for current instructionsgit clone https://github.com/keylase/nvidia-patch.gitcd nvidia-patchbash patch.sh
Patching the NVIDIA driver modifies a system binary. This is not officially supported by NVIDIA. After driver updates, you must re-apply the patch. Some cloud providers (AWS, GCP) may not allow driver patching on managed GPU instances.
Any supported NVIDIA GPU works. For cost efficiency, an RTX 3060 12GB or RTX 4060 Ti 16GB provides good transcoding throughput at low power draw. Patch the NVENC limit to handle more concurrent sessions.Budget pick: GTX 1660 Super (6 GB) — cheapest entry for transcoding-only.
16 GB VRAM minimum. RTX 4070 Ti Super (16 GB) or RTX 3090 (24 GB) are common choices. 24 GB allows running 2–3 warm AI models alongside transcoding.Best value: RTX 3090 24 GB — widely available used, high VRAM, strong community track record.
24 GB+ VRAM. For LLM inference at reasonable speed, RTX 4090 (24 GB) or data centre cards (A100, L40S). Multiple-GPU setups for serving many warm models.Production pick: A100 40/80 GB or L40S 48 GB in a data centre.