Use this page to verify hardware. Continue to the after these tests when you are ready for on-chain activation and earning.
Prerequisites
Confirm each item before starting:Video transcoding test
What this proves: the orchestrator accepts video segments, transcodes them on the GPU, and delivers HLS output.Video flow summary
The gateway received the RTMP stream, split it into segments, routed them to the orchestrator, the orchestrator transcoded each segment on the GPU, and the gateway reassembled the output as an HLS stream. The-network offchain flag kept the test local and bypassed blockchain interaction.
GPU transcoding works on this machine. Continue to the AI test, or skip to the for production configuration.
AI inference test
The diffusion test requires 24 GB VRAM. For GPUs with 8-16 GB VRAM, skip to the LLM alternative below.
http://llm_runner:8000 as the url value in aiModels.json when you switch to the Ollama-based LLM path.
LLM alternative for 8-16 GB VRAM GPUs
LLM alternative for 8-16 GB VRAM GPUs
The Create a Docker volume and pull a model:Add the LLM entry to For the full LLM pipeline setup including the Docker network configuration, see .
llm pipeline uses the Ollama runner instead of livepeer/ai-runner. It runs quantised LLMs within 8 GB VRAM.Pull the Ollama runner:aiModels.json:AI flow summary
The orchestrator started in off-chain mode with the AI worker enabled. On first start,livepeer/ai-runner was spawned as a child Docker container via Docker-out-of-Docker, downloaded model weights from HuggingFace, and loaded them into GPU VRAM. The test inference request travelled from curl to the orchestrator, through the AI runner container, and returned a generated PNG.
AI inference works on this machine. Model weights remain cached in ~/.lpData/models/ for production use.
Next steps
Setup Guide
Configure for production: on-chain setup, staking, Arbitrum activation, reward calling.
Operator Rationale
Still evaluating? Review the cost-benefit analysis before committing.
Join a Pool
Earn without full solo setup by contributing GPU capacity to a pool.
Workload Options
Compare workloads and determine which pipelines fit your hardware.