This tutorial gets an orchestrator earning from AI inference in under two hours. One GPU, one warm diffusion model, no active set membership needed. Estimated time: 1.5 to 2.5 hours (most of this is model download time - approximately 6 GB). You will verify:
- go-livepeer starts with the AI worker enabled
- The warm model registers at
tools.livepeer.cloud/ai/network-capabilities - A local test inference returns a result
- The node is live on the Livepeer AI network
Prerequisites
Step 1: Install go-livepeer and pull the AI runner
Step 2: Write aiModels.json
aiModels.json declares which AI pipelines and models the node serves, and at what price.
Step 3: Pre-download the model
Download model weights before starting the node. The warm model must be available at startup or the AI worker will attempt to download it on-demand, which delays the first inference.Step 4: Start go-livepeer with AI worker
Step 5: Register AI capabilities
Step 6: Test a local inference
Send a test inference request directly to the orchestrator to confirm the pipeline is serving:test-output.png: PNG image data with a non-zero file size. The PNG is the generated image.
The first inference after startup is slower than later requests because the CUDA kernels warm up. That is normal. From the second request onward, SDXL-Lightning at 512×512 should complete in under 5 seconds.
What happened
The node completed the AI inference path:- go-livepeer started with
-aiWorker- it readaiModels.json, pulled thelivepeer/ai-runnercontainer, mounted the model weights from the host, and loadedByteDance/SDXL-Lightninginto GPU VRAM as a warm model. - Capability advertisement - go-livepeer registered the
text-to-imagepipeline and warm model status on-chain (via the AI Service Registry) and announced it to the network. Gateways that query fortext-to-imagecapability now see this node in their routing pool. - The test inference travelled from the local HTTP call to the AI worker, to the AI runner container, through SDXL-Lightning’s inference pipeline, and back as a PNG. In production, this same path is triggered by a gateway routing an inference request.
- Earning begins when a gateway routes a job to this node. Each completed job sends a PM ticket worth approximately
price_per_unit × pixels_in_outputwei. See for how winning tickets translate to ETH.
Related pages
Add AI to a Video Node
Add AI to an existing video orchestrator without touching the video configuration.
Model Demand Reference
Which models are receiving the most gateway requests and at what price tiers.
Pricing Strategy
Set AI pricing in aiModels.json: wei vs USD notation and competitive positioning.
AI Model Management
Warm vs cold strategy, VRAM allocation, and optimisation flags.