Enabling node metrics
go-livepeer exposes a standard Prometheus endpoint when you pass the-monitor flag at startup.
Enable monitoring flags
http://localhost:7935/metrics
This is on the same port as the go-livepeer CLI (7935 by default). The -monitor flag activates the /metrics path on that port.
Monitoring flags:
Split orchestrator and transcoder setups should pass
-monitor on both processes when both sides need to be scraped. Each process exposes its own /metrics endpoint on its respective CLI port.What metrics are exposed
go-livepeer exposes metrics across multiple categories. The full reference is at Prometheus Metrics Reference. The metrics you will actually act on:Session metrics
Current active sessions, max session capacity, sessions per GPU. These tell you whether your node is at capacity or idle.
Segment metrics
Segments received, transcoded, and failed. Success rate over time is your core transcoding health signal.
Ticket metrics
Winning tickets received and redeemed. A gap between the two indicates ETH balance or redemption issues.
Latency metrics
Processing time per segment. High latency means your GPU is saturated or a pipeline is slow — both affect gateway scoring.
- GPU utilisation (where reported by the NVIDIA driver)
- ETH balance and pending fees
- Round number and reward call status
Option A: Docker monitoring stack (fastest setup)
Livepeer maintains a Docker image that bundles Prometheus, Grafana, and starter dashboard templates. This is the quickest path from zero to a working dashboard.Run the Docker monitoring stack
http://localhost:3000 (default credentials: admin / admin).
Multi-node:
Run the monitoring stack for multiple nodes
prometheus.io/scrape label are discovered automatically without specifying LP_NODES.
livepeer/livepeer-monitoring
Source, Dockerfile, prometheus.yml, and Grafana dashboard templates.
Option B: Custom Prometheus and Grafana
Operators already running a monitoring stack should add go-livepeer as a scrape target:Prometheus scrape target
kill -HUP <prometheus-pid> or use the reload API at http://localhost:9090/-/reload.
Useful Grafana panels to build:
Monitoring AI runner containers
AI inference workloads run in theai-runner Docker container alongside go-livepeer. Monitoring the container separately from the node gives you a faster signal for AI-specific issues.
Check container health:
Check AI runner health
Query registered capabilities
Log capture and verbose output
By default,livepeer sends all logs to stdout only. For long-running production nodes, capture logs to a file:
Capture logs with tee
Enable verbose logging
-v 6 you will see individual segment reception and transcoding activity, which is the fastest way to confirm your node is receiving and processing work without a full Prometheus setup.
Useful log search patterns:
Search logs for key events
journalctl instead:
Inspect systemd logs
Alerting
Prometheus users should add alerting rules for the most impactful failure modes:Prometheus alert rules