Live VLM WebUI

20 MIN

Real-time Vision Language Model interaction with webcam streaming

SymptomCauseFix
pip install shows "error: externally-managed-environment"Python 3.12+ prevents system-wide pip installsUse virtual environment: python3 -m venv live-vlm-env && source live-vlm-env/bin/activate && pip install live-vlm-webui
Browser shows "Your connection is not private" warningApplication uses self-signed SSL certificateClick "Advanced" → "Proceed to <IP> (unsafe)" - this is safe and expected behavior
Camera not accessible or "Permission Denied"Browser requires HTTPS for webcam accessEnsure you're using https:// (not http://). Accept self-signed certificate warning and grant camera permissions when prompted
"Failed to connect to VLM" or "Connection refused"Ollama or VLM backend not runningVerify Ollama is running with curl http://localhost:11434/v1/models. If not running, start with sudo systemctl start ollama
VLM responses are very slow (>5 seconds per frame)Model too large for available VRAM or incorrect configurationTry a smaller model (gemma3:4b instead of larger models). Increase Frame Processing Interval to 60+ frames. Reduce Max Tokens to 100-200
GPU stats show "N/A" for all metricsNVML not available or GPU driver issuesVerify GPU access with nvidia-smi. Ensure NVIDIA drivers are properly installed
"No models available" in model dropdownAPI endpoint incorrect or models not downloadedVerify API endpoint is http://localhost:11434/v1 for Ollama. Download models with ollama pull gemma3:4b
Server fails to start with "port already in use"Port 8090 already occupied by another serviceStop the conflicting service or use --port flag to specify a different port: live-vlm-webui --port 8091
Cannot access from remote browser on networkFirewall blocking port 8090 or wrong IP addressVerify firewall allows port 8090: sudo ufw allow 8090. Use correct IP from hostname -I command
Video stream is laggy or frozenNetwork issues or browser performanceUse Chrome or Edge browser. Access from a separate PC on the network rather than locally. Check network bandwidth
Analysis results in unexpected languageModel supports multilingual and detected language in promptExplicitly specify output language in prompt: "Answer in English: describe what you see"
pip install fails with dependency errorsConflicting Python package versionsTry installing with --user flag: pip install --user live-vlm-webui
Command live-vlm-webui not found after installBinary path not in PATHAdd ~/.local/bin to PATH: export PATH="$HOME/.local/bin:$PATH" then run source ~/.bashrc
Camera works but no VLM analysis results appear, browser shows InvalidStateErrorAccessing via SSH port forwarding from remote machineWebRTC requires direct network connectivity and doesn't work through SSH tunnels (SSH only forwards TCP, WebRTC needs UDP). Solution 1: Access the web UI directly from a browser on the same network as the server. Solution 2: Use the server machine's browser directly. Solution 3: Use X11 forwarding (ssh -X) to display the browser remotely