Relay: The Lightweight LLM Proxy for Self-Hosted AI Stacks
Tired of 'proxy fatigue' juggling Ollama, Llama.cpp, and Groq? Relay is a lightweight, config-driven TypeScript proxy that orchestrates heterogeneous LLM backends—like nginx for reasoning. Runs stable for 11+ days.