Private AI infrastructure running multiple language models on local hardware. Multi-provider architecture supporting Ollama local models and external APIs through a unified interface.
🔒All inference runs on local hardware. Zero data leaves this machine.