I’ve built a system that lets local LLMs (via Ollama) control self-hosted applications through a multi-agent architecture:

Router agent analyzes requests and delegates to specialized experts 25+ agents for different domains (knowledge bases, DAWs, home automation, git repos) Uses n8n for workflows and MCP servers for integration Works with qwen3, llama3.1, mistral, or any model with function calling

The goal was to create a unified interface to all my self-hosted services that keeps everything local and privacy-focused while still being practical.

Everything’s open-source with full documentation, Docker configs, system prompts, and n8n workflows.

GitHub: dujonwalker/project-nova

I’d love feedback from anyone interested in local LLM integrations with self-hosted services!


💬 Discussion r/LocalLLM (33 points, 0 commentaires)