Put this in the local llama sub but thought I’d share here too!
I found out recently that Amazon/Alexa is going to use ALL users vocal data with ZERO opt outs for their new Alexa+ service so I decided to build my own that is 1000x better and runs fully local.
The stack uses Home Assistant directly tied into Ollama. The long and short term memory is a custom automation design that I’ll be documenting soon and providing for others.
This entire set up runs 100% local and you could probably get away with the whole thing working within / under 16 gigs of VRAM.
💬 Discussion r/LocalLLM (386 points, 33 commentaires) 🔗 Source