pull down to refresh
21 sats \ 3 replies \ @mrsu 15 Jun 2024 \ parent \ on: Making my local LLM voice assistant faster and more scalable with RAG tech
Probably much sooner than you think. AI is moving pretty quick.
Even for self hosting like this guys doing?
reply
Yes. Its getting much easier. You can easily spin up a back end now without much tech knowledge. You don't even need a GPU (although that speeds things up significantly).
If you're interested, look into Ollama or LM Studio. They provide APIs to interface with the LLMs. Then there are a bunch of clients you can install and point to this endpoint.
reply
Thanks
I’ll have a look
reply