

1·
2 days agoNot on top of my head, but there must be something. llama.cpp and vllm have basically solved the inference problem for LLMs. What you need is a RAG solution on top that also combines it with web search.
Not on top of my head, but there must be something. llama.cpp and vllm have basically solved the inference problem for LLMs. What you need is a RAG solution on top that also combines it with web search.
for coding tasks you need web search and RAG. It’s not the size of the model that matters, since even the largest models find solutions online.
I don’t use Signal because they don’t release the app in F-Droid. Signal devs refuse to release the app outside of Google Play Store, which is very evil.