When will “local” LLMs hit?

Given the burst of LLM applications and RAG’s improvement, I am surprised no existing projects build a local, private llm for more secure and proprietary use cases that needs a local llm app with local rag that functions without internet.

Comments URL: https://news.ycombinator.com/item?id=39597455

Points: 1

# Comments: 1

Leave a Reply