Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I kinda do want to build a local RAG? I want some significant subset of Wikipedia (I assume most people know about these) on a dedicated machine with a RAG front-end. I would have then an offline Wikipedia "librarian" I could query.

But I'm lazy and assumed that someone has already built such a thing. I'm just not aware of this "Wikipedia-RAG-in-a-box".





In this case do you even need a RAG? Most models will have been trained on Wikipedia anyway.

Give Jan (https://www.jan.ai/) a try for instance. You'll need to do a bit of research as to what model will give you the best perf on your system but one of the quantized Llama or Qwen models will probably suit you well.


Thank you.

Standard Ollama probably covers a lot of this



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: