Content
@
0 reply
0 recast
0 reaction
Jake Casey
@jakeacasey
Hey all, maybe you can help me? Thinking of a small RAG project, but would prefer to run it locally. My idea would be to use langchain, vectordb, but I'm not quite sure on the inference/embeddings part yet. Is there a commonly used way to generate embeddings from documents? Are there local ways to generate them? Thanks for helping a dummy out!
2 replies
0 recast
2 reactions
Rafi
@rafi
You can simplify your stack further by just iterating over all your embeddings. There is no need for vector dbs for single-machine low-volume use case, like yours. Plus, if it’s a learning exercise, you can learn a lot of about matrices :)
1 reply
0 recast
1 reaction
Nicholas Charriere
@pushix
Yes. There are local ways. You need to pick an embedding model and a vector DB. Are you doing python or JavaScript? Your best bet is to find a langchain or llama index guide. There should be some doing exactly this
1 reply
0 recast
1 reaction