zlacker

[parent] [thread] 1 comments
1. TruthW+(OP)[view] [source] 2024-02-13 18:54:53
I think RAG approach with Vector DB is more likely. Just like when you add a file to your prompt / custom GPTs.

Adding the entire file (or memory in this case) would take up too much of the context. So just query the DB and if there's a match add it to the prompt after the conversation started.

replies(1): >>lxgr+n
2. lxgr+n[view] [source] 2024-02-13 18:56:58
>>TruthW+(OP)
These "memories" seem rather short, much shorter than the average document in a knowledge base or FAQ, for example. Maybe they do get compressed to embedding vectors, though.

I could imagine that once there's too many, it would indeed make sense to classify them as a database, though: "Prefers cats over dogs" is probably not salient information in too many queries.

[go to top]