zlacker

[parent] [thread] 2 comments
1. lxgr+(OP)[view] [source] 2024-02-13 18:47:20
I wouldn't be surprised if they essentially just add it to the prompt. ("You are ChatGPT... You are talking to a user that prefers cats over dogs and is afraid of spiders, prefers bullet points over long text...").
replies(1): >>TruthW+k1
2. TruthW+k1[view] [source] 2024-02-13 18:54:53
>>lxgr+(OP)
I think RAG approach with Vector DB is more likely. Just like when you add a file to your prompt / custom GPTs.

Adding the entire file (or memory in this case) would take up too much of the context. So just query the DB and if there's a match add it to the prompt after the conversation started.

replies(1): >>lxgr+H1
◧◩
3. lxgr+H1[view] [source] [discussion] 2024-02-13 18:56:58
>>TruthW+k1
These "memories" seem rather short, much shorter than the average document in a knowledge base or FAQ, for example. Maybe they do get compressed to embedding vectors, though.

I could imagine that once there's too many, it would indeed make sense to classify them as a database, though: "Prefers cats over dogs" is probably not salient information in too many queries.

[go to top]