Long Term Memory extension #675
Replies: 13 comments 8 replies
-
Wow! It's looks ineteresing and great feature for chat mode! |
Beta Was this translation helpful? Give feedback.
-
@wawawario2 I have a question: So does LTM only work the next time you load chat? Can this function as additional context? Or the only way to use LTM is to keep reloading your character? |
Beta Was this translation helpful? Give feedback.
-
By merging the extension here, you end up losing your ability to easily modify it. Since this is a very complex extension in active development, I encourage you to convert your repository into something that can be installed like this:
I have added a mention to your extension in the wiki: https://github.com/oobabooga/text-generation-webui/wiki/Extensions Do you want me to merge the changes to modules/chat.py? |
Beta Was this translation helpful? Give feedback.
-
@wawawario2 This is great; I haven't looked at the code yet, can you describe how it works? Also, if you open up the issues queue on your repro, I'd gladly test and possibly contribute. Lack of long term memory (be it through context, lora, softprompt, or some other voodoo) is the biggest thing holding these bots back from becoming parts of our lives (at least, for the very bleeding edge early adopters like us). I look forward to this! Let me know if I can help! |
Beta Was this translation helpful? Give feedback.
-
This is so great, this link describes how the LTM module works under the hood. |
Beta Was this translation helpful? Give feedback.
-
Yeah, this is very cool. I'm excited to see where this goes. Thanks for your efforts so far. |
Beta Was this translation helpful? Give feedback.
-
I think that this commit broke the extension: e722c24 The fix should be simple. I will also need to add a "continue" parameter to |
Beta Was this translation helpful? Give feedback.
-
I never realized this extension was the vector embedding storage system. Could I feed an entire github repository with it's various files, and query related lines using the extension? that would be magic, I never really played with chatgpt and its vector embedding storage so I don't really have a feel to its effectiveness for code. |
Beta Was this translation helpful? Give feedback.
-
This is a really great feature, I hope it can be quickly integrated into the code. |
Beta Was this translation helpful? Give feedback.
-
I installed this extension and it worked fine for two days, but then my models started to go crazy. I don't know why, I switched to any model, got irrelevant responses, mixed up the persons in the responses, or the Chronos model put in a specific detail from LTM in one of the responses (with syntax!) |
Beta Was this translation helpful? Give feedback.
-
Hey all, as a heads up this extension is no longer in development. Anyone still using this extension should migrate. I really appreciate all the support over the past few months. |
Beta Was this translation helpful? Give feedback.
-
What is the relationship and differences between RAG ("just read my files") vs long-term memory? Is it merely a coherence issue? |
Beta Was this translation helpful? Give feedback.
-
Hey wawawario, I've really been meaning to get in contact with you to ask a simple question about your extension. I want to know if I edit long_term_memory.db with DataBase Browser for SQLite (as I have been able to), will the embeddings be automatically updated or will it still use the old embeddings? I want to retroactively change memories as we are trying to implement oobabooga in a small business setting with business reports as memories, and it would be much easier than re-submitting all of the memories in order to change the contents of one of them. |
Beta Was this translation helpful? Give feedback.
-
EDIT: As a quick heads up, the repo has been converted to a proper extension, so you no longer have to manage a fork of ooba's repo.
So I'm working on a long-term memory module. Right now, I'm using this UI as a means to field-test it and make improvements, but if there's any interest in merging this module directly into this repo, I can align some of my priorities accordingly (mainly adding some features necessary to make it usable for a general-purpose audience). It's currently a decoupled extension so it should have minimal, if any, impact on the rest of the repo. The current implementation is detailed in the readme. I'm open to any suggestions.
Beta Was this translation helpful? Give feedback.
All reactions