Replies: 3 comments 1 reply
-
from this example https://github.com/ggml-org/llama.cpp/blob/master/examples/save-load-state/save-load-state.cpp |
Beta Was this translation helpful? Give feedback.
-
Yes! I might release a version with the support for these functions separately earlier once I can assure the interface for this is stable and won't break in a future version. Keep in mind that if your chat is longer than the context size, then a context shift will have to occur and reevaluate most of the context state either way. |
Beta Was this translation helpful? Give feedback.
-
thank you |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
I want this feature to restore a very long context history.
With normal setChatHistory, especially for my CPU LLM, it will take hours to reload.
But with "llama_state_get_data" and "llama_state_set_data" functions, it will take only few minute or even seconds to reload.
I already try to implement in AddonContext.cpp, I think the concept is work, it can remember something from reload file, but somehow the chat is broke.
I think there are states in javascript part that also need to be saved. But I don't knows what part and how to.
I hope that this project will implement this feature, it really useful for many use cases.
Beta Was this translation helpful? Give feedback.
All reactions