What should be a good approach to achieve the context based discussion keeping token limit in mind? #344
naveengujjar29
started this conversation in
General
Replies: 1 comment
-
There are two ways I have found. 1st approach is to pass the last "K" conversations with the context window size in mind. 2nd approach is to make use of summarising the previous discussion to compress the information being passed for tracking the historical context of the discussion. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I have used the spring-ai for RAG on custom data and use the PineCone VectorStore DB.
With my current implementation, I perform the VectorStore similarity search for the provided input by user then pass that context as well in UserMessage but it is causing issue in "Token Limit Exceed" after some time.
What should be the general practice for such a scenario? How are we going to maintain the chat context?
Below code for reference I have used currently.
Beta Was this translation helpful? Give feedback.
All reactions