-
Notifications
You must be signed in to change notification settings - Fork 13.4k
server: add warning message for deprecated context field in /api/generate #7878
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
2535ba4 to
3f17495
Compare
3f17495 to
16f353f
Compare
|
Can we revisit this please? @jmorganca said..
How so? Are you talking about this?.. What we need is to be able to create a context array. Then use it to seed with prompts over and over again. This isn't about streaming chat, if that's what you meant. This is about being able to restart the request/response chain from any prior point. For instance, maybe the prompt engineering was off and we want to rephrase the second question after we're 10 deep. Can you demonstrate how we can have 10 turns and then go back up to number 2, use the original context to go in different directions? Can we do that with /api/chat and if so, how? We all would very much like to know how to do that. |
|
Please undo this pull or at least leave the context parameter as is. Using the chat endpoint would, instead, be a lot more inefficient because the LLM would have to re-interpret all the initial part of the chat. By leaving the context parameter as is, context caching is easy to implement and experiments are a lot faster, and the experiments would be repeatable. |
|
My understanding is that If the context parameter is removed from The Is it really the same as: It seems to me that the latter would be far more unstable as the interpretation of vs |
The
contextparameter in/api/generatehas been longtime replaced by functionality in the/api/chatendpoint. This PR adds a deprecation warning in the logs when used.