You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi,
I use the openai llm model in production, and we would like to incorporate conversational history with the ChatML approach, with the User/Assistant alternating streams that openai supports: [{"role": "user", "content": messge1},{"role": "assistant", "content": messge2},{"role": "user", "content": messge3},{"role": "assistant", "content": messge4},]
Such that we could provide a list of dicts and update the sent request by appending before the request with the current prompt
Hi,
I use the openai llm model in production, and we would like to incorporate conversational history with the ChatML approach, with the User/Assistant alternating streams that openai supports:
[{"role": "user", "content": messge1},{"role": "assistant", "content": messge2},{"role": "user", "content": messge3},{"role": "assistant", "content": messge4},]
Such that we could provide a list of dicts and update the sent request by appending before the request with the current prompt
spacy-llm/spacy_llm/models/rest/openai/model.py
Line 129 in 117f689
So I was wondering if that parameter could be added to the OPENAI rest model
The text was updated successfully, but these errors were encountered: