Use conversation template for api proxy, fix eventsource format#2383
Use conversation template for api proxy, fix eventsource format#2383zeyugao wants to merge 6 commits intoggml-org:masterfrom
Conversation
Fix eventsource format
|
Thank you! The PHPStorm plugin I was using, codeGPT didn't work with the main branch api_like_OAI.py. With yours it works smoothly! With the new llama-2 based wizard-13b I finally have a usable local-only assistant that integrates seamlessly in my existing workflows. :D |
|
The pr has been merged in upstream, and due to the limitation from GitHub (https://github.com/orgs/community/discussions/5634), it seems that I cannot allow editing by maintainer. |
|
Thank you! The llama-cpp-python generated result is missing some key words |
|
I am observing this error with a 70B Llama 2 model when attempting to run the guidance tutorial notebook and dropping in |
In this pr, it adds a
--chat-prompt-modelparameter enables the use of a model registered in fastchat/conversation.py. As model prompt templates, like llama 2, become more intricate, handling them exclusively with tools such as --chat-prompt and --user-name becomes less manageable. Thus, a community-maintained conversation template has been developed as a more user-friendly solution.Currently, the customized system message is pending the merge of lm-sys/FastChat#2069. Yet, the current fschat version should operate without exceptions.
Furthermore, there exists an issue when presenting data in an event-source format. The data must conclude with two
\ncharacters, rather than just one\n, implying the necessity for a blank line that contains only a single\ncharacter, which is what OpenAI did.