We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
baichuan2-13B-chat
I encountered the same issue while using `baichuan2-13B-chat`..
I extracted the chat parameters from baichuan2's generation_config.json, and when I call the tgi interface, the result is as follows.
When I invoke the chat method, the result is as follows.
chat
Here are the deployment parameters for the tgi.
--max-batch-prefill-tokens 4096 --max-input-length 4096 --max-total-tokens 4608
Originally posted by @zTaoplus in #981 (comment)
The text was updated successfully, but these errors were encountered:
Successfully merging a pull request may close this issue.
I extracted the chat parameters from baichuan2's generation_config.json, and when I call the tgi interface, the result is as follows.
When I invoke the
chat
method, the result is as follows.Here are the deployment parameters for the tgi.
Originally posted by @zTaoplus in #981 (comment)
The text was updated successfully, but these errors were encountered: