We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
功能描述 / Feature Description 是否支持计算传给模型的引用文本、提示词、用户问题等内容的token数量,并且在超出模型能接受的最大token的时候进行动态限制,防止报错: 请求内容超出模型最大token数量
解决的问题 / Problem Solved 解决 请求内容超出模型最大token数量的报错
实现建议 / Implementation Suggestions
替代方案 / Alternative Solutions
其他信息 / Additional Information
The text was updated successfully, but these errors were encountered:
动态限制? 如果你是直接从xinfer三方模型服务平台接进来的 那么你要从chatchat报错里改了
Sorry, something went wrong.
我是使用vllm启动的模型
得从chatchat这部分改 我后面看下 处理好后回复你
好的,谢谢你
同问,自己也在看。这部分居然没有做限制,一旦超过max-model-len直接让它报错。
这个问题已经被标记为 stale ,因为它已经超过 30 天没有任何活动。
stale
No branches or pull requests
功能描述 / Feature Description
是否支持计算传给模型的引用文本、提示词、用户问题等内容的token数量,并且在超出模型能接受的最大token的时候进行动态限制,防止报错: 请求内容超出模型最大token数量
解决的问题 / Problem Solved
解决 请求内容超出模型最大token数量的报错
实现建议 / Implementation Suggestions
替代方案 / Alternative Solutions
其他信息 / Additional Information
The text was updated successfully, but these errors were encountered: