-
-
Notifications
You must be signed in to change notification settings - Fork 11.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Request] 能否支持在服务端配置模型是否使用客户端请求模式 #5273
Comments
Thank you for raising an issue. We will investigate into the matter and get back to you as soon as possible. |
🥰 Description of requirementsCan it support configuring the model on the server side and whether to use the client request mode? 🧐 SolutionAdd an environment variable to control whether client request mode is enabled. 📝 Supplementary informationIf -language_model_settings is configured, the user will not be able to modify whether to use client request mode |
目前 Google AI Studio 模型也不支持客户端请求模式,导致在较大 Context 的情况下 Vercel 的部署会出现请求超时的问题。 Edited: 不确定是否 Google AI Studio 是否支持这种模式,可能是 API 限制。 |
Currently, the Google AI Studio model does not support client request mode, resulting in request timeout issues in the deployment of Vercel when the Context is large. |
客户端请求模式会明文携带你的 api key 的,如果在服务端配置然后客户端请求的话,你服务端配置的 key 会被完全暴漏 |
The client request mode will carry your api key in plain text. If you configure it on the server and then request it from the client, the key configured on the server will be completely exposed. |
能否只是支持ollama呢?本地运行ollama还是比较常用的场景 |
Can we just support ollama? Running ollama locally is still a common scenario. |
🥰 需求描述
能否支持在服务端配置模型是否使用客户端请求模式
🧐 解决方案
添加一个环境变量来控制是否启用客户端请求模式.
📝 补充信息
如果配置了-language_model_settings,用户将无法修改是否使用客户端请求模式
经过测试,本地部署的Ollama模型如果关闭客户端请求模式将无法使用,必须开启才行.
The text was updated successfully, but these errors were encountered: