Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Request] 能否支持在服务端配置模型是否使用客户端请求模式 #5273

Open
EnTaroYan opened this issue Jan 3, 2025 · 4 comments
Labels
🌠 Feature Request New feature or request | 特性与建议 ollama Relative to Ollama Provider and ollama models

Comments

@EnTaroYan
Copy link

🥰 需求描述

能否支持在服务端配置模型是否使用客户端请求模式

🧐 解决方案

添加一个环境变量来控制是否启用客户端请求模式.

📝 补充信息

如果配置了-language_model_settings,用户将无法修改是否使用客户端请求模式
经过测试,本地部署的Ollama模型如果关闭客户端请求模式将无法使用,必须开启才行.

@EnTaroYan EnTaroYan added the 🌠 Feature Request New feature or request | 特性与建议 label Jan 3, 2025
@lobehubbot
Copy link
Member

👀 @EnTaroYan

Thank you for raising an issue. We will investigate into the matter and get back to you as soon as possible.
Please make sure you have given us as much context as possible.
非常感谢您提交 issue。我们会尽快调查此事,并尽快回复您。 请确保您已经提供了尽可能多的背景信息。

@lobehubbot
Copy link
Member

Bot detected the issue body's language is not English, translate it automatically. 👯👭🏻🧑‍🤝‍🧑👫🧑🏿‍🤝‍🧑🏻👩🏾‍🤝‍👨🏿👬🏿


🥰 Description of requirements

Can it support configuring the model on the server side and whether to use the client request mode?

🧐 Solution

Add an environment variable to control whether client request mode is enabled.

📝 Supplementary information

If -language_model_settings is configured, the user will not be able to modify whether to use client request mode
After testing, the locally deployed Ollama model cannot be used if the client request mode is turned off. It must be turned on.

@dosubot dosubot bot added the ollama Relative to Ollama Provider and ollama models label Jan 3, 2025
@youngzyl
Copy link

youngzyl commented Jan 3, 2025

目前 Google AI Studio 模型也不支持客户端请求模式,导致在较大 Context 的情况下 Vercel 的部署会出现请求超时的问题。

Edited: 不确定是否 Google AI Studio 是否支持这种模式,可能是 API 限制。

@lobehubbot
Copy link
Member

Bot detected the issue body's language is not English, translate it automatically. 👯👭🏻🧑‍🤝‍🧑👫🧑🏿‍🤝‍🧑🏻👩🏾‍🤝‍👨🏿👬🏿


Currently, the Google AI Studio model does not support client request mode, resulting in request timeout issues in the deployment of Vercel when the Context is large.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
🌠 Feature Request New feature or request | 特性与建议 ollama Relative to Ollama Provider and ollama models
Projects
None yet
Development

No branches or pull requests

3 participants