-
-
Notifications
You must be signed in to change notification settings - Fork 10.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Request] LLM 适配 #151
Comments
Would like to add OpenAI's fine-tuned model call |
Yes, with custom model name |
本地 LLM 是否有一个明确的支持计划? |
Does the local LLM have a clear support plan? |
Google's Gemini Pro API is also available now. |
计划在1月份开始做 |
Plan to start in January |
any plans to add support for mistral? |
yes |
This should be caused by the built-in max_tokens of gpt-4-vision being too small. |
Are there any plans for Gemini Pro? Thank you |
Hi Team! |
ollama wrapped in litellm should be a piece of cake, I am going to try it. |
@cocobeach Thanks for your introduction! We will add ollama support also~ 😁 |
Are thers any plants to add a key pool and calling weight to dynamic call those keys? |
No. There is not plan about key polling. |
有没有可能,直接接入 langchain 内 chat model 的支持 ? |
Is it possible to directly access the chat model support in langchain? |
@zhangheli 试过了,不行:#737 (reply in thread) |
@zhangheli tried it, but it didn’t work: #737 (reply in thread) |
That's great, it would allow Mac users to use the interface with local models easily. Also for the best inference but that's for nvidia therefore mostly windows station vLLM is quite unique in the way it manages gpu memory, for instance for me it's the only setup that allows me to actually take advantage of my dual RTX a4000 setup but I can run Mixtral on it quite smoothly and with excellent results. I really like your platform, but I don't understand why the total amount of tokens allowed for a whole chat thread is limited to the model's context windows. Meaning if I am using gpt 3.5 turbo 16k, I get to chat up to 16k thread size, then it gives me an error? |
@cocobeach you can set a limit of current topic context. We don't set a limit by default. |
Tanks a lot!! |
Please add Moonshot support! See https://platform.moonshot.cn |
|
✅ @arvinxx This issue is closed, If you have any questions, you can comment and reply. |
🎉 This issue has been resolved in version 0.127.0 🎉 The release is available on: Your semantic-release bot 📦🚀 |
[AgentSubmit] travel-assistant @ccdanpian
独立 LLM 设置 Tab,为上层助手角色提供丰富的大模型选项。
清单:
完整 RFC 在: #737
The text was updated successfully, but these errors were encountered: