Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Request] LLM 适配 #151

Closed
5 of 9 tasks
arvinxx opened this issue Sep 6, 2023 · 35 comments · Fixed by #1173 or #1293
Closed
5 of 9 tasks

[Request] LLM 适配 #151

arvinxx opened this issue Sep 6, 2023 · 35 comments · Fixed by #1173 or #1293
Assignees
Labels
🌠 Feature Request New feature or request | 特性与建议 Model Provider 模型服务商 released

Comments

@arvinxx
Copy link
Contributor

arvinxx commented Sep 6, 2023

独立 LLM 设置 Tab,为上层助手角色提供丰富的大模型选项。

清单:

完整 RFC 在: #737

@arvinxx arvinxx converted this from a draft issue Sep 6, 2023
@arvinxx arvinxx self-assigned this Sep 6, 2023
@canisminor1990 canisminor1990 changed the title LLM适配 [Request] LLM适配 Sep 6, 2023
@canisminor1990 canisminor1990 added the 🌠 Feature Request New feature or request | 特性与建议 label Sep 6, 2023
@arvinxx arvinxx moved this from Todo - Chat to In Progress in Lobe Chat Routine Sep 9, 2023
@arvinxx arvinxx changed the title [Request] LLM适配 [Request] LLM 适配 Sep 9, 2023
@arvinxx arvinxx removed the status in Lobe Chat Routine Oct 6, 2023
@arvinxx arvinxx moved this to Roadmap - Chat 1.0 in Lobe Chat Routine Oct 6, 2023
@ddwinhzy
Copy link

Would like to add OpenAI's fine-tuned model call

@arvinxx
Copy link
Contributor Author

arvinxx commented Oct 11, 2023

Would like to add OpenAI's fine-tuned model call

Yes, with custom model name

@lobehubbot lobehubbot added the Inactive No response in 30 days | 超过 30 天未活跃 label Nov 11, 2023
@arvinxx arvinxx moved this from Roadmap - Chat 1.x to In Progress in Lobe Chat Routine Nov 27, 2023
@gavinliu
Copy link

本地 LLM 是否有一个明确的支持计划?

@lobehubbot
Copy link
Member

Bot detected the issue body's language is not English, translate it automatically. 👯👭🏻🧑‍🤝‍🧑👫🧑🏿‍🤝‍🧑🏻👩🏾‍🤝‍👨🏿👬🏿


Does the local LLM have a clear support plan?

@jyboy
Copy link

jyboy commented Dec 13, 2023

Google's Gemini Pro API is also available now.
Please see the official tweet.

@arvinxx
Copy link
Contributor Author

arvinxx commented Dec 14, 2023

本地 LLM 是否有一个明确的支持计划?

计划在1月份开始做

@lobehubbot
Copy link
Member

Bot detected the issue body's language is not English, translate it automatically. 👯👭🏻🧑‍🤝‍🧑👫🧑🏿‍🤝‍🧑🏻👩🏾‍🤝‍👨🏿👬🏿


Does the local LLM have a clear support plan?

Plan to start in January

@hazelnutcloud
Copy link

any plans to add support for mistral?

@arvinxx
Copy link
Contributor Author

arvinxx commented Dec 18, 2023

yes

@structure-charger
Copy link

@WayneShao
Copy link

我是用了 2023-12-01-preview 版本的API调用 Azure 的 gpt-4-vision 模型,可以成功获得回复,但是回复一句之后就会直接截断
image
控制台输出
image

@lobehubbot
Copy link
Member

Bot detected the issue body's language is not English, translate it automatically. 👯👭🏻🧑‍🤝‍🧑👫🧑🏿‍🤝‍🧑🏻👩🏾‍🤝‍👨🏿👬🏿


I used the 2023-12-01-preview version of the API to call Azure's gpt-4-vision model. I could successfully get a reply, but the reply would be truncated directly after one sentence.
image
console output
image

@arvinxx
Copy link
Contributor Author

arvinxx commented Jan 3, 2024

我是用了 2023-12-01-preview 版本的API调用 Azure 的 gpt-4-vision 模型,可以成功获得回复,但是回复一句之后就会直接截断

image

控制台输出

image

这个应该是 gpt-4-vision 内置的max_tokens 太小了导致的

@lobehubbot
Copy link
Member

Bot detected the issue body's language is not English, translate it automatically. 👯👭🏻🧑‍🤝‍🧑👫🧑🏿‍🤝‍🧑🏻👩🏾‍🤝‍👨🏿👬🏿


I used the 2023-12-01-preview version of the API to call Azure's gpt-4-vision model. I could successfully get a reply, but the reply would be truncated directly after one sentence.

image

Console output

image

This should be caused by the built-in max_tokens of gpt-4-vision being too small.

@Met-Du
Copy link

Met-Du commented Jan 27, 2024

Are there any plans for Gemini Pro? Thank you

@cocobeach
Copy link

独立 LLM 设置 Tab,为上层助手角色提供丰富的大模型选项。

清单:

完整 RFC 在: #737

Hi Team!
There is a very simple simple way of running models locally with ollama https://github.com/ollama/ollama and wrap it up with https://github.com/BerriAI/litellm and using https://github.com/BerriAI/liteLLM-proxy it shouldn't be too difficult to make it work with a lot of OS LLM.
Best Regards

@cocobeach
Copy link

Bot detected the issue body's language is not English, translate it automatically. 👯👭🏻🧑‍🤝‍🧑👫🧑🏿‍🤝‍🧑🏻👩🏾‍🤝‍👨🏿👬🏿

Are there any plans to adapt to ollama?

ollama wrapped in litellm should be a piece of cake, I am going to try it.

@arvinxx
Copy link
Contributor Author

arvinxx commented Jan 30, 2024

@cocobeach Thanks for your introduction! We will add ollama support also~ 😁

@j0ole
Copy link

j0ole commented Feb 1, 2024

Are thers any plants to add a key pool and calling weight to dynamic call those keys?

@arvinxx
Copy link
Contributor Author

arvinxx commented Feb 2, 2024

Are thers any plants to add a key pool and calling weight to dynamic call those keys?

No. There is not plan about key polling.

@arvinxx arvinxx added Model Provider 模型服务商 and removed Inactive No response in 30 days | 超过 30 天未活跃 labels Feb 2, 2024
@arvinxx arvinxx linked a pull request Feb 2, 2024 that will close this issue
42 tasks
@github-project-automation github-project-automation bot moved this from In Progress to Done in Lobe Chat Routine Feb 5, 2024
@lobehub lobehub deleted a comment from lobehubbot Feb 5, 2024
@arvinxx arvinxx reopened this Feb 5, 2024
@zhangheli
Copy link

有没有可能,直接接入 langchain 内 chat model 的支持 ?

@lobehubbot
Copy link
Member

Bot detected the issue body's language is not English, translate it automatically. 👯👭🏻🧑‍🤝‍🧑👫🧑🏿‍🤝‍🧑🏻👩🏾‍🤝‍👨🏿👬🏿


Is it possible to directly access the chat model support in langchain?

@arvinxx
Copy link
Contributor Author

arvinxx commented Feb 5, 2024

@zhangheli 试过了,不行:#737 (reply in thread)

@lobehubbot
Copy link
Member

Bot detected the issue body's language is not English, translate it automatically. 👯👭🏻🧑‍🤝‍🧑👫🧑🏿‍🤝‍🧑🏻👩🏾‍🤝‍👨🏿👬🏿


@zhangheli tried it, but it didn’t work: #737 (reply in thread)

@cocobeach
Copy link

@cocobeach Thanks for your introduction! We will add ollama support also~ 😁

That's great, it would allow Mac users to use the interface with local models easily.

Also for the best inference but that's for nvidia therefore mostly windows station vLLM is quite unique in the way it manages gpu memory, for instance for me it's the only setup that allows me to actually take advantage of my dual RTX a4000 setup but I can run Mixtral on it quite smoothly and with excellent results.
Now that gpt4 got "lazy" I think there is a good opening in the market for open source models, if they are managed properly.

I really like your platform, but I don't understand why the total amount of tokens allowed for a whole chat thread is limited to the model's context windows. Meaning if I am using gpt 3.5 turbo 16k, I get to chat up to 16k thread size, then it gives me an error?
Can I configure that differently?
Best Regards

@arvinxx
Copy link
Contributor Author

arvinxx commented Feb 5, 2024

@cocobeach you can set a limit of current topic context. We don't set a limit by default.
image

@cocobeach
Copy link

Tanks a lot!!

@ifsheldon
Copy link

Please add Moonshot support! See https://platform.moonshot.cn
The APIs are a compatible subset with OpenAI's

@sjy
Copy link
Contributor

sjy commented Feb 7, 2024

@cocobeach Thanks for your introduction! We will add ollama support also~ 😁

This a PR WIP for Ollama Provider, #1265 @arvinxx

@lobehubbot
Copy link
Member

@arvinxx

This issue is closed, If you have any questions, you can comment and reply.
此问题已经关闭。如果您有任何问题,可以留言并回复。

@lobehubbot
Copy link
Member

🎉 This issue has been resolved in version 0.127.0 🎉

The release is available on:

Your semantic-release bot 📦🚀

skylinemusiccds referenced this issue in skylinemusiccds/BrainWeaver-development May 12, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
🌠 Feature Request New feature or request | 特性与建议 Model Provider 模型服务商 released
Projects
Archived in project