We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
我们想实现 turbomind 的离线推理模型 llava-interleave-qwen-7b-hf, 请问有什么可以参考的案例吗?包括模型参数配置、模型转换/加载过程以及模型推理实现中的注意问题等。
No response
The text was updated successfully, but these errors were encountered:
为了兼容pytorch/turbomind两个后端,我们把vision模型从vlm中拆了出来。
要支持turbomind离线推理,主要部分有三个:
我觉得可以参考一下这个PR是如何做的,基本上覆盖了上面说的几个部分,如果有问题的话可以再讨论。 https://github.com/InternLM/lmdeploy/pull/1425/files
Sorry, something went wrong.
bugfix: llava-hf/llava-interleave-qwen-7b-hf (InternLM#2497)
6ef3a7e
- fix init raise exception because tie_word_embeddings config - max_batch_size option for start
092d960
- fix init raise exception because tie_word_embeddings config
irexyc
No branches or pull requests
Motivation
我们想实现 turbomind 的离线推理模型 llava-interleave-qwen-7b-hf, 请问有什么可以参考的案例吗?包括模型参数配置、模型转换/加载过程以及模型推理实现中的注意问题等。
Related resources
Additional context
No response
The text was updated successfully, but these errors were encountered: