Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

TASKS=llm,rag模式下,出现线程问题报错:RuntimeError: Cannot re-initialize CUDA in forked subprocess. To use CUDA with multiprocessing, you must use the 'spawn' start method #308

Open
2 tasks done
syusama opened this issue Aug 21, 2024 · 3 comments

Comments

@syusama
Copy link

syusama commented Aug 21, 2024

提交前必须检查以下项目 | The following items must be checked before submission

  • 请确保使用的是仓库最新代码(git pull),一些问题已被解决和修复。 | Make sure you are using the latest code from the repository (git pull), some issues have already been addressed and fixed.
  • 我已阅读项目文档FAQ章节并且已在Issue中对问题进行了搜索,没有找到相似问题和解决方案 | I have searched the existing issues / discussions

问题类型 | Type of problem

模型推理和部署 | Model inference and deployment

操作系统 | Operating system

Linux

详细描述问题 | Detailed description of the problem

Ubuntu系统
docker-compose部署
镜像api-llm:vllm

当同时部署llm和embedding模型时

TASKS=llm,rag

会报错:
RuntimeError: Cannot re-initialize CUDA in forked subprocess. To use CUDA with multiprocessing, you must use the 'spawn' start method

单独部署llm则没有问题

运行日志或截图 | Runtime logs or screenshots

`
微信截图_20240821105553

@FreeRotate
Copy link
Contributor

你好,请问这个问题解决了吗?我也报同样的错误

@syusama
Copy link
Author

syusama commented Oct 12, 2024

你好,请问这个问题解决了吗?我也报同样的错误

没有解决,我后来是分别启一个llm,一个rag

@FreeRotate
Copy link
Contributor

你好,请问这个问题解决了吗?我也报同样的错误

没有解决,我后来是分别启一个llm,一个rag

我刚解决了,回退vllm和torch版本,使用vllm==0.4.2,torch==2.3.0

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants