-
-
Notifications
You must be signed in to change notification settings - Fork 5.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: vllm crashes when preemption of priority scheduling is triggered on vllm-0.6.3.dev173+g36ea7907.d20241011 #9342
Comments
Any update or suggestion for this? |
Encountering the exact same thing. |
Encountering the same situation。 |
Encountering the exact same thing. Is there any one fix? |
I use --disable-async-output-proc to avoid this problem,so I think #7049 this pr raise the bug |
This solved it for me, thank you! 🙏 |
Your current environment
The output of `python collect_env.py`
Model Input Dumps
No response
🐛 Describe the bug
As I mentioned in #9272, even if the priority is propagated successfully, vllm always crashes as long as preemption happens.
I just tested with vllm-0.6.3.dev173+g36ea7907.d20241011. The only change I made is following fix and some logs:
#9277
Could you please help to check it?
Reproduce procedure:
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: