-
-
Notifications
You must be signed in to change notification settings - Fork 14.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
vllm: init at v0.3.1 with rocm support #283842
Conversation
Flask/Werkzeug can't go into master and needs a lot more work on reverse dependencies. |
vllm as is still needed some work and wasn't working properly with cuda. If you've got it working for rocm, we could merge an early version with that only. (if you ever want to submit a PR for vllm, you don't need to credit me at all, you can just copy paste what you feel is useful and put me as a maintainer). |
19e6aa3
to
a5d537a
Compare
a5d537a
to
27666ab
Compare
Managed to get it working on Cuda with this PR: |
27666ab
to
7420a6a
Compare
7420a6a
to
1f409b4
Compare
1f409b4
to
1846a26
Compare
Currently re-testing CUDA build after rebase on master, for ROCm it requires |
gpuTargets | ||
else | ||
# vllm supports less gpu targets than rocm clr, supported target list is taken from ROCM_SUPPORTED_ARCHS in setup.py | ||
lib.lists.intersectLists rocmPackages.clr.gpuTargets ["gfx90a" "gfx908" "gfx906" "gfx1030" "gfx1100"] |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks like vllm setup.py calculates the intersection by itself now, this needs to be removed.
Tested with CUDA, it works. Edit: retested with cuda, i was testing on base branch without #285249, I have patched dependencies for torch 2.2.0. |
bd63471
to
3adc470
Compare
3adc470
to
dac6982
Compare
Works on ROCm. |
@CertainLach thank you very much for this! |
Description of changes
Depends on:
werkzeug
3.x psf/httpbin#36 (As the needed for new werkzeug version is not released. VLLM doesn't depend directly on this, but httpbin is still in the dependency graph, and needs to be built.) - fix backported to stagingI am not very familiar with python packaging in nixos (Though, anything is better than requirements.txt)
Couple of changes were made to fix build due to conflicting dependencies:
I think the only working way here is to use
nixpkgs.config.torchSupport
/nixpkgs.config.cudaSupport
, as overriding torch dependency doesn't work very well in this case.It also builds with CUDA for me, though I haven't properly tested it in runtime
Things done
nix.conf
? (See Nix manual)sandbox = relaxed
sandbox = true
nix-shell -p nixpkgs-review --run "nixpkgs-review rev HEAD"
. Note: all changes have to be committed, also see nixpkgs-review usage./result/bin/
)Add a 👍 reaction to pull requests you find important.