-
-
Notifications
You must be signed in to change notification settings - Fork 879
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Enable Ascend NPU support #1758
Conversation
Good day! @winglian I tried to create a class Thus, finnaly I refactor the whole This brings a lot changes, while making the model loading pipeline more clearly. Moreover, the changes of member variables such as Please review the latest code and give me some suggestions. Thanks a lot! |
a9b5ca4
to
8d39332
Compare
Looks like it includes two parts in this commits Model Loaders reafactor and Ascend NPU support. Maybe we could spilit it as two PRs, the first one is Model Loaders reafactor, then we will rebase the Ascend NPU support PR after it. Or do you have any other suggestions? @winglian Please feel free let us know if you have any more concern. Thanks! |
58afbe0
to
080f4eb
Compare
@winglian Hi, Ascend NPU support is done on the latest branch, plz review it, thanks! |
8b72705
to
bfbc3a4
Compare
7caec61
to
d9b71ed
Compare
@NanoCode012 I just pushed the lasted code, but I accidentally closed PR before that, and now CI is stopped. What should I do now? |
@MengqingCao , no worries. I restarted them. I checked the PR, and all my points have been resolved. Thank you for addressing them. I plan to let the multi-gpu CI to run as well to ensure no issues there. |
Description
Enable Ascend NPU backend for finetuning, inferencing and gradio webui.
Main changes:
device
Motivation and Context
There are two benefits:
Example
Screenshots
NPU supported CLI inference
NPU supported Gradio webui inference
Config
lora.yaml