You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, I recently wanted to use litserve to deploy a service and use litgpt to load the model. However, due to video memory reasons, I wanted to use an AWQ model. However, I did not see the method of loading the awq model in the python API document. Do you have any methods?
The text was updated successfully, but these errors were encountered:
But if you have weights in AWQ format and want to load them to LItGPT - it's not supported.
There was an attempt to support AutoGPTQ (that should also support AWQ format) in #924, but it was never merged.
(Making that PR up-to-date might could be a cool contribution 😉).
Hi, I recently wanted to use litserve to deploy a service and use litgpt to load the model. However, due to video memory reasons, I wanted to use an AWQ model. However, I did not see the method of loading the awq model in the python API document. Do you have any methods?
The text was updated successfully, but these errors were encountered: