-
-
Notifications
You must be signed in to change notification settings - Fork 5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: Loading mistralai/Mixtral-8x22B-Instruct-v0.1 raises TypeError: a bytes-like object is required, not 'str' #9821
Comments
I managed to fix this by explicitly casting the device name to diff --git a/vllm/model_executor/layers/fused_moe/fused_moe.py b/vllm/model_executor/layers/fused_moe/fused_moe.py
index 1cf5c225..b43bb39a 100644
--- a/vllm/model_executor/layers/fused_moe/fused_moe.py
+++ b/vllm/model_executor/layers/fused_moe/fused_moe.py
@@ -288,7 +288,7 @@ def invoke_fused_moe_kernel(A: torch.Tensor, B: torch.Tensor, C: torch.Tensor,
def get_config_file_name(E: int, N: int, dtype: Optional[str]) -> str:
- device_name = current_platform.get_device_name().replace(" ", "_")
+ device_name = str(current_platform.get_device_name()).replace(" ", "_")
dtype_selector = "" if not dtype else f",dtype={dtype}"
return f"E={E},N={N},device_name={device_name}{dtype_selector}.json" |
It look likes your pip install nvidia-ml-py==12.560.30 |
@jeejeelee Thank you for the quick reply. What do you think of setting a minimum version for - nvidia-ml-py # for pynvml package
+ nvidia-ml-py >= 12.560.30 # for pynvml package |
IMHO, we should add a minimum version, cc @youkaichao |
sounds good, we can specify |
Your current environment
The output of `python collect_env.py`
🐛 Describe the bug
When trying to serve mistralai/Mixtral-8x22B-Instruct-v0.1 in a single-node with 8 GPUs I get the error below:
You can reproduce the error with the cli above (and a similar environment) or with the snippet below:
The text was updated successfully, but these errors were encountered: