Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support Multi-GPU inference on CUDA devices #101

Merged
merged 6 commits into from
Jan 14, 2025

Conversation

guoqingbao
Copy link
Collaborator

Run Multi-GPU inference with NCCL feature

cargo run --release --features cuda,nccl -- --port 2000 --device-ids "0,1" --weight-path /home/Meta-Llama-3.1-8B-Instruct/ llama3 --temperature 0. --penalty 1.0

If you encoutered problems under Multi-GPU setttings, you may:

export NCCL_P2P_LEVEL=LOC # use local devices (mutiple cards within a server, PCIE, etc.)
export NCCL_P2P_DISABLE=1 # diable p2p cause this feature can cause illegal memory access in certain environments
export NCCL_IB_DISABLE=1 # diable ibnet/infiniband (optional)

Note: quantized models are not supported yet under multi-gpu setting.

@guoqingbao guoqingbao merged commit 8fc4c00 into EricLBuehler:master Jan 14, 2025
6 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant