-
Notifications
You must be signed in to change notification settings - Fork 863
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Use Case: Enhancing LLM Serving with Torch Compiled RAG on AWS Graviton #3276
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Overall LGTM, left some minor to nit comments.
|
||
### Download Llama | ||
|
||
Follow [this instruction](https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct) to get permission |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Would be good to update this to Llama 3.1 now that its released https://huggingface.co/meta-llama/Meta-Llama-3.1-70B-Instruct
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I tried 3.1. One interesting observation was , when I ask it "What's new with Llama 3.1", it gives an acceptable answer, which shouldn't be possible. :D
So, for this usecase Llama 3 drives home the point better
|
||
### Download Llama | ||
|
||
Follow [this instruction](https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct) to get permission |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Follow [this instruction](https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct) to get permission | |
Follow [this instruction](https://huggingface.co/meta-llama/Meta-Llama-3.1-70B-Instruct) to get permission |
huggingface-cli login --token $HUGGINGFACE_TOKEN | ||
|
||
```bash | ||
python ../Download_model.py --model_path model --model_name meta-llama/Meta-Llama-3-8B-Instruct |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
python ../Download_model.py --model_path model --model_name meta-llama/Meta-Llama-3-8B-Instruct | |
python ../Download_model.py --model_path model --model_name meta-llama/Meta-Llama-3.1-8B-Instruct |
```bash | ||
python ../Download_model.py --model_path model --model_name meta-llama/Meta-Llama-3-8B-Instruct | ||
``` | ||
Model will be saved in the following path, `model/models--meta-llama--Meta-Llama-3-8B-Instruct`. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Model will be saved in the following path, `model/models--meta-llama--Meta-Llama-3-8B-Instruct`. | |
Model will be saved in the following path, `model/models--meta-llama--Meta-Llama-3.1-8B-Instruct`. |
Co-authored-by: Matthias Reso <13337103+mreso@users.noreply.github.com>
Co-authored-by: Matthias Reso <13337103+mreso@users.noreply.github.com>
Description
This PR shows a use case of TorchServe for GenAI deployment.
The use-case shows how torch.compile can be used on embedding model in RAG to improve throughput and how this RAG endpoint can be used along with an LLM endpoint to improve results from the LLM
Fixes #(issue)
Type of change
Please delete options that are not relevant.
Feature/Issue validation/testing
Please describe the Unit or Integration tests that you ran to verify your changes and relevant result summary. Provide instructions so it can be reproduced.
Please also list any relevant details for your test configuration.
Test A
Logs for Test A
Test B
Logs for Test B
Checklist: