-
Notifications
You must be signed in to change notification settings - Fork 32
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Update benchmark to run openorca dataset #21
Conversation
5f5540f
to
54194d3
Compare
benchmarks/benchmark_serving.py
Outdated
@@ -42,7 +42,7 @@ | |||
python -m benchmarks.benchmark_serving \ | |||
--request-rate 1 | |||
|
|||
e2e example: python3 benchmark_serving.py --tokenizer /home/rwitten/maxtext/assets/tokenizer --num-prompts 100 --dataset ~/ShareGPT_V3_unfiltered_cleaned_split.json | |||
e2e example: python3 benchmark_serving.py --tokenizer /home/rwitten/maxtext/assets/tokenizer --num-prompts 100 --dataset ~/ShareGPT_V3_unfiltered_cleaned_split.json |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We can also update the example here with your change. And also the README in /benchmark.
54194d3
to
7477e24
Compare
7477e24
to
71c111c
Compare
e9e3ac7
to
4f41058
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We can merge this change first, since I need to release a new JetStream py package. We can do refactor later since I see the current sample filter logic is identical for both dataset.
# Tokenize the prompts and completions. | ||
prompts = dataset["prompts"] | ||
outputs = dataset["results"] | ||
n = len(prompts) | ||
prompt_token_ids = tokenizer.tokenize(prompts) | ||
output_token_ids = tokenizer.tokenize(outputs) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I feel we could extract this part out as a func for different dataset, and the rest are identical and thus we could keep them in the sample_request function?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Discussed offline, there are some existing data processing that may not be necessary. Will revisit and refactor the data preprocessing part if needed.
No description provided.