Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[python] Fix new logprobs computation in vllm_utils #2146

Merged
merged 1 commit into from
Jul 4, 2024

Conversation

sindhuvahinis
Copy link
Contributor

Description

This bug is only in master, not in 0.28.0. This fixes the LMI no-code low code CI failures.

Even if we set logprobs=1, sometimes, vLLM sends more than one log probabilities. Here for new log probs, we add all the log probabilities that are return by vLLM to new_logprobs dict.

But when we determine whether it is last token or not, i == (len(new_logprobs) -1) and this fails, because now it has more than one new probs, this case will never be true. So last_token never occurred, so it returned broken json without any details. Hence the CI failed.

Will add unit test cases for this use-cases as well in the next PR.

@sindhuvahinis sindhuvahinis requested review from zachgk, frankfliu and a team as code owners July 4, 2024 01:24
@sindhuvahinis sindhuvahinis merged commit 46e05cb into deepjavalibrary:master Jul 4, 2024
9 checks passed
@sindhuvahinis sindhuvahinis deleted the ci branch July 10, 2024 19:06
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants