Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bug]: Chunked prefill returning gibberish in some cases. #4697

Closed
fmmoret opened this issue May 8, 2024 · 2 comments
Closed

[Bug]: Chunked prefill returning gibberish in some cases. #4697

fmmoret opened this issue May 8, 2024 · 2 comments
Labels
bug Something isn't working stale

Comments

@fmmoret
Copy link

fmmoret commented May 8, 2024

Your current environment

main branch Dockerfile.rocm default dependencies.

🐛 Describe the bug

--max-num-batched-tokens=131072 --enable-chunked-prefill -- perfect response. temp 0
--max-num-batched-tokens=16384 --enable-chunked-prefill -- gibberish response. temp 0

Using a prompt of 100001 seq length & generate 100 tokens.

With temp 0, the gibberish does NOT match itself across iterations
E.g.: Good response 1 = What? What?”\n\n“Why, the bridge was mined [...]
Bad response 1 = So far as Jiedgilliesgillies-illies-illies-er. A Jemel-er-illies-ied-: \xa0 [...]
Bad response 2 is entirely different from 1 = \xa0gillies in England-ied. A Jiedgeld-eren [...]

I haven't looked into the VLLM impl yet. It seems like maybe the tensors are not initialized correctly somewhere and are inheriting whatever values were already in memory at the time.

I have seen this kind of thing happen before when someone uses x = torch.empty(size) -- which initializes to whatever memory already had set for that segment -- when they meant to use / wanted zeros.

@fmmoret fmmoret added the bug Something isn't working label May 8, 2024
@fmmoret fmmoret changed the title [Bug]: batched prefill returning gibberish in some cases. [Bug]: Chunked prefill returning gibberish in some cases. May 10, 2024
Copy link

This issue has been automatically marked as stale because it has not had any activity within 90 days. It will be automatically closed if no further activity occurs within 30 days. Leave a comment if you feel this issue should remain open. Thank you!

@github-actions github-actions bot added the stale label Oct 27, 2024
Copy link

This issue has been automatically closed due to inactivity. Please feel free to reopen if you feel it is still relevant. Thank you!

@github-actions github-actions bot closed this as not planned Won't fix, can't repro, duplicate, stale Nov 28, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working stale
Projects
None yet
Development

No branches or pull requests

1 participant