Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

bug: Adaptative batching with max_batch_size=1 crashes the API #4856

Closed
bruno-hays opened this issue Jul 11, 2024 · 0 comments · Fixed by #4895
Closed

bug: Adaptative batching with max_batch_size=1 crashes the API #4856

bruno-hays opened this issue Jul 11, 2024 · 0 comments · Fixed by #4895
Labels
bug Something isn't working

Comments

@bruno-hays
Copy link

Describe the bug

With this decorator on my function:

    @bentoml.api(batchable=True,
                 max_batch_size=1,
                 max_latency_ms=3600000)

I get this incomprehensible error:

  File "/home/gcpuser/sky_workdir/whisperapi/venv/lib/python3.11/site-packages/bentoml/_internal/utils/metrics.py", line 44, in exponential_buckets
    assert start < end
           ^^^^^^^^^^^
AssertionError

now use this header:

    @bentoml.api(batchable=True,
                 max_batch_size=2,
                 max_latency_ms=3600000)

And it works.

Amusingly, this happens when I run my API on debian and I am not able to reproduce this error on my mac M1

To reproduce

No response

Expected behavior

No response

Environment

bentoml==1.2.19
python==3.11.9

@bruno-hays bruno-hays added the bug Something isn't working label Jul 11, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
1 participant