You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Is your feature request related to a problem? Please describe.
The feature request is not related to a bug. But at my company, we have this following problem: we would like to be able to dismiss prediction requests after a certain amount of time. Why ? Because they are not relevant anymore after this amount of time. As of today, it is possible to define a timeout, responseTimeout (https://pytorch.org/serve/configuration.html), but it only applies to model inference (handler). It does not include the time spent in queue.
Describe the solution
To avoid any regression, I would suggest to add a new model parameter: predictionRequestTimeout. This parameter defines the timeout after which a prediction request is dismissed. It includes the time spend in queue and inference time. From an implementation point of view, I would suggest to modify:
Is your feature request related to a problem? Please describe.
The feature request is not related to a bug. But at my company, we have this following problem: we would like to be able to dismiss prediction requests after a certain amount of time. Why ? Because they are not relevant anymore after this amount of time. As of today, it is possible to define a timeout, responseTimeout (https://pytorch.org/serve/configuration.html), but it only applies to model inference (handler). It does not include the time spent in queue.
Describe the solution
To avoid any regression, I would suggest to add a new model parameter: predictionRequestTimeout. This parameter defines the timeout after which a prediction request is dismissed. It includes the time spend in queue and inference time. From an implementation point of view, I would suggest to modify:
If you agree with this feature requests, I'm willing to push a PR.
The text was updated successfully, but these errors were encountered: