Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add duration of audio and VAD removed duration to BatchedInferencePipeline #1186

Merged
merged 14 commits into from
Dec 23, 2024

Conversation

greenw0lf
Copy link
Contributor

With the non-batched version of the WhisperModel, you would get logging output like:

Processing audio with duration 01:33:59.990
VAD filter removed 06:55.648 of audio

Whereas, when calling the BatchedInferencePipeline's transcribe() method, that is no longer the case.

This PR tries to bring that logging back as I believe it does not add extra overhead to the logging and it is quite useful for developers who wish to know how much audio gets processed in the end.

If this is not the case and it is an issue related to my usage of the model, I apologize in advance!

@@ -114,6 +114,7 @@ def __init__(
self,
model,
):
self.logger = get_logger()
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

self.model already has a logger, so I'd rather we use the same logger instead of having duplicate loggers in both the pipeline and the model

@MahmoudAshraf97 MahmoudAshraf97 merged commit 9516429 into SYSTRAN:master Dec 23, 2024
3 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants