Skip to content

Conversation

@greenw0lf
Copy link
Contributor

With the non-batched version of the WhisperModel, you would get logging output like:

Processing audio with duration 01:33:59.990
VAD filter removed 06:55.648 of audio

Whereas, when calling the BatchedInferencePipeline's transcribe() method, that is no longer the case.

This PR tries to bring that logging back as I believe it does not add extra overhead to the logging and it is quite useful for developers who wish to know how much audio gets processed in the end.

If this is not the case and it is an issue related to my usage of the model, I apologize in advance!

self,
model,
):
self.logger = get_logger()
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

self.model already has a logger, so I'd rather we use the same logger instead of having duplicate loggers in both the pipeline and the model

@MahmoudAshraf97 MahmoudAshraf97 merged commit 9516429 into SYSTRAN:master Dec 23, 2024
3 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants