Skip to content

Issue with WER evaluation when using a pre-trained model #7893

Answered by titu1994
vdovichevnick asked this question in Q&A
Discussion options

You must be logged in to vote

If you just need transcription, you can simply use model.transcribe() and pass a list of files. If there's a whole manifest of files, then you can use the transcribe_speech.py script in ASR examples.

For your code though, wer Decoding can be found for ctc under asr.metrics.wer.py - https://github.com/NVIDIA/NeMo/blob/main/nemo/collections/asr/metrics/wer.py#L993

This is Char decoding, for subwords use CTCBPEDecoding in ctc_bpe.py

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@vdovichevnick
Comment options

Answer selected by vdovichevnick
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants