Add substitutions option in ASR evaluation#15343
Open
avmonte wants to merge 2 commits intoNVIDIA-NeMo:mainfrom
Open
Add substitutions option in ASR evaluation#15343avmonte wants to merge 2 commits intoNVIDIA-NeMo:mainfrom
avmonte wants to merge 2 commits intoNVIDIA-NeMo:mainfrom
Conversation
Signed-off-by: avmonte <unstoppablehay@gmail.com>
Signed-off-by: avmonte <unstoppablehay@gmail.com>
Author
|
@titu1994, @redoctopus, @jbalam-nv, @okuchaiev Please take a look :) |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Hi there!
I've been experimenting with Armenian ASR and noticed a consistent evaluation issue: the model never outputs the single-character Armenian
և(U+0587). Instead it outputs the decomposed formեւ(two characters). As the ground-truth transcripts containև, this mismatch inflates WER despite the fact that the predicted text is linguistically equivalent.This PR adds an evaluation-time normalization option to address that type of issues.
text_processing.substitutionsfor evaluation-time text normalization:SRC~DST;SRC2~DST2;...\u0587.Reproduce
Below is the configuration I used to run my evaluation
Evaluation results
Base MCV test WER (from the HF model card): 9.90
Normalized MCV test WER: 5.42