Skip to content

[Question] Loss explodes instantly when fine-tuning Conformer-CTC on custom dataset #4421

Answered by sqrk
sqrk asked this question in Q&A
Discussion options

You must be logged in to vote

After some investigation, I think that the exploding loss could have been caused by the vocabulary size being too large. I initially thought the vocab size was supposed to be the number of unique words in my dataset, but since we're using CTC it should be at least the number of characters in the dataset or more. And in order to use the pre-trained weights, the vocabulary size needs to match.

Replies: 4 comments 7 replies

Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Answer selected by sqrk
Comment options

You must be logged in to vote
6 replies
@sqrk
Comment options

@VahidooX
Comment options

@titu1994
Comment options

@sqrk
Comment options

@titu1994
Comment options

Comment options

You must be logged in to vote
1 reply
@titu1994
Comment options

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
5 participants
Converted from issue

This discussion was converted from issue #4400 on June 22, 2022 16:22.