-
Notifications
You must be signed in to change notification settings - Fork 73
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Regarding the Language Model used #85
Comments
This is equivalent to the implementation before adding the language model. If you wanted to replicate the paper, you would spit out the top 1000 beams from the DS model and rescore them with a LM. (This would take a bit of extra work of course...) |
Yeah I "actually" went through the code this time and already realized there isn't an LM. Oh. That sounds alright. Can you give me some pointers on how to
|
|
And note...the paper mentioned some weighting between the score from the DS model and the score from the LM....wasn't clear if this was estimated or set like a hyper-param |
I'm still playing with the base model code, but once I get better results, I'd be happy to help with this part...but I'm a month or two from where I'll have time... |
I won't be able to implement the language model due to time constraints but it definitely is a large part to the project and would improve the model's performance substantially. A lot of reference can be found in the original deepscribe 2 paper |
Dear all, |
DeepSpeech paper [1] says that at inference time CTC models are paired with a language model.
Which language model does this implementation use? Where is the language model written/stored/called in the code?
How can I use my own language model with the network?
The text was updated successfully, but these errors were encountered: