Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

nndep Classifier and Config: Prevent fine-tuning with flag. #144

Closed
wants to merge 1 commit into from

Conversation

cinjon
Copy link
Contributor

@cinjon cinjon commented Feb 15, 2016

Add a flag (default: true) for performing the word embedding gradient update. The main purpose is for those who want to provide their own embeddings and not fine-tune them.

…g the word embedding gradient update. The main purpose is for those who want to provide their own embeddings and not finetune them.
@manning manning added this to the v.3.6.1 milestone Feb 19, 2016
@manning
Copy link
Member

manning commented Feb 19, 2016

Looks good, thanks!

@manning
Copy link
Member

manning commented Feb 19, 2016

Merged.

@manning manning closed this Feb 19, 2016
@cinjon
Copy link
Contributor Author

cinjon commented Feb 19, 2016

This implementation has a problem actually in that it deals poorly with any word embeddings in the training data that aren't in the provided embedFile. They are initialized to random vectors and then are neither treated as -UNKNOWN- nor updated in the gradient step.

I think that the more faithful solution would be to treat them as -UNKNOWN-, but that's not the intention of the code. An alternative would be to selectively update those embeddings only.

What do you think? I can make the change.

@manning
Copy link
Member

manning commented Feb 21, 2016

Maybe @hans or @danqi could give a more expert opinion, but it would seem okay to me to treat them as UNK.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants