Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Random initialization is better? #1

Open
zhiweihu1103 opened this issue Sep 7, 2021 · 4 comments
Open

Random initialization is better? #1

zhiweihu1103 opened this issue Sep 7, 2021 · 4 comments

Comments

@zhiweihu1103
Copy link

Problem description: I initialize the parameters of NewLook randomly (that is, set the --init or --init_checkpoint parameter to None in run_model_NewLook.py), and find that the effect of using random initialization directly is better than the effect presented in the paper (take FB15k-237 as example). Can you explain it?
Below is the results table:
微信截图_20210907182814
Notes:
NLK refers to the results of your paper;
VAL refers to the valid set results after training 30000 steps, and use the saved model to predict.
TST refers to the test set results with the random initialization parameters for NewLook model.
We can see, the effect of TST is often better, Very confused about this result. Looking forward to your reply

@lihuiliullh
Copy link
Owner

what is the command you use for random initialization?

@zhiweihu1103
Copy link
Author

what is the command you use for random initialization?

I modify the run_model_NewLook.py only removed the train related code, and set the --init parameter to None for random initialization. Just only do this.

@lihuiliullh
Copy link
Owner

Thanks for pointing the problem out. I just fixed the problem. It should work correctly now.

@zhiweihu1103
Copy link
Author

zhiweihu1103 commented Sep 13, 2021

Thanks for pointing the problem out. I just fixed the problem. It should work correctly now.

Thank you for your quickly fix. Except above, I have other questions:
Q1: Would you mind share the code about how you create the difference datasets (such as the test_triples_3d.pkl and so on with the d flag);
Q2: There are also some doubts about the generation of test and valid hard version data sets. Whether it is convenient for you to open source the code for generating hard version data(such as the test_ans_ci_hard.pkl)?
Looking forward to your reply. Thx.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants