-
Notifications
You must be signed in to change notification settings - Fork 6
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Would you mind add GCN model in baselines? #1
Comments
You can put the code of https://github.com/tkipf/gcn into Please make sure that the GCN had been installed on that directory. Besides, you can see whether the GCN works well by setting the variable debug as True in |
Thanks a lot. Besides that, you said tuning the size of hidden layer, this parameter is 'flags.DEFINE_integer('hidden1', 16, 'Number of units in hidden layer 1.')'? Or you tuning the number of hidden layers? |
Yes,it means the number of units in the hidden layer 1. |
OK,thanks a lot. I'll try, please keep in touch! |
I find the GCN source code provided dataset format is different from the pubmed dataset your provide and has it load data function, is that we use the GCN provided pubmed dataset to train GCN model, and use AutoNE provided dataset to train classification? We don't need to rewrite the GCN load data model? Are the Pubmed dataset you provided is same as GCN provided? |
For easily processing, I change the format of the pubmed. You can see the convert_gcn_data function in the src/utils.py to get how to preprocess it. Besides, you can find the download link of the preprocessed datasets in the readme. |
Did you change the GCN source code to save it embedding results? If not, did you directly use the classification result providing by GCN as the AutoNE final result? If that, does it need to change other part of AutoNE (Because it seems need to put embeddings as input)? |
Sorry, I have modified the GCN code to filtering the I/O. The gcn and AROPE codes are shown in here. |
Maybe I have fixed this problem, it need to run convert_gcn_data function at first and then running cmd needs to set hyper-parameter '--output_dir' at first to save gcn embeddings...Thanks a lot |
It is very strange. The first run of BayesOpt is just random, it should be similar to the random search. But in the first trial in your experiments, the random search is much better. In fact, I also find the BayesOpt maybe sometimes worse than Random in the beginning, for example, Figure 3 a, b. As for 4 hyper-parameters, I think in your figure, the higher curve should be better, so AutoNE has a better result than NaiveAutoNE. The problem of random may be similar to the previous one and please run more times. |
In fact, in hyperparameter optimization experiments, the AUC in trials T is the best AUC value we had found. So the curve of each method should be nondecreasing. In my code, I do this in random_search function (main.py Line 274-277). But in Bayes_opt and AutoNE, I do not do that for debug. You should do that when you plot the figure. As for that random may be better than Bayes_opt, I think it is OK. In my results, the Bayes_opt is also not always better than random. |
In the KDD19 paper, besides Deepwalk and AROPE, you also use GCN model to prove the performance of AutoNE framework, but your source code do not contain this part? Would you mind add this?
The text was updated successfully, but these errors were encountered: