-
Notifications
You must be signed in to change notification settings - Fork 89
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
About the experiment in GPT-GNN #19
Comments
And if I use no_pretrain, I saw your code that is only consider train_data, valid_data, test_data in finetune_reddit.py. Am I right? |
For all of the pre-training baselines (including GAE, GraphSAGE-unsuper, and our method), the setting all follows the pretrain-finetune paradigm, which means we first pre-train the model using the self-supervised task on the pre-training dataset (in your example, 0-70%), then use the pre-trained model to finetune on the training set (70-80%), with model selection using the valid set, and get generalization performance on the test set. Yes, for no_pretrain, we don't leverage the pre-training data. |
Very thanks for your reply! I'm still a little confused about GraphSAGE and GAE. Did you put all pre_train data into model or sample data just like your paper described? |
Pre-training will use all the pre_train data, but we conduct mini-batch training by subgraph sampling to avoid memory issue (as the whole graph is too big for the GPU memory) |
Hi authors
Thanks for your amazing work in Pre-train GNN, It can solve larger datasizes Graph in GNN model. but I have some questions when I try to understand the part of experiment. I saw your code in github, and I noticed that it includes pre-train and fine-tune. And I have a question about experiment in your paper.
Looking forward to your reply, thank you!
The text was updated successfully, but these errors were encountered: