Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Why batch training? #78

Open
jiaojiaoguan opened this issue Sep 16, 2022 · 2 comments
Open

Why batch training? #78

jiaojiaoguan opened this issue Sep 16, 2022 · 2 comments

Comments

@jiaojiaoguan
Copy link

jiaojiaoguan commented Sep 16, 2022

Hello everyone:

I have a error when I run e = Wh1 + Wh2.T

03c1df1c422d9e571631705a0858377

because the dimension of e is N by N, in my graph, n is about 400000. so it takes 400000 x 40000 x 4/1024/1024/1024 GB (about 600G). I want to know whether batch size is useful.

Thanks in advance!

@LQchen1
Copy link

LQchen1 commented Sep 21, 2022

Hello, I think it's useless, because GAT actually needs to compute the Attention of the entire graph

@Preston2jager
Copy link

From my understanding batch will work, for a shallow GAT example 2 layers, a single node only gathers information from nodes 2 steps away, nodes beyond that distance have no affect.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants