Hi!
I’m a beginner in Graph Neural Networks. After I followed the instructions in the tutorial. I’m curious about full graph training on distributed platforms. All the tutorials used graph sampler to reduce memory consumption. However, if I have an abundant amount of device memory, is there any method to put all the graph on GPUs to shorten the communication overhead? (For instance, partition the graph and load each partition to GPUs)
It seems like the difference between full-batch training and mini-batch training. All the sample code right now is mini-batch training. I’m wondering whether full-batch training on distributed platform is supported in DGL or not. If so, is there any sample code or API references available?
I have tried to write a toy code. But it seems not full batch training, just partition the graph and train them separately.
Thank you very much for your time in reading my question.