Are there benchmarks on the speed of using user-defined message/reduce functions? Besides built-in functions that invoke g-SDDMM or g-SpMM for message passing, I’m curious about the performance of running user-defined message functions and reduce functions using DGL. How well do the node_batch
and edge_batch
methods perform under different conditions, such as sparse/dense graphs and large/small graphs? Are there statistics that you can share?
As you can see, with the progress of GNN (Graph Neural Networks), there are various message functions. I believe that having access to such statistics would be highly beneficial for the community. It would not only help in understanding the performance nuances of DGL under various conditions but also assist in optimizing our own implementations more effectively.