If we have a smaller graph, It’s easy to store the node features along with the graph.
I am dealing with a bigger graph and I have all the node features in spark dataframes, What I was doing is as follows: spark_dataframe → pandas_dataframe → numpy_array → torch_tensor but as the graph is large, I am facing the driver memory issues with this approach.
Any efficient method to convert the pyspark dataframe to torch tensors?