Hi,
I am trying to taining a new pinSage model on my own dataset(10M edge, 0.8M Item nodes with feature). However, I noticed that the memroy of GPU used increase slowly ,and the whole procedure failed at last after ten epochs always. Dose any one hit the same issue ? Wish your help, thanks.
here is some error message
out, (argX, argY) = _gspmm(gidx, op, reduce_op, X, Y)
File “/home/dolphinfs_lilifeng/anaconda3/envs/myEnv/lib/python3.7/site-packages/dgl/sparse.py”, line 233, in _gspmm
arg_e_nd)
File “dgl/_ffi/_cython/./function.pxi”, line 293, in dgl._ffi._cy3.core.FunctionBase.call
File “dgl/_ffi/_cython/./function.pxi”, line 239, in dgl._ffi._cy3.core.FuncCall
dgl._ffi.base.DGLError: [15:27:43] /opt/dgl/src/runtime/cuda/cuda_device_api.cc:114: Check failed: e == cudaSuccess || e == cudaErrorCudartUnloading: CUDA: out of memory