Problem with cuda
aipursuing opened this issue · comments
When I run run_raft.sh with RTX3090Ti, the cuda error occurs
`model = torch.nn.DataParallel(RAFT(args))
File "/home/pai/anaconda3/envs/aigc/lib/python3.10/site-packages/torch/nn/parallel/data_parallel.py", line 145, in init
self.module.to(self.src_device_obj)
File "/home/pai/anaconda3/envs/aigc/lib/python3.10/site-packages/torch/nn/modules/module.py", line 927, in to
return self._apply(convert)
File "/home/pai/anaconda3/envs/aigc/lib/python3.10/site-packages/torch/nn/modules/module.py", line 579, in _apply
module._apply(fn)
File "/home/pai/anaconda3/envs/aigc/lib/python3.10/site-packages/torch/nn/modules/module.py", line 579, in _apply
module._apply(fn)
File "/home/pai/anaconda3/envs/aigc/lib/python3.10/site-packages/torch/nn/modules/module.py", line 602, in _apply
param_applied = fn(param)
File "/home/pai/anaconda3/envs/aigc/lib/python3.10/site-packages/torch/nn/modules/module.py", line 925, in convert
return t.to(device, dtype if t.is_floating_point() or t.is_complex() else None, non_blocking)
RuntimeError: CUDA error: out of memory
CUDA kernel errors might be asynchronously reported at some other API call,so the stacktrace below might be incorrect.
For debugging consider passing CUDA_LAUNCH_BLOCKING=1.`
How can I solve it? My cuda version is 11.4.
Really appreciate your time, looking forward to your reply.