Originally created by: Athena-I
I tried to make the inference on A30, while an error occurred: RuntimeError: CUDA out of memory. How to inference on multi cards?
Log in to post a comment.