Inference speed is too slow, how to optimize it and accelerate inference?
SkylerZheng opened this issue · comments
Hi, I used HPSv2 on A100(1 single GPU), it's about 21-23 seconds per run. Are there anyway we can try to improve the latency?
Do you have a minimal script to reproduce the issue?