salesforce / ctrl

Conditional Transformer Language Model for Controllable Generation

Home Page:https://arxiv.org/abs/1909.05858

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

How to finetuning with lower memory fp16 version for p100 GPUs?

xurongqiang opened this issue · comments

For finetuning with lower memory fp16 version(for fp32 version , OOMs occur. ), How should I modify the training.py script?

This is quite delicate and doesn't quite seem to work out of the box. I'm going to need more time to look into this.

This is quite delicate and doesn't quite seem to work out of the box. I'm going to need more time to look into this.

The background to this problem is that we have a large number of p100 machines, but they cannot run on the fp32 version. Thank you for your improvement.

Is there some update on this matter? We have the same problem, unfortunately.

Yes same here for us. Both huggingface and this repo seem to have the same OOM error when running on Google Colab free GPU like p100. Any fix or workaround yet?

The problem still persists, unfortunately. Fine-tuning doesn't really work with collab resources..