ictnlp / STEMM

Code for ACL 2022 main conference paper "STEMM: Self-learning with Speech-text Manifold Mixup for Speech Translation".

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

About fine-tuning

zhouyan19 opened this issue · comments

In the paper, Section 2.2 , you say "We combine those pretrained modules and finetune the whole model for ST". Did you freeze the Wav2Vec2.0 Model during training ? If not , I wonder if it's because of the mix-up training strategy , so as to bridge the modality gap.

Thanks for your question. We did not freeze the Wav2vec2.0 module, so that the audio representation can be tuned with training.

Thanks for your question. We did not freeze the Wav2vec2.0 module, so that the audio representation can be tuned with training.

Thank you !