marian-nmt / marian-dev

Fast Neural Machine Translation in C++ - development repository

Home Page:https://marian-nmt.github.io

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Documentation of BERT command line options

tomsbergmanis opened this issue · comments

Marian command line options include:
--bert-mask-symbol TEXT=[MASK] Masking symbol for BERT masked-LM training
--bert-sep-symbol TEXT=[SEP] Sentence separator symbol for BERT next sentence prediction training
--bert-class-symbol TEXT=[CLS] Class symbol BERT classifier training
--bert-masking-fraction FLOAT=0.15 Fraction of masked out tokens during training
--bert-train-type-embeddings=true Train bert type embeddings, set to false to use static sinusoidal embeddings
--bert-type-vocab-size INT=2 Size of BERT type vocab (sentence A and B)
Yet I can not find example of how and for what these can be used.
It would be good to have simple use case example on how to use these to do BERT pretraining.

Even a code example in a comment below would be really good!

Hi Toms, I've never used it and don't know what it actually does. Pinging @emjotde, who may know. It might be that it was an experimental feature.

@tomsbergmanis Did you ever figure out how to use these?

(I have need of a cpu-decoding classifier for Norwegian text and marian seems promising
cf. https://groups.google.com/g/marian-nmt/c/iVq-jGa3N8M but I don't know if it's possible to use it that way)