huu4ontocord / M3rlin

Multilingual, Multimodal, Multidomain (M3) Model

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

M3rlin

Multilingual, Multimodal, Multidomain (M3) Model

Details

  • We are using training code Openlm and FMengine which can run on JUWELS
  • The code in this repo is the M3rlin specific code, which is data loading and interleaving of embeddings and an extra mse loss.
  • Clone or pip install the openlm or FMengine code directly to use for training.

TODO:

  • add extraction of image or embeddings form hf dataset or jsonl (jsonl is usually faster)
  • test that the embedding is saved to webdataset format
  • test loading embeddings and token ids in train.py
  • write the code to insert token_id into the token sequences
  • embeddings should be saved in a 3D tensor (batch, embedding_id, embedding_dim) and returned
  • positions are an array of 3D tensor (batch, sequence_id, column_id)
  • need to add MSE loss
  • confirm and test MSE loss
  • Add up-projection and down-projection for embedings input and output embedding
  • Add Peft and freezing base model

About

Multilingual, Multimodal, Multidomain (M3) Model

License:Apache License 2.0


Languages

Language:Python 93.5%Language:Shell 6.5%