Atnafu Lambebo Tonja's repositories
LLM-Finetuning-Toolkit
Toolkit for fine-tuning, ablating and unit-testing open-source LLMs.
academicpages.github.io
Github Pages template for academic personal websites, forked from mmistakes/minimal-mistakes
ALMA
This is repository for ALMA translation models.
alpaca-lora
Instruct-tune LLaMA on consumer hardware
arxiv2024-triple-encoders
triple-encoders is a library for contextualizing distributed Sentence Transformers representations.
Aurora
🐳 Aurora is a [Chinese Version] MoE model. Aurora is a further work based on Mixtral-8x7B, which activates the chat capability of the model's Chinese open domain.
axolotl
Go ahead and axolotl questions
bible_scapper
python script that allows to scrape any version of bible from bible.com
Chinese-LLaMA-Alpaca
中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
Chinese-Mixtral
中文Mixtral混合专家大模型(Chinese Mixtral MoE LLMs)
datatrove
Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.
evals
Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.
fairseq
Facebook AI Research Sequence-to-Sequence Toolkit written in Python.
indaba-pracs-2023
Notebooks for the Practicals at the Deep Learning Indaba 2023.
Llama-2
All the projects related to Llama
llm-translator
Mixtral-based Ja-En (En-Ja) Translation model
lm-evaluation-harness
A framework for few-shot evaluation of language models.
masakhane-news
MasakhaNEWS: News Topic Classification for African Languages
masakhane-pos
POS for African languages
mgpt
Multilingual Generative Pretrained Model
mot
Multilingual Open Text
OpenMoE
A family of open-sourced Mixture-of-Experts (MoE) Large Language Models
sentence-transformers
Multilingual Sentence & Image Embeddings with BERT
simple-nmt
This repo contains a simple source code for advanced neural machine translation based on sequence-to-sequence.
TinyLlama
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
TPU-Alignment
Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free