Haoxiang Wang's starred repositories
LLaMA-Factory
A WebUI for Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
mistral-src
Reference implementation of Mistral AI 7B v0.1 model.
llama-recipes
Scripts for fine-tuning Llama2 with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization & question answering. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment.Demo apps to showcase Llama2 for WhatsApp & Messenger
mmsegmentation
OpenMMLab Semantic Segmentation Toolbox and Benchmark.
lm-evaluation-harness
A framework for few-shot evaluation of language models.
opencompass
OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.
chain-of-thought-hub
Benchmarking large language models' complex reasoning ability with chain-of-thought prompting
ml-fastvit
This repository contains the official implementation of the research paper, "FastViT: A Fast Hybrid Vision Transformer using Structural Reparameterization" ICCV 2023
tensor_parallel
Automatically split your PyTorch models on multiple GPUs for training & inference
panopticapi
COCO 2018 Panoptic Segmentation Task API (Beta version)
prometheus
[ICLR 2024 & NeurIPS 2023 WS] An Evaluator LM that is open-source, offers reproducible evaluation, and inexpensive to use. Specifically designed for fine-grained evaluation on a customized score rubric, Prometheus is a good alternative for human evaluation and GPT-4 evaluation.
mint-bench
Official Repo for ICLR 2024 paper MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback by Xingyao Wang*, Zihan Wang*, Jiateng Liu, Yangyi Chen, Lifan Yuan, Hao Peng and Heng Ji.
understanding-forgetting
Understanding Catastrophic Forgetting in Language Models via Implicit Inference