Haoxiang Wang (Haoxiang-Wang)

Haoxiang-Wang

Geek Repo

Company:University of Illinois, Urbana-Champaign

Location:Urbana

Home Page:https://haoxiang-wang.github.io/

Twitter:@Haoxiang__Wang

Github PK Tool:Github PK Tool


Organizations
AI-secure

Haoxiang Wang's starred repositories

fms-fsdp

🚀 Efficiently (pre)training foundation models with native PyTorch features, including FSDP for training and SDPA implementation of Flash attention v2.

Language:PythonLicense:Apache-2.0Stargazers:130Issues:0Issues:0

SciCode

A benchmark that challenges language models to code solutions for scientific problems

Language:PythonLicense:Apache-2.0Stargazers:57Issues:0Issues:0

PINE

Offcial Repo of Paper "Eliminating Position Bias of Language Models: A Mechanistic Approach""

Language:PythonStargazers:6Issues:0Issues:0

Adam-mini

Code for Adam-mini: Use Fewer Learning Rates To Gain More https://arxiv.org/abs/2406.16793

Language:PythonStargazers:226Issues:0Issues:0

reward-bench

RewardBench: the first evaluation tool for reward models.

Language:PythonLicense:Apache-2.0Stargazers:308Issues:0Issues:0

bedirt.github.io

My personal website

Language:HTMLStargazers:2Issues:0Issues:0
Language:HTMLStargazers:1Issues:0Issues:0

hugo-PaperMod

A fast, clean, responsive Hugo theme.

Language:HTMLLicense:MITStargazers:9290Issues:0Issues:0

rlftqc

Reinforcement Learning for Fault-Tolerant Quantum Circuit Discovery

Language:PythonLicense:MITStargazers:6Issues:0Issues:0

qdx

Quantum error correction code AI-discovery with Jax

Language:Jupyter NotebookLicense:MITStargazers:10Issues:0Issues:0

CodeUltraFeedback

CodeUltraFeedback: aligning large language models to coding preferences

Language:PythonLicense:MITStargazers:57Issues:0Issues:0

2025QuantInternships

Public quant internship repository, maintained by NUFT but available for everyone.

Stargazers:1079Issues:0Issues:0

RLHF-Reward-Modeling

Recipes to train reward model for RLHF.

Language:PythonLicense:Apache-2.0Stargazers:508Issues:0Issues:0

grok-1

Grok open release

Language:PythonLicense:Apache-2.0Stargazers:49218Issues:0Issues:0

Directional-Preference-Alignment

Directional Preference Alignment

License:Apache-2.0Stargazers:43Issues:0Issues:0

flash-attention

Fast and memory-efficient exact attention

Language:PythonLicense:BSD-3-ClauseStargazers:12659Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:545Issues:0Issues:0

easse

Easier Automatic Sentence Simplification Evaluation

Language:RoffLicense:GPL-3.0Stargazers:156Issues:0Issues:0

NeMo

A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)

Language:PythonLicense:Apache-2.0Stargazers:11106Issues:0Issues:0

prometheus

[ICLR 2024 & NeurIPS 2023 WS] An Evaluator LM that is open-source, offers reproducible evaluation, and inexpensive to use. Specifically designed for fine-grained evaluation on a customized score rubric, Prometheus is a good alternative for human evaluation and GPT-4 evaluation.

Language:PythonLicense:MITStargazers:281Issues:0Issues:0

Otter

🦦 Otter, a multi-modal model based on OpenFlamingo (open-sourced version of DeepMind's Flamingo), trained on MIMIC-IT and showcasing improved instruction-following and in-context learning ability.

Language:PythonLicense:MITStargazers:3527Issues:0Issues:0

trl

Train transformer language models with reinforcement learning.

Language:PythonLicense:Apache-2.0Stargazers:8900Issues:0Issues:0
Language:PythonLicense:MITStargazers:8Issues:0Issues:0

LLaMA-Factory

A WebUI for Efficient Fine-Tuning of 100+ LLMs (ACL 2024)

Language:PythonLicense:Apache-2.0Stargazers:28056Issues:0Issues:0

FLASK

[ICLR 2024 Spotlight] FLASK: Fine-grained Language Model Evaluation based on Alignment Skill Sets

Language:PythonStargazers:207Issues:0Issues:0

lm-evaluation-harness

A framework for few-shot evaluation of language models.

Language:PythonLicense:MITStargazers:5996Issues:0Issues:0

axolotl

Go ahead and axolotl questions

Language:PythonLicense:Apache-2.0Stargazers:7099Issues:0Issues:0

tensor_parallel

Automatically split your PyTorch models on multiple GPUs for training & inference

Language:PythonLicense:MITStargazers:605Issues:0Issues:0
Language:PythonStargazers:127Issues:0Issues:0

mint-bench

Official Repo for ICLR 2024 paper MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback by Xingyao Wang*, Zihan Wang*, Jiateng Liu, Yangyi Chen, Lifan Yuan, Hao Peng and Heng Ji.

Language:PythonLicense:Apache-2.0Stargazers:100Issues:0Issues:0