wangxi123's starred repositories

onnx-modifier

A tool to modify ONNX models in a visualization fashion, based on Netron and Flask.

Language:JavaScriptLicense:MITStargazers:1192Issues:0Issues:0

reward-bench

RewardBench: the first evaluation tool for reward models.

Language:PythonLicense:Apache-2.0Stargazers:281Issues:0Issues:0

unsloth

Finetune Llama 3, Mistral, Phi & Gemma LLMs 2-5x faster with 80% less memory

Language:PythonLicense:Apache-2.0Stargazers:12402Issues:0Issues:0

LLaMA-Factory

A WebUI for Efficient Fine-Tuning of 100+ LLMs (ACL 2024)

Language:PythonLicense:Apache-2.0Stargazers:25902Issues:0Issues:0

Yi

A series of large language models trained from scratch by developers @01-ai

Language:PythonLicense:Apache-2.0Stargazers:7451Issues:0Issues:0

text2vec

text2vec, text to vector. 文本向量表征工具,把文本转化为向量矩阵,实现了Word2Vec、RankBM25、Sentence-BERT、CoSENT等文本表征、文本相似度计算模型,开箱即用。

Language:PythonLicense:Apache-2.0Stargazers:4254Issues:0Issues:0

SuperCLUE-Math6

SuperCLUE-Math6:新一代中文原生多轮多步数学推理数据集的探索之旅

Language:PythonStargazers:32Issues:0Issues:0

RMT

(CVPR2024)RMT: Retentive Networks Meet Vision Transformer

Language:PythonStargazers:246Issues:0Issues:0

Qwen

The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.

Language:PythonLicense:Apache-2.0Stargazers:12515Issues:0Issues:0

alpaca_eval

An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:1301Issues:0Issues:0

self-rewarding-lm-pytorch

Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAI

Language:PythonLicense:MITStargazers:1268Issues:0Issues:0

direct-preference-optimization

Reference implementation for DPO (Direct Preference Optimization)

Language:PythonLicense:Apache-2.0Stargazers:1850Issues:0Issues:0

awesome-chatgpt-prompts-zh

ChatGPT 中文调教指南。各种场景使用指南。学习怎么让它听你的话。

License:MITStargazers:51409Issues:0Issues:0

narrativeqa

This repository contains the NarrativeQA dataset. It includes the list of documents with Wikipedia summaries, links to full stories, and questions and answers.

Language:ShellLicense:Apache-2.0Stargazers:442Issues:0Issues:0

Medusa

Medusa: Simple Framework for Accelerating LLM Generation with Multiple Decoding Heads

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:2016Issues:0Issues:0

bigcode-evaluation-harness

A framework for the evaluation of autoregressive code generation language models.

Language:PythonLicense:Apache-2.0Stargazers:704Issues:0Issues:0

human-eval

Code for the paper "Evaluating Large Language Models Trained on Code"

Language:PythonLicense:MITStargazers:2151Issues:0Issues:0

chain-of-thought-hub

Benchmarking large language models' complex reasoning ability with chain-of-thought prompting

Language:Jupyter NotebookLicense:MITStargazers:2443Issues:0Issues:0

vllm

A high-throughput and memory-efficient inference and serving engine for LLMs

Language:PythonLicense:Apache-2.0Stargazers:22417Issues:0Issues:0

LEval

[ACL'24] Data and code for L-Eval, a comprehensive long context language models evaluation benchmark

Language:PythonLicense:GPL-3.0Stargazers:311Issues:0Issues:0

LongChat

Official repository for LongChat and LongEval

Language:PythonLicense:Apache-2.0Stargazers:500Issues:0Issues:0

LongBench

LongBench: A Bilingual, Multitask Benchmark for Long Context Understanding

Language:PythonLicense:MITStargazers:539Issues:0Issues:0

opencompass

OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.

Language:PythonLicense:Apache-2.0Stargazers:3215Issues:0Issues:0

open_llama

OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset

License:Apache-2.0Stargazers:7273Issues:0Issues:0

memorizing-transformers-pytorch

Implementation of Memorizing Transformers (ICLR 2022), attention net augmented with indexing and retrieval of memories using approximate nearest neighbors, in Pytorch

Language:PythonLicense:MITStargazers:620Issues:0Issues:0

long_llama

LongLLaMA is a large language model capable of handling long contexts. It is based on OpenLLaMA and fine-tuned with the Focused Transformer (FoT) method.

Language:PythonLicense:Apache-2.0Stargazers:1437Issues:0Issues:0

recurrent-memory-transformer

[NeurIPS 22] [AAAI 24] Recurrent Transformer-based long-context architecture.

Language:Jupyter NotebookStargazers:747Issues:0Issues:0

Baichuan-7B

A large-scale 7B pretraining language model developed by BaiChuan-Inc.

Language:PythonLicense:Apache-2.0Stargazers:5660Issues:0Issues:0

FlagAI

FlagAI (Fast LArge-scale General AI models) is a fast, easy-to-use and extensible toolkit for large-scale model.

Language:PythonLicense:Apache-2.0Stargazers:3808Issues:0Issues:0

ceval

Official github repo for C-Eval, a Chinese evaluation suite for foundation models [NeurIPS 2023]

Language:PythonLicense:MITStargazers:1551Issues:0Issues:0