chenggr (123penny123)

123penny123

Geek Repo

Company:Southeast University

Github PK Tool:Github PK Tool

chenggr's starred repositories

grok-1

Grok open release

Language:PythonLicense:Apache-2.0Stargazers:49198Issues:560Issues:202

DeepSpeed

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

Language:PythonLicense:Apache-2.0Stargazers:33979Issues:341Issues:2655

LLaMA-Factory

A WebUI for Efficient Fine-Tuning of 100+ LLMs (ACL 2024)

Language:PythonLicense:Apache-2.0Stargazers:27521Issues:185Issues:4364

Qwen

The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.

Language:PythonLicense:Apache-2.0Stargazers:12763Issues:98Issues:1032

trl

Train transformer language models with reinforcement learning.

Language:PythonLicense:Apache-2.0Stargazers:8853Issues:76Issues:1007

Qwen2

Qwen2 is the large language model series developed by Qwen team, Alibaba Cloud.

GroundingDINO

[ECCV 2024] Official implementation of the paper "Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection"

Language:PythonLicense:Apache-2.0Stargazers:5767Issues:37Issues:287

trlx

A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)

Language:PythonLicense:MITStargazers:4403Issues:49Issues:285

alignment-handbook

Robust recipes to align language models with human and AI preferences

Language:PythonLicense:Apache-2.0Stargazers:4257Issues:111Issues:124

awesome-RLHF

A curated list of reinforcement learning with human feedback resources (continually updated)

MedicalGPT

MedicalGPT: Training Your Own Medical GPT Model with ChatGPT Training Pipeline. 训练医疗大模型,实现了包括增量预训练(PT)、有监督微调(SFT)、RLHF、DPO、ORPO。

Language:PythonLicense:Apache-2.0Stargazers:3011Issues:33Issues:368

DecryptPrompt

总结Prompt&LLM论文,开源数据&模型,AIGC应用

RL4LMs

A modular RL library to fine-tune language models to human preferences

Language:PythonLicense:Apache-2.0Stargazers:2134Issues:26Issues:54

vall-e

PyTorch implementation of VALL-E(Zero-Shot Text-To-Speech), Reproduced Demo https://lifeiteng.github.io/valle/index.html

Language:PythonLicense:Apache-2.0Stargazers:1951Issues:50Issues:125

direct-preference-optimization

Reference implementation for DPO (Direct Preference Optimization)

Language:PythonLicense:Apache-2.0Stargazers:1896Issues:19Issues:77

Artificial-Intelligence-Terminology-Database

A comprehensive mapping database of English to Chinese technical vocabulary in the artificial intelligence domain

Llama-X

Open Academic Research on Improving LLaMA to SOTA LLM

Language:PythonLicense:Apache-2.0Stargazers:1582Issues:41Issues:21

self-rewarding-lm-pytorch

Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAI

Language:PythonLicense:MITStargazers:1273Issues:23Issues:17

MOSS-RLHF

MOSS-RLHF

Language:PythonLicense:Apache-2.0Stargazers:1235Issues:34Issues:51

machine-learning-interview

算法工程师-机器学习面试题总结

Xwin-LM

Xwin-LM: Powerful, Stable, and Reproducible LLM Alignment

AI-interview-cards

最完整的AI算法面试题目仓库,1000道,25个类目

Paper-Reading-ConvAI

📖 Paper reading list in conversational AI (constantly updating 🤗).

LLM-Blender

[ACL2023] We introduce LLM-Blender, an innovative ensembling framework to attain consistently superior performance by leveraging the diverse strengths of multiple open-source LLMs. LLM-Blender cut the weaknesses through ranking and integrate the strengths through fusing generation to enhance the capability of LLMs.

Language:PythonLicense:Apache-2.0Stargazers:833Issues:15Issues:23

alpaca_farm

A simulation framework for RLHF and alternatives. Develop your RLHF method without collecting human data.

Language:PythonLicense:Apache-2.0Stargazers:742Issues:8Issues:41

KnowledgeEditingPapers

[知识编辑] Must-read Papers on Knowledge Editing for Large Language Models.

Language:JavaScriptLicense:NOASSERTIONStargazers:718Issues:8Issues:0

Multitask-Learning

Awesome Multitask Learning Resources

UltraFeedback

A large-scale, fine-grained, diverse preference dataset (and models).

Language:PythonLicense:MITStargazers:285Issues:10Issues:13

Pixel-Navigator

Official GitHub Repository for Paper "Bridging Zero-shot Object Navigation and Foundation Models through Pixel-Guided Navigation Skill", ICRA 2024