Hao (TissueC)

TissueC

Geek Repo

Company:CoAI of Tsinghua University @thu-coai

Location:Beijing

Home Page:sunhao.site

Github PK Tool:Github PK Tool


Organizations
thu-coai

Hao's starred repositories

PaperMemory

Your browser's reference manager: automatic paper detection (Arxiv, OpenReview & more), publication venue matching and code repository discovery! Also enhances ArXiv: BibTex citation, Markdown link, direct download and more!

Language:JavaScriptLicense:MITStargazers:478Issues:0Issues:0

GLM-4

GLM-4 series: Open Multilingual Multimodal Chat LMs | 开源多语言多模态对话模型

Language:PythonLicense:Apache-2.0Stargazers:3899Issues:0Issues:0
Language:PythonStargazers:775Issues:0Issues:0

lighteval

LightEval is a lightweight LLM evaluation suite that Hugging Face has been using internally with the recently released LLM data processing library datatrove and LLM training library nanotron.

Language:PythonLicense:MITStargazers:495Issues:0Issues:0

elasticsearch

Free and Open, Distributed, RESTful Search Engine

Language:JavaLicense:NOASSERTIONStargazers:68711Issues:0Issues:0

LLM-Extrapolation

Official repository for paper "Weak-to-Strong Extrapolation Expedites Alignment"

Language:PythonStargazers:56Issues:0Issues:0

DeepSeek-VL

DeepSeek-VL: Towards Real-World Vision-Language Understanding

Language:PythonLicense:MITStargazers:1900Issues:0Issues:0

mup

maximal update parametrization (µP)

Language:Jupyter NotebookLicense:MITStargazers:1255Issues:0Issues:0

fm-cheatsheet

Website for hosting the Open Foundation Models Cheat Sheet.

Language:JavaScriptStargazers:251Issues:0Issues:0

Qwen2

Qwen2 is the large language model series developed by Qwen team, Alibaba Cloud.

Language:ShellStargazers:6498Issues:0Issues:0

outlines

Structured Text Generation

Language:PythonLicense:Apache-2.0Stargazers:7325Issues:0Issues:0

OLMo

Modeling, training, eval, and inference code for OLMo

Language:PythonLicense:Apache-2.0Stargazers:4248Issues:0Issues:0

MiniCPM

MiniCPM-2B: An end-side LLM outperforming Llama2-13B.

Language:PythonLicense:Apache-2.0Stargazers:4460Issues:0Issues:0

llama-moe

⛷️ LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training

Language:PythonLicense:Apache-2.0Stargazers:810Issues:0Issues:0
Language:PythonLicense:MITStargazers:41Issues:0Issues:0

DeepSeek-MoE

DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models

Language:PythonLicense:MITStargazers:931Issues:0Issues:0
Stargazers:71Issues:0Issues:0

FollowBench

Code for "FollowBench: A Multi-level Fine-grained Constraints Following Benchmark for Large Language Models (ACL 2024)"

Language:PythonLicense:Apache-2.0Stargazers:62Issues:0Issues:0

Pai-Megatron-Patch

The official repo of Pai-Megatron-Patch for LLM & VLM large scale training developed by Alibaba Cloud.

Language:PythonLicense:Apache-2.0Stargazers:579Issues:0Issues:0
Language:PythonStargazers:108Issues:0Issues:0

mergekit

Tools for merging pretrained large language models.

Language:PythonLicense:LGPL-3.0Stargazers:4172Issues:0Issues:0

Yuan-2.0

Yuan 2.0 Large Language Model

Language:PythonLicense:NOASSERTIONStargazers:672Issues:0Issues:0

tensor_parallel

Automatically split your PyTorch models on multiple GPUs for training & inference

Language:PythonLicense:MITStargazers:604Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:271Issues:0Issues:0

opencompass

OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.

Language:PythonLicense:Apache-2.0Stargazers:3392Issues:0Issues:0

LongQLoRA

LongQLoRA: Extent Context Length of LLMs Efficiently

Language:PythonStargazers:152Issues:0Issues:0

DeepSeek-Coder

DeepSeek Coder: Let the Code Write Itself

Language:PythonLicense:MITStargazers:6170Issues:0Issues:0

Yi

A series of large language models trained from scratch by developers @01-ai

Language:PythonLicense:Apache-2.0Stargazers:7506Issues:0Issues:0

ray

Ray is a unified framework for scaling AI and Python applications. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.

Language:PythonLicense:Apache-2.0Stargazers:32232Issues:0Issues:0

FasterTransformer

Transformer related optimization, including BERT, GPT

Language:C++License:Apache-2.0Stargazers:5681Issues:0Issues:0