Christmas's starred repositories

build-your-own-x

Master programming by recreating your favorite technologies from scratch.

Stargazers:283557Issues:0Issues:0

Natural_Language_Processing_with_Transformers

Natural Language Processing with Transformers 中译本,最权威Transformers教程

Stargazers:333Issues:0Issues:0

opencompass

OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.

Language:PythonLicense:Apache-2.0Stargazers:3220Issues:0Issues:0

how-to-optim-algorithm-in-cuda

how to optimize some algorithm in cuda.

Language:CudaStargazers:1187Issues:0Issues:0

CEPE

[ACL 2024] Long-Context Language Modeling with Parallel Encodings

Language:PythonLicense:MITStargazers:117Issues:0Issues:0

mmrazor

OpenMMLab Model Compression Toolbox and Benchmark.

Language:PythonLicense:Apache-2.0Stargazers:1412Issues:0Issues:0

EAGLE

Official Implementation of EAGLE-1 and EAGLE-2

Language:PythonLicense:Apache-2.0Stargazers:646Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:1040Issues:0Issues:0

fastllm

纯c++的全平台llm加速库,支持python调用,chatglm-6B级模型单卡可达10000+token / s,支持glm, llama, moss基座,手机端流畅运行

Language:C++License:Apache-2.0Stargazers:3200Issues:0Issues:0

ntk_alibi

NTK scaled version of ALiBi position encoding in Transformer.

Stargazers:61Issues:0Issues:0

laser

The Truth Is In There: Improving Reasoning in Language Models with Layer-Selective Rank Reduction

Language:PythonLicense:MITStargazers:344Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:56Issues:0Issues:0

roformer

Rotary Transformer

Language:PythonLicense:Apache-2.0Stargazers:740Issues:0Issues:0

Chinese-Vicuna

Chinese-Vicuna: A Chinese Instruction-following LLaMA-based Model —— 一个中文低资源的llama+lora方案,结构参考alpaca

Language:CLicense:Apache-2.0Stargazers:4144Issues:0Issues:0

lm-evaluation-harness

A framework for few-shot evaluation of language models.

Language:PythonLicense:MITStargazers:5809Issues:0Issues:0

exllama

A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.

Language:PythonLicense:MITStargazers:2668Issues:0Issues:0

baichuan-Dynamic-NTK-ALiBi

百川Dynamic NTK-ALiBi的代码实现:无需微调即可推理更长文本

Language:PythonStargazers:45Issues:0Issues:0

LongBench

LongBench: A Bilingual, Multitask Benchmark for Long Context Understanding

Language:PythonLicense:MITStargazers:539Issues:0Issues:0

VisualGLM-6B

Chinese and English multimodal conversational language model | 多模态中英双语对话语言模型

Language:PythonLicense:Apache-2.0Stargazers:4036Issues:0Issues:0

CogView2

official code repo for paper "CogView2: Faster and Better Text-to-Image Generation via Hierarchical Transformers"

Language:PythonLicense:Apache-2.0Stargazers:932Issues:0Issues:0

CogView

Text-to-Image generation. The repo for NeurIPS 2021 paper "CogView: Mastering Text-to-Image Generation via Transformers".

Language:PythonLicense:Apache-2.0Stargazers:1631Issues:0Issues:0

CodeGeeX2

CodeGeeX2: A More Powerful Multilingual Code Generation Model

Language:PythonLicense:Apache-2.0Stargazers:7600Issues:0Issues:0

CodeGeeX

CodeGeeX: An Open Multilingual Code Generation Model (KDD 2023)

Language:PythonLicense:Apache-2.0Stargazers:7966Issues:0Issues:0

ChatGLM3

ChatGLM3 series: Open Bilingual Chat LLMs | 开源双语对话语言模型

Language:PythonLicense:Apache-2.0Stargazers:13037Issues:0Issues:0

reformer-pytorch

Reformer, the efficient Transformer, in Pytorch

Language:PythonLicense:MITStargazers:2078Issues:0Issues:0

yarn

YaRN: Efficient Context Window Extension of Large Language Models

Language:PythonLicense:MITStargazers:1256Issues:0Issues:0

Long-Context

This repository contains code and tooling for the Abacus.AI LLM Context Expansion project. Also included are evaluation scripts and benchmark tasks that evaluate a model’s information retrieval capabilities with context expansion. We also include key experimental results and instructions for reproducing and building on them.

Language:PythonLicense:Apache-2.0Stargazers:568Issues:0Issues:0

CLEX

[ICLR 2024] CLEX: Continuous Length Extrapolation for Large Language Models

Language:PythonLicense:MITStargazers:70Issues:0Issues:0

LongLoRA

Code and documents of LongLoRA and LongAlpaca (ICLR 2024 Oral)

Language:PythonLicense:Apache-2.0Stargazers:2552Issues:0Issues:0

LLMLingua

To speed up LLMs' inference and enhance LLM's perceive of key information, compress the prompt and KV-Cache, which achieves up to 20x compression with minimal performance loss.

Language:PythonLicense:MITStargazers:4178Issues:0Issues:0