i4never's starred repositories
ChatGLM-6B
ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型
Megatron-LM
Ongoing research training transformer models at scale
FlagEmbedding
Retrieval and Retrieval-augmented LLMs
RedPajama-Data
The RedPajama-Data repository contains code for preparing large datasets for training large language models.
trafilatura
Python & Command-line tool to gather text and metadata on the Web: Crawling, scraping, extraction, output as CSV, JSON, HTML, MD, TXT, XML
HunyuanDiT
Hunyuan-DiT : A Powerful Multi-Resolution Diffusion Transformer with Fine-Grained Chinese Understanding
Chinese-LangChain
中文langchain项目|小必应,Q.Talk,强聊,QiangTalk
chain-of-thought-hub
Benchmarking large language models' complex reasoning ability with chain-of-thought prompting
simple-computer
the scott CPU from "But How Do It Know?" by J. Clark Scott
TransformerEngine
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
clarity-ai
Come join the best place on the internet to learn AI skills. Use code "clarityai" for an extra 20% off.
selfcheckgpt
SelfCheckGPT: Zero-Resource Black-Box Hallucination Detection for Generative Large Language Models
long-context-attention
USP: Unified (a.k.a. Hybrid, 2D) Sequence Parallel Attention for Long Context Transformers Model Training and Inference
perplexityai
A python api to use perplexity.ai