Yingfei(Jeremy) Xiang's repositories
AttnPacker
Code and Pre-Trained Models for "AttnPacker: An end-to-end deep learning method for protein side-chain packing"
ContinualLM
An Extensible Continual Learning Framework Focused on Language Models (LMs)
efficient-evolution
Efficient evolution from protein language models
Fengshenbang-LM
Fengshenbang-LM(封神榜大模型)是IDEA研究院认知计算与自然语言研究中心主导的大模型开源体系,成为中文AIGC和认知智能的基础设施。
how-to-train-tokenizer
怎么训练一个LLM分词器
long_llama
LongLLaMA is a large language model capable of handling long contexts. It is based on OpenLLaMA and fine-tuned with the Focused Transformer (FoT) method.
mixture-of-experts
PyTorch Re-Implementation of "The Sparsely-Gated Mixture-of-Experts Layer" by Noam Shazeer et al. https://arxiv.org/abs/1701.06538
multipack_sampler
Multipack distributed sampler for fast padding-free training of LLMs
PaddleFleetX
Paddle Distributed Training Examples. 飞桨分布式训练示例 Resnet Bert GPT MOE DataParallel ModelParallel PipelineParallel HybridParallel AutoParallel Zero Sharding Recompute GradientMerge Offload AMP DGC LocalSGD Wide&Deep
RFdiffusion
Code for running RFdiffusion
sentencepiece_chinese_bpe
使用sentencepiece中BPE训练中文词表,并在transformers中进行使用。
sk-iterative-planner
Iterative Planner for Semantic Kernel
Stable-Alignment
Multi-agent Social Simulation + Efficient, Effective, and Stable alternative of RLHF. Code for the paper "Training Socially Aligned Language Models in Simulated Human Society".
mesh
Mesh TensorFlow: Model Parallelism Made Easier
MT-LLaMA
Multi-Task instruction-tuned LLaMA
PdfGptIndexer
An efficient tool for indexing and searching PDF text data using OpenAI API and FAISS (Facebook AI Similarity Search) index, designed for rapid information retrieval and superior search accuracy.
PLSC
Paddle Large Scale Classification Tools,supports ArcFace, CosFace, PartialFC, Data Parallel + Model Parallel. Model includes ResNet, ViT, Swin, DeiT, CaiT, FaceViT, MoCo, MAE, ConvMAE, CAE.