There are 5 repositories under pre-training topic.
A one-stop data processing system to make data higher-quality, juicier, and more digestible for LLMs! 🍎 🍋 🌽 ➡️ ➡️🍸 🍹 🍷为大语言模型提供更高质量、更丰富、更易”消化“的数据!
Papers about pretraining and self-supervised learning on Graph Neural Networks (GNN).
ms-swift: Use PEFT or Full-parameter to finetune 200+ LLMs or 15+ MLLMs
Awesome resources for in-context learning and prompt engineering: Mastery of the LLMs such as ChatGPT, GPT-3, and FlanT5, with up-to-date and cutting-edge updates.
Code for TKDE paper "Self-supervised learning on graphs: Contrastive, generative, or predictive"
Awesome list for research on CLIP (Contrastive Language-Image Pre-Training).
Tencent Pre-training framework in PyTorch & Pre-trained Model Zoo
Pre-training of Deep Bidirectional Transformers for Language Understanding: pre-train TextCNN
A professional list on Large (Language) Models and Foundation Models (LLM, LM, FM) for Time Series, Spatiotemporal, and Event Data.
Unified Training of Universal Time Series Forecasting Transformers
[ICLR 2024] Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruning
Large Language Model-enhanced Recommender System Papers
Conceptual 12M is a dataset containing (image-URL, caption) pairs collected for vision-and-language pre-training.
Probing the representations of Vision Transformers.
The repository of ET-BERT, a network traffic classification model on encrypted traffic. The work has been accepted as The Web Conference (WWW) 2022 accepted paper.
[CVPR2023] All in One: Exploring Unified Video-Language Pre-training
💐Kaleido-BERT: Vision-Language Pre-training on Fashion Domain
[MIR-2023-Survey] A continuously updated paper list for multi-modal pre-trained big models
OpenAI GPT2 pre-training and sequence prediction implementation in Tensorflow 2.0
GearNet and Geometric Pretraining Methods for Protein Structure Representation Learning, ICLR'2023 (https://arxiv.org/abs/2203.06125)
[Survey] Masked Modeling for Self-supervised Representation Learning on Vision and Beyond (https://arxiv.org/abs/2401.00897)
Paper List of Pre-trained Foundation Recommender Models
The official repo for [NeurIPS'23] "SAMRS: Scaling-up Remote Sensing Segmentation Dataset with Segment Anything Model"
A simple and working implementation of Electra, the fastest way to pretrain language models from scratch, in Pytorch