There are 16 repositories under instruction-tuning topic.
Unify Efficient Fine-Tuning of 100+ LLMs
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
:sparkles::sparkles:Latest Papers and Datasets on Multimodal Large Language Models, and Their Evaluation.
Instruction Tuning with GPT-4
Aligning pretrained language models with instruction data generated by themselves.
We unified the interfaces of instruction-tuning data (e.g., CoT data), multiple LLMs and parameter-efficient methods (e.g., lora, p-tuning) together for easy use. We welcome open-source enthusiasts to initiate any meaningful PR on this repo and integrate as many LLM related technologies as possible. 我们打造了方便研究人员上手和使用大模型等微调平台,我们欢迎开源爱好者发起任何有意义的pr!
Video-LLaVA: Learning United Visual Representation by Alignment Before Projection
总结Prompt&LLM论文,开源数据&模型,AIGC应用
InternLM-XComposer2 is a groundbreaking vision-language large model (VLLM) excelling in free-form text-image composition and comprehension.
A one-stop data processing system to make data higher-quality, juicier, and more digestible for LLMs! 🍎 🍋 🌽 ➡️ ➡️🍸 🍹 🍷为大语言模型提供更高质量、更丰富、更易”消化“的数据!
A collection of open-source dataset to train instruction-following LLMs (ChatGPT,LLaMA,Alpaca)
Video Foundation Models & Data for Multimodal Understanding
DataDreamer: Prompt. Generate Synthetic Data. Train & Align Models. 🤖💤
Crosslingual Generalization through Multitask Finetuning
DialogStudio: Towards Richest and Most Diverse Unified Dataset Collection and Instruction-Aware Models for Conversational AI
DISC-FinLLM,中文金融大语言模型(LLM),旨在为用户提供金融场景下专业、智能、全面的金融咨询服务。DISC-FinLLM, a Chinese financial large language model (LLM) designed to provide users with professional, intelligent, and comprehensive financial consulting services in financial scenarios.
Generative Representational Instruction Tuning
Papers and Datasets on Instruction Tuning and Following. ✨✨✨
MindSpore online courses: Step into LLM
Research Trends in LLM-guided Multimodal Learning.
🐳 Aurora is a [Chinese Version] MoE model. Aurora is a further work based on Mixtral-8x7B, which activates the chat capability of the model's Chinese open domain.
A curated list of awesome instruction tuning datasets, models, papers and repositories.
Preprint: Less: Selecting Influential Data for Targeted Instruction Tuning
All available datasets for Instruction Tuning of Large Language Models