There are 12 repositories under instruction-tuning topic.
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
Unify Efficient Fine-tuning of 100+ LLMs
:sparkles::sparkles:Latest Papers and Datasets on Multimodal Large Language Models, and Their Evaluation.
Instruction Tuning with GPT-4
Aligning pretrained language models with instruction data generated by themselves.
We unified the interfaces of instruction-tuning data (e.g., CoT data), multiple LLMs and parameter-efficient methods (e.g., lora, p-tuning) together for easy use. We welcome open-source enthusiasts to initiate any meaningful PR on this repo and integrate as many LLM related technologies as possible. 我们打造了方便研究人员上手和使用大模型等微调平台,我们欢迎开源爱好者发起任何有意义的pr!
Video-LLaVA: Learning United Visual Representation by Alignment Before Projection
总结Prompt&LLM论文,开源数据&模型,AIGC应用
InternLM-XComposer2 is a groundbreaking vision-language large model (VLLM) excelling in free-form text-image composition and comprehension.
A one-stop data processing system to make data higher-quality, juicier, and more digestible for LLMs! 🍎 🍋 🌽 ➡️ ➡️🍸 🍹 🍷为大语言模型提供更高质量、更丰富、更易”消化“的数据!
A collection of open-source dataset to train instruction-following LLMs (ChatGPT,LLaMA,Alpaca)
InternVideo: General Video Foundation Models via Generative and Discriminative Learning (https://arxiv.org/abs/2212.03191)
DataDreamer: Prompt. Generate Synthetic Data. Train & Align Models. 🤖💤
Crosslingual Generalization through Multitask Finetuning
DialogStudio: Towards Richest and Most Diverse Unified Dataset Collection and Instruction-Aware Models for Conversational AI
DISC-FinLLM,中文金融大语言模型(LLM),旨在为用户提供金融场景下专业、智能、全面的金融咨询服务。DISC-FinLLM, a Chinese financial large language model (LLM) designed to provide users with professional, intelligent, and comprehensive financial consulting services in financial scenarios.
Papers and Datasets on Instruction Tuning and Following. ✨✨✨
Generative Representational Instruction Tuning
Research Trends in LLM-guided Multimodal Learning.
MindSpore online courses: Step into LLM
🐳 Aurora is a [Chinese Version] MoE model. Aurora is a further work based on Mixtral-8x7B, which activates the chat capability of the model's Chinese open domain.
A curated list of awesome instruction tuning datasets, models, papers and repositories.
All available datasets for Instruction Tuning of Large Language Models
Code for instruction-tuning Stable Diffusion.