There are 10 repositories under qlora topic.
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
Fine-tuning & Reinforcement Learning for LLMs. 𦄠Train OpenAI gpt-oss, Qwen3, Llama 4, DeepSeek-R1, Gemma 3, TTS 2x faster with 70% less VRAM.
Accessible large language models via k-bit quantization for PyTorch.
Firefly: 大樔åč®ē»å·„å ·ļ¼ęÆęč®ē»Qwen2.5ćQwen2ćYi1.5ćPhi-3ćLlama3ćGemmaćMiniCPMćYićDeepseekćOrionćXversećMixtral-8x7BćZephyrćMistralćBaichuan2ćLlma2ćLlamaćQwenćBaichuanćChatGLM2ćInternLMćZiya2ćVicunaćBloomē大樔å
Fine-tuning ChatGLM-6B with PEFT | åŗäŗ PEFT ēé«ę ChatGLM å¾®č°
š¦ šš²š®šæš» about ššš š, ššš š¢š½š, and šš²š°šš¼šæ ššš for free by designing, training, and deploying a real-time financial advisor LLM system ~ š“š°š¶š³š¤š¦ š¤š°š„š¦ + š·šŖš„š¦š° & š³š¦š¢š„šŖšÆšØ š®š¢šµš¦š³šŖš¢šš“
chatglm 6b finetuning and alpaca finetuning
Toolkit for fine-tuning, ablating and unit-testing open-source LLMs.
Easy and Efficient Finetuning LLMs. (Supported LLama, LLama2, LLama3, Qwen, Baichuan, GLM , Falcon) 大樔åé«ęéåč®ē»+éØē½².
šæåęéäøęå»ē大樔å(Sunsimiao)ļ¼ęä¾å®å ØćåÆé ćę®ę ēäøęå»ē大樔å
FireflyäøęLLaMA-2大樔åļ¼ęÆęå¢éé¢č®ē»Baichuan2ćLlama2ćLlamaćFalconćQwenćBaichuanćInternLMćBloomē大樔å
End to End Generative AI Industry Projects on LLM Models with Deployment_Awesome LLM Projects
Finetuning of Falcon-7B LLM using QLoRA on Mental Health Conversational Dataset
The official codes for "Aurora: Activating chinese chat capability for Mixtral-8x7B sparse Mixture-of-Experts through Instruction-Tuning"
对llama3čæč”å Øåå¾®č°ćloraå¾®č°ä»„åqloraå¾®č°ć
LongQLoRA: Extent Context Length of LLMs Efficiently
Large language Model fintuning bloom , opt , gpt, gpt2 ,llama,llama-2,cpmant and so on
使ēØqlora对äøę大čÆčØęØ”åčæč”å¾®č°ļ¼å å«ChatGLMćChinese-LLaMA-AlpacaćBELLE
Fine-Tuning Falcon-7B, LLAMA 2 with QLoRA to create an advanced AI model with a profound understanding of the Indian legal context.
Use QLoRA to tune LLM in PyTorch-Lightning w/ Huggingface + MLflow
VerifAI initiative to build open-source easy-to-deploy generative question-answering engine that can reference and verify answers for correctness (using posteriori model)
Finetune any model on HF in less than 30 seconds
Implements pre-training, supervised fine-tuning (SFT), and reinforcement learning from human feedback (RLHF), to train and fine-tune the LLaMA2 model to follow human instructions, similar to InstructGPT or ChatGPT, but on a much smaller scale.
Tuning the Finetuning: An exploration of achieving success with QLoRA
A Gradio web UI for Large Language Models. Supports LoRA/QLoRA finetuning,RAG(Retrieval-augmented generation) and Chat
baichuan and baichuan2 finetuning and alpaca finetuning
META LLAMA3 GENAI Real World UseCases End To End Implementation Guide
An LLM training library for instruction-tuning.