JustQJ / LLM_Materials

记录平时阅读的有关大模型的一些材料

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

LLM Materials

Recording some materials about LLM.

LLM Training

Techniques

Datasets

LLM Evaluation

LLM Architecture

LLM Compression

Survey

  • [Arxiv 2024] A Survey of Resource-efficient LLM and Multimodal Foundation Models. This survey delves into the critical importance of such research, examining both algorithmic and systemic aspects. It offers a comprehensive analysis and valuable insights gleaned from existing literature, encompassing a broad array of topics from cutting-edge model architectures and training/serving algorithms to practical system designs and implementations.
  • [Arxiv 2024] LLM Inference Unveiled: Survey and Roofline Model Insights. not only summarizing the current state of research but also by introducing a framework based on roofline model for systematic analysis of LLM inference techniques. This framework identifies the bottlenecks when deploying LLMs on hardware devices and provides a clear understanding of practical problems, such as why LLMs are memory-bound, how much memory and computation they need, and how to choose the right hardware. Code
  • [Arxiv 2023] A Survey on Model Compression for Large Language Models.

Low-rank Decomposition

Pruning

Sparsity

Quantization

LLM System

Survey

  • [Arxiv 2024] A Survey of Resource-efficient LLM and Multimodal Foundation Models. This survey delves into the critical importance of such research, examining both algorithmic and systemic aspects. It offers a comprehensive analysis and valuable insights gleaned from existing literature, encompassing a broad array of topics from cutting-edge model architectures and training/serving algorithms to practical system designs and implementations.
  • [Arxiv 2024] LLM Inference Unveiled: Survey and Roofline Model Insights. not only summarizing the current state of research but also by introducing a framework based on roofline model for systematic analysis of LLM inference techniques. This framework identifies the bottlenecks when deploying LLMs on hardware devices and provides a clear understanding of practical problems, such as why LLMs are memory-bound, how much memory and computation they need, and how to choose the right hardware. Code

Training System

Inference System

Some Collections

Projects

Multimodel LLM

Survey

About

记录平时阅读的有关大模型的一些材料