RunningLeon's repositories
mmdetection
OpenMMLab Detection Toolbox and Benchmark
Efficient-AI-Backbones
Efficient AI Backbones including GhostNet, TNT and MLP, developed by Huawei Noah's Ark Lab.
InternLM
InternLM has open-sourced a 7 billion parameter base model, a chat model tailored for practical scenarios and the training system.
llama.cpp
LLM inference in C/C++
llamafile
Distribute and run LLMs with a single file.
mmclassification
OpenMMLab Image Classification Toolbox and Benchmark
MobileSAM
This is the official code for MobileSAM project that makes SAM lightweight for mobile applications and beyond!
Multimodal-GPT
Multimodal-GPT
MyST-Parser
An extended commonmark compliant parser, with bridges to docutils/sphinx
onnx-simplifier
Simplify your onnx model
opencompass
OpenCompass is an LLM evaluation platform, supporting a wide range of models (LLaMA, LLaMa2, ChatGLM2, ChatGPT, Claude, etc) over 50+ datasets.
TensorRT
NVIDIA® TensorRT™, an SDK for high-performance deep learning inference, includes a deep learning inference optimizer and runtime that delivers low latency and high throughput for inference applications.
TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.