Heo, Sung's repositories
Anima
33B Chinese LLM, DPO QLORA, 100K context, AirLLM 70B inference with single 4GB GPU
autogluon
Fast and Accurate ML in 3 Lines of Code
AutoRAG
RAG AutoML Tool - Find optimal RAG pipeline for your own data.
ChatGee
카카오톡 AI 챗봇 빌더
ChatRTX
A developer reference project for creating Retrieval Augmented Generation (RAG) chatbots on Windows using TensorRT-LLM
FLAML
A fast library for AutoML and tuning. Join our Discord: https://discord.gg/Cppx2vSPVP.
fsi-samples
A collection of open-source GPU accelerated Python tools and examples for quantitative analyst tasks and leverages RAPIDS AI project, Numba, cuDF, and Dask.
langchain-kr
LangChain 공식 Document, Cookbook, 그 밖의 실용 예제를 바탕으로 작성한 한국어 튜토리얼입니다. 본 튜토리얼을 통해 LangChain을 더 쉽고 효과적으로 사용하는 방법을 배울 수 있습니다.
langserve_ollama
무료로 한국어🇰🇷 파인튜닝 모델 받아서 로컬 LLM 호스팅. LangServe, Ollama, streamlit + RAG
law-chatbot
법원판례 챗봇 서비스
Live2Diff
Live2Diff: A Pipeline that processes Live video streams by a uni-directional video Diffusion model.
LLaMA-Factory
Unify Efficient Fine-Tuning of 100+ LLMs
magic-animate
MagicAnimate: Temporally Consistent Human Image Animation using Diffusion Model
mediapipe
Cross-platform, customizable ML solutions for live and streaming media.
NeMo
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
nim-anywhere
Accelerate your Generative AI with NIMs.
ollama
Get up and running with Llama 2, Mistral, Gemma, and other large language models.
open-webui
ChatGPT-Style WebUI for LLMs (Formerly Ollama WebUI)
priompt
Prompt design using JSX.
Socratic
A simple Mathematical utility for Humans 🦉
SolarLLMZeroToAll
Full Stack SolarLLM Zero to All
STUDY_RAPIDS
NVIDIA 기술랩 - Data Engineering with RAPIDS code 정리 용 repository
TensorRT-Model-Optimizer
TensorRT Model Optimizer is a unified library of state-of-the-art model optimization techniques such as quantization and sparsity. It compresses deep learning models for downstream deployment frameworks like TensorRT-LLM or TensorRT to optimize inference speed on NVIDIA GPUs.
Titanic_RAPIDS
Titanic Machine Learning with RAPIDS
workbench-example-hybrid-rag
An NVIDIA AI Workbench example project for Retrieval Augmented Generation (RAG)
workbench-example-rapids-cudf
An NVIDIA AI Workbench example project for exploring the RAPIDS cuDF library
workbench-example-rapids-cuml
An NVIDIA AI Workbench example project for exploring the RAPIDS cuML library
workbench-llamafactory
This is an NVIDIA AI Workbench example project that demonstrates an end-to-end model development workflow using Llamafactory.