Yura Choi (Yuuraa)

Yuuraa

Geek Repo

Company:Yonsei Univeristy

Home Page:https://velog.io/@yoorachoi

Twitter:@Yura02786865

Github PK Tool:Github PK Tool

Yura Choi's starred repositories

llm-course

Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:37445Issues:389Issues:67

FastChat

An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.

Language:PythonLicense:Apache-2.0Stargazers:36530Issues:348Issues:1769

mem0

The Memory layer for your AI apps

Language:PythonLicense:Apache-2.0Stargazers:21951Issues:125Issues:641

LLaVA

[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.

Language:PythonLicense:Apache-2.0Stargazers:19455Issues:160Issues:1489

peft

🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.

Language:PythonLicense:Apache-2.0Stargazers:15947Issues:106Issues:1031

MiniCPM-V

MiniCPM-V 2.6: A GPT-4V Level MLLM for Single Image, Multi Image and Video on Your Phone

Language:PythonLicense:Apache-2.0Stargazers:12046Issues:101Issues:525

Awesome-Multimodal-Large-Language-Models

:sparkles::sparkles:Latest Advances on Multimodal Large Language Models

segment-anything-2

The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:10998Issues:64Issues:245

LAVIS

LAVIS - A One-stop Library for Language-Vision Intelligence

Language:Jupyter NotebookLicense:BSD-3-ClauseStargazers:9692Issues:97Issues:649

RAG_Techniques

This repository showcases various advanced techniques for Retrieval-Augmented Generation (RAG) systems. RAG systems combine information retrieval with generative models to provide accurate and contextually rich responses.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:6657Issues:74Issues:13

Voyager

An Open-Ended Embodied Agent with Large Language Models

Language:JavaScriptLicense:MITStargazers:5503Issues:63Issues:146

anthropic-cookbook

A collection of notebooks/recipes showcasing some fun and effective ways of using Claude.

Language:Jupyter NotebookLicense:MITStargazers:5113Issues:150Issues:31

agentscope

Start building LLM-empowered multi-agent applications in an easier way.

Language:PythonLicense:Apache-2.0Stargazers:4889Issues:30Issues:132

MeloTTS

High-quality multi-lingual text-to-speech library by MyShell.ai. Support English, Spanish, French, Chinese, Japanese and Korean.

Language:PythonLicense:MITStargazers:4445Issues:39Issues:163

LMOps

General technology for enabling AI capabilities w/ LLMs and MLLMs

Language:PythonLicense:MITStargazers:3581Issues:55Issues:116

Ask-Anything

[CVPR2024 Highlight][VideoChatGPT] ChatGPT with video understanding! And many more supported LMs such as miniGPT4, StableLM, and MOSS.

Language:PythonLicense:MITStargazers:2994Issues:36Issues:226

RouteLLM

A framework for serving and evaluating LLM routers - save LLM costs without compromising quality!

Language:PythonLicense:Apache-2.0Stargazers:2925Issues:25Issues:46

Awesome-LLMs-for-Video-Understanding

🔥🔥🔥Latest Papers, Codes and Datasets on Vid-LLMs.

MetaCLIP

ICLR2024 Spotlight: curation/training code, metadata, distribution and pre-trained models for MetaCLIP; CVPR 2024: MoDE: CLIP Data Experts via Clustering

Language:PythonLicense:NOASSERTIONStargazers:1176Issues:12Issues:27

Video-ChatGPT

[ACL 2024 🔥] Video-ChatGPT is a video conversation model capable of generating meaningful conversation about videos. It combines the capabilities of LLMs with a pretrained visual encoder adapted for spatiotemporal video representation. We also introduce a rigorous 'Quantitative Evaluation Benchmarking' for video-based conversational models.

Language:PythonLicense:CC-BY-4.0Stargazers:1163Issues:14Issues:119

awesome-instruction-dataset

A collection of open-source dataset to train instruction-following LLMs (ChatGPT,LLaMA,Alpaca)

Everything-LLMs-And-Robotics

The world's largest GitHub Repository for LLMs + Robotics

License:BSD-3-ClauseStargazers:752Issues:21Issues:0

ALLaVA

Harnessing 1.4M GPT4V-synthesized Data for A Lite Vision-Language Model

Language:PythonLicense:Apache-2.0Stargazers:239Issues:11Issues:11

BIKE

【CVPR'2023】Bidirectional Cross-Modal Knowledge Exploration for Video Recognition with Pre-trained Vision-Language Models

Language:PythonLicense:MITStargazers:157Issues:11Issues:23
Language:PythonLicense:MITStargazers:101Issues:4Issues:11

vlm-rlaif

ACL'24 (Oral) Tuning Large Multimodal Models for Videos using Reinforcement Learning from AI Feedback

Language:PythonLicense:Apache-2.0Stargazers:43Issues:3Issues:3

ECCV2024_MBS

[ECCV2024] Mitigating Background Shift in Class-Incremental Semantic Segmentation

Language:PythonLicense:MITStargazers:23Issues:2Issues:1

Pseudo-RIS

[ECCV 2024] Official code for "Pseudo-RIS: Distinctive Pseudo-supervision Generation for Referring Image Segmentation"

srt

i-SRT:Aligning Large Multimodal Models for Videos by Iterative Self-Retrospective Judgement

Language:PythonStargazers:10Issues:1Issues:0