Minjoon Jung (minjoong507)

minjoong507

Geek Repo

Company:Seoul National University

Location:Seoul, South Korea

Home Page:https://minjoong507.github.io/

Twitter:@minjoon507

Github PK Tool:Github PK Tool

Minjoon Jung's starred repositories

SelecMix

SelecMix: Debiased Learning by Contradicting-pair Sampling (NeurIPS 2022)

Language:PythonLicense:MITStargazers:11Issues:0Issues:0

Fine-Grained-Causal-RL

Fine-Grained Causal Dynamics Learning with Quantization for Improving Robustness in Reinforcement Learning (ICML 2024)

Language:PythonLicense:MITStargazers:7Issues:0Issues:0

FreeVA

FreeVA: Offline MLLM as Training-Free Video Assistant

Language:PythonLicense:Apache-2.0Stargazers:36Issues:0Issues:0

Chain-of-ThoughtsPapers

A trend starts from "Chain of Thought Prompting Elicits Reasoning in Large Language Models".

Stargazers:1870Issues:0Issues:0

VideoMamba

[ECCV2024] VideoMamba: State Space Model for Efficient Video Understanding

Language:PythonLicense:Apache-2.0Stargazers:680Issues:0Issues:0

Ask-Anything

[CVPR2024 Highlight][VideoChatGPT] ChatGPT with video understanding! And many more supported LMs such as miniGPT4, StableLM, and MOSS.

Language:PythonLicense:MITStargazers:2861Issues:0Issues:0

llama

Inference code for Llama models

Language:PythonLicense:NOASSERTIONStargazers:54160Issues:0Issues:0

GenAI_LLM_timeline

ChatGPT, GenerativeAI and LLMs Timeline

Stargazers:937Issues:0Issues:0

VTimeLLM

[CVPR'2024 Highlight] Official PyTorch implementation of the paper "VTimeLLM: Empower LLM to Grasp Video Moments".

Language:PythonLicense:NOASSERTIONStargazers:168Issues:0Issues:0

esper

ESPER

Language:PythonStargazers:23Issues:0Issues:0

viper

Code for the paper "ViperGPT: Visual Inference via Python Execution for Reasoning"

Language:Jupyter NotebookLicense:NOASSERTIONStargazers:1635Issues:0Issues:0

VTG-GPT

VTG-GPT: Tuning-Free Zero-Shot Video Temporal Grounding with GPT

Language:PythonLicense:MITStargazers:68Issues:0Issues:0

LaViLa

Code release for "Learning Video Representations from Large Language Models"

Language:PythonLicense:MITStargazers:458Issues:0Issues:0

TimeChat

[CVPR 2024] TimeChat: A Time-sensitive Multimodal Large Language Model for Long Video Understanding

Language:PythonLicense:BSD-3-ClauseStargazers:238Issues:0Issues:0
Language:PythonStargazers:34Issues:0Issues:0
Language:PythonLicense:MITStargazers:50Issues:0Issues:0

MPGN

[EMNLP 2022] Pytorch code for "Modal-specific Pseudo Query Generation for Video Corpus Moment Retrieval"

Language:PythonLicense:MITStargazers:6Issues:0Issues:0

LLaVA

[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.

Language:PythonLicense:Apache-2.0Stargazers:17946Issues:0Issues:0

NExT-GQA

Can I Trust Your Answer? Visually Grounded Video Question Answering (CVPR'24, Highlight)

Language:PythonLicense:MITStargazers:37Issues:0Issues:0

Awesome-LLMs-for-Video-Understanding

🔥🔥🔥Latest Papers, Codes and Datasets on Vid-LLMs.

Stargazers:921Issues:0Issues:0

actionformer_release

Code release for ActionFormer (ECCV 2022)

Language:PythonLicense:MITStargazers:405Issues:0Issues:0

vision-language-models-are-bows

Experiments and data for the paper "When and why vision-language models behave like bags-of-words, and what to do about it?" Oral @ ICLR 2023

Language:PythonLicense:MITStargazers:217Issues:0Issues:0

CGDETR

Official pytorch repository for CG-DETR "Correlation-guided Query-Dependency Calibration in Video Representation Learning for Temporal Grounding"

Language:PythonLicense:NOASSERTIONStargazers:92Issues:0Issues:0

LLaMA-Adapter

[ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters

Language:PythonLicense:GPL-3.0Stargazers:5614Issues:0Issues:0

BM-DETR

[arXiv 23] Pytorch code for "Overcoming Weak Visual-Textual Alignment for Video Moment Retrieval"

Language:PythonLicense:MITStargazers:11Issues:0Issues:0

GVCCI

[IROS 2023] GVCCI: Lifelong Learning of Visual Grounding for Language-Guided Robotic Manipulation

Language:PythonStargazers:7Issues:0Issues:0

sglkt-visdial

🌈 PyTorch Implementation for EMNLP'21 Findings "Reasoning Visual Dialog with Sparse Graph Learning and Knowledge Transfer"

Language:PythonLicense:MITStargazers:13Issues:0Issues:0

dan-visdial

✨ Official PyTorch Implementation for EMNLP'19 Paper, "Dual Attention Networks for Visual Reference Resolution in Visual Dialog"

Language:PythonLicense:MITStargazers:45Issues:0Issues:0

gst-visdial

:speech_balloon: Official PyTorch Implementation for CVPR'23 Paper, "The Dialog Must Go On: Improving Visual Dialog via Generative Self-Training"

Language:PythonLicense:MITStargazers:18Issues:0Issues:0

caption_contest_corpus

Corpus to accompany: "Do Androids Laugh at Electric Sheep? Humor "Understanding" Benchmarks from The New Yorker Caption Contest"

Language:PythonLicense:MITStargazers:49Issues:0Issues:0