Weoshin's starred repositories
MidJourney-Styles-and-Keywords-Reference
A reference containing Styles and Keywords that you can use with MidJourney AI. There are also pages showing resolution comparison, image weights, and much more!
Video-LLaVA
Video-LLaVA: Learning United Visual Representation by Alignment Before Projection
Awesome-Multimodal-Large-Language-Models
:sparkles::sparkles:Latest Advances on Multimodal Large Language Models
LLMs_interview_notes
该仓库主要记录 大模型(LLMs) 算法工程师相关的面试题
Ask-Anything
[CVPR2024 Highlight][VideoChatGPT] ChatGPT with video understanding! And many more supported LMs such as miniGPT4, StableLM, and MOSS.
SAEval-Benchmark
SAEval: A benchmark for sentiment analysis to evaluate the model's performance on various subtasks.
speechbrain
A PyTorch-based Speech Toolkit
torchscale
Foundation Architecture for (M)LLMs
ch-sims-v2
Make Acoustic and Visual Cues Matter: CH-SIMS v2.0 Dataset and AV-Mixup Consistent Module
MultiBench
[NeurIPS 2021] Multiscale Benchmarks for Multimodal Representation Learning
DialogueCRN
Source code for ACL-IJCNLP 2021 paper "DialogueCRN: Contextual Reasoning Networks for Emotion Recognition in Conversations"
Time-Series-Library
A Library for Advanced Deep Time Series Models.
NAACL-19-CIM
Multi-task Learning for Multi-modal Emotion Recognition and Sentiment Analysis
Multimodal-datasets
This repository is build in association with our position paper on "Multimodality for NLP-Centered Applications: Resources, Advances and Frontiers". As a part of this release we share the information about recent multimodal datasets which are available for research purposes. We found that although 100+ multimodal language resources are available in literature for various NLP tasks, still publicly available multimodal datasets are under-explored for its re-usage in subsequent problem domains.