唐月明's starred repositories

Raspberry-PI-Smart-Home

利用python语言进行编程设计,结合树莓派和小型电动机等硬件,通过软件的设计与硬件的相互协调,保证系统的正常运行。本文主要研究如何利用树莓派进行智能家居控制的程序设计,利用Python语言进行程序编写和其他的几个硬件设备完成自动窗帘的控制和温湿度的自动调控,并在自动模式之外可以进行手动的控制。进行本次研究,可以探寻树莓派在智能家居控制系统的设计方面的应用;同时通过与其他硬件搭配进行的设计,对探索不需要大幅度更改原有家居设备的情况下进行智能家居控制的方法有着重要意义。

Language:C++Stargazers:5Issues:0Issues:0

acl-style-files

Official style files for papers submitted to venues of the Association for Computational Linguistics

Language:TeXStargazers:628Issues:0Issues:0

awesome-llm-understanding-mechanism

awesome papers in LLM interpretability

Stargazers:187Issues:0Issues:0
Language:PythonStargazers:28Issues:0Issues:0

MAIA-Data-Processing

This repository has all the multimodal data processing scripts

Language:PythonStargazers:4Issues:0Issues:0

fairseq

Facebook AI Research Sequence-to-Sequence Toolkit written in Python.

Language:PythonLicense:MITStargazers:29660Issues:0Issues:0

Neural-Network-Parameter-Diffusion

We introduce a novel approach for parameter generation, named neural network parameter diffusion (p-diff), which employs a standard latent diffusion model to synthesize a new set of parameters

Language:PythonStargazers:781Issues:0Issues:0

MISA

MISA: Modality-Invariant and -Specific Representations for Multimodal Sentiment Analysis

Language:PythonLicense:MITStargazers:172Issues:0Issues:0

MMSA-FET

A Tool for extracting multimodal features from videos.

Language:PythonLicense:GPL-3.0Stargazers:121Issues:0Issues:0

MMSA

MMSA is a unified framework for Multimodal Sentiment Analysis.

Language:PythonLicense:MITStargazers:605Issues:0Issues:0

Vary

Official code implementation of Vary: Scaling Up the Vision Vocabulary of Large Vision Language Models.

Language:PythonStargazers:1628Issues:0Issues:0

mPLUG-DocOwl

mPLUG-DocOwl: Modularized Multimodal Large Language Model for Document Understanding

Language:PythonLicense:Apache-2.0Stargazers:1079Issues:0Issues:0

UniSA

UniSA: Unified Generative Framework for Sentiment Analysis

Language:PythonLicense:MITStargazers:41Issues:0Issues:0

how-to-train-tokenizer

怎么训练一个LLM分词器

Language:PythonStargazers:108Issues:0Issues:0

clip_text_span

official implementation of "Interpreting CLIP's Image Representation via Text-Based Decomposition"

Language:Jupyter NotebookLicense:MITStargazers:122Issues:0Issues:0

ALMT

Learning Language-guided Adaptive Hyper-modality Representation for Multimodal Sentiment Analysis

Language:PythonLicense:MITStargazers:50Issues:0Issues:0
Language:PythonLicense:MITStargazers:35Issues:0Issues:0

Osprey

[CVPR2024] The code for "Osprey: Pixel Understanding with Visual Instruction Tuning"

Language:PythonLicense:Apache-2.0Stargazers:711Issues:0Issues:0

img2dataset

Easily turn large sets of image urls to an image dataset. Can download, resize and package 100M urls in 20h on one machine.

Language:PythonLicense:MITStargazers:3403Issues:0Issues:0
Language:PythonLicense:NOASSERTIONStargazers:8165Issues:0Issues:0

CLIP

CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image

Language:Jupyter NotebookLicense:MITStargazers:23252Issues:0Issues:0

GroundingGPT

[ACL 2024] GroundingGPT: Language-Enhanced Multi-modal Grounding Model

Language:PythonLicense:Apache-2.0Stargazers:257Issues:0Issues:0
Language:PythonStargazers:110Issues:0Issues:0

Awesome-Multimodal-Large-Language-Models

:sparkles::sparkles:Latest Advances on Multimodal Large Language Models

Stargazers:10179Issues:0Issues:0

Awesome-LLM-KG

Awesome papers about unifying LLMs and KGs

Stargazers:1715Issues:0Issues:0

awesome-multimodal-ml

Reading list for research topics in multimodal machine learning

License:MITStargazers:5615Issues:0Issues:0

top-class-resource-library

SCUT拔尖基地班线上资源库

Stargazers:2Issues:0Issues:0

Tech_Aarticle

主要是我是日常看过的不错的文章的资源汇总,方便自己也分享给大家。有些我看过的,就会做简单的解读,没看过的,就先罗列一下,然后之后看了把解读更新上;涉及到搜索/推荐/自然语言处理。

Stargazers:1722Issues:0Issues:0