tdye24

tdye24

Geek Repo

Company:ECNU

Location:Shanghai

Home Page:http://tdye24.github.io

Github PK Tool:Github PK Tool

tdye24's starred repositories

llm-inference-benchmark

LLM Inference benchmark

Language:PythonLicense:MITStargazers:318Issues:0Issues:0

LLMs_interview_notes

LLMs interview notes and answers:该仓库主要记录大模型(LLMs)算法工程师相关的面试题和参考答案

License:MITStargazers:222Issues:0Issues:0

Personalized_PCA

An implementation for personalized PCA

Language:PythonStargazers:2Issues:0Issues:0

llama3-from-scratch

llama3 implementation one matrix multiplication at a time

Language:Jupyter NotebookLicense:MITStargazers:12754Issues:0Issues:0

llama3

The official Meta Llama 3 GitHub site

Language:PythonLicense:NOASSERTIONStargazers:25723Issues:0Issues:0

Awesome-LLM-Compression

Awesome LLM compression research papers and tools.

License:MITStargazers:986Issues:0Issues:0

dap-cl

Official code of "Generating Instance-level Prompts for Rehearsal-free Continual Learning (ICCV 2023)"

Language:PythonLicense:NOASSERTIONStargazers:39Issues:0Issues:0

LLM-Pruner

[NeurIPS 2023] LLM-Pruner: On the Structural Pruning of Large Language Models. Support Llama-3/3.1, Llama-2, LLaMA, BLOOM, Vicuna, Baichuan, TinyLlama, etc.

Language:PythonLicense:Apache-2.0Stargazers:780Issues:0Issues:0

llm-action

本项目旨在分享大模型相关技术原理以及实战经验。

Language:HTMLLicense:Apache-2.0Stargazers:8710Issues:0Issues:0

Monkey

【CVPR 2024 Highlight】Monkey (LMM): Image Resolution and Text Label Are Important Things for Large Multi-modal Models

Language:PythonLicense:MITStargazers:1668Issues:0Issues:0
Language:PythonStargazers:7Issues:0Issues:0

idl_data

OCR Annotations from Amazon Textract for Industry Documents Library

Language:PythonStargazers:93Issues:0Issues:0

MM-NIAH

This is the official implementation of the paper "Needle In A Multimodal Haystack"

Language:PythonStargazers:72Issues:0Issues:0

GAOKAO-Bench

GAOKAO-Bench is an evaluation framework that utilizes GAOKAO questions as a dataset to evaluate large language models.

Language:PythonLicense:Apache-2.0Stargazers:505Issues:0Issues:0

CogVLM2

GPT4V-level open-source multi-modal model based on Llama3-8B

Language:PythonLicense:Apache-2.0Stargazers:1816Issues:0Issues:0

CHIP2022_MedTable-MedInvoice_CogVLM

阿里天池算法竞赛-CHIP2022医疗清单发票OCR要素提取任务-solution

Language:PythonStargazers:3Issues:0Issues:0

DuReader

Baseline Systems of DuReader Dataset

Language:PythonStargazers:1121Issues:0Issues:0

ChiQA

The implementations of various baselines in our CIKM 2022 paper: ChiQA: A Large Scale Image-based Real-World Question Answering Dataset for Multi-Modal Understanding.

Language:PythonStargazers:30Issues:0Issues:0

Efficient-Multimodal-LLMs-Survey

Efficient Multimodal Large Language Models: A Survey

License:Apache-2.0Stargazers:219Issues:0Issues:0

TaiSu

TaiSu(太素)--a large-scale Chinese multimodal dataset(亿级大规模中文视觉语言预训练数据集)

Language:PythonLicense:NOASSERTIONStargazers:171Issues:0Issues:0

CLUEDatasetSearch

搜索所有中文NLP数据集,附常用英文NLP数据集

Language:PythonStargazers:4064Issues:0Issues:0

screen_qa

ScreenQA dataset was introduced in the "ScreenQA: Large-Scale Question-Answer Pairs over Mobile App Screenshots" paper. It contains ~86K question-answer pairs collected by human annotators for ~35K screenshots from Rico. It should be used to train and evaluate models capable of screen content understanding via question answering.

License:CC-BY-4.0Stargazers:82Issues:0Issues:0
Language:PythonLicense:MITStargazers:51Issues:0Issues:0
Language:PythonLicense:GPL-3.0Stargazers:153Issues:0Issues:0

DeepSpeed

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

Language:PythonLicense:Apache-2.0Stargazers:34468Issues:0Issues:0

Awesome-Multimodal-Large-Language-Models

:sparkles::sparkles:Latest Advances on Multimodal Large Language Models

Stargazers:11325Issues:0Issues:0

InternVL

[CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4o. 接近GPT-4o表现的开源多模态对话模型

Language:PythonLicense:MITStargazers:5204Issues:0Issues:0

FedPR

【CVPR 2023】Learning Federated Visual Prompt in Null Space for MRI Reconstruction

Language:PythonStargazers:42Issues:0Issues:0

VLMEvalKit

Open-source evaluation toolkit of large vision-language models (LVLMs), support ~100 VLMs, 30+ benchmarks

Language:PythonLicense:Apache-2.0Stargazers:904Issues:0Issues:0

invoice

Collaboration with wangxupeng(https://github.com/wangxupeng)

Language:CLicense:MITStargazers:1787Issues:0Issues:0