Srijan Das (srijandas07)

srijandas07

Geek Repo

Company:UNC Charlotte

Location:Charlotte, USA

Home Page:https://srijandas07.github.io/

Twitter:@srijandas07

Github PK Tool:Github PK Tool

Srijan Das's starred repositories

Awesome-Transformer-Attention

An ultimately comprehensive paper list of Vision Transformer/Attention, including papers, codes, and related websites

pytorchvideo

A deep learning library for video understanding research.

Language:PythonLicense:Apache-2.0Stargazers:3250Issues:160Issues:180

ffcv

FFCV: Fast Forward Computer Vision (and other ML workloads!)

Language:PythonLicense:Apache-2.0Stargazers:2812Issues:20Issues:276

TimeSformer

The official pytorch implementation of our paper "Is Space-Time Attention All You Need for Video Understanding?"

Language:PythonLicense:NOASSERTIONStargazers:1496Issues:28Issues:128

VideoMAE

[NeurIPS 2022 Spotlight] VideoMAE: Masked Autoencoders are Data-Efficient Learners for Self-Supervised Video Pre-Training

Language:PythonLicense:NOASSERTIONStargazers:1296Issues:16Issues:119

ICCV-2023-Papers

ICCV 2023 Papers: Discover cutting-edge research from ICCV 2023, the leading computer vision conference. Stay updated on the latest in computer vision and deep learning, with code included. ⭐ support visual intelligence development!

Language:PythonLicense:MITStargazers:908Issues:13Issues:10

MultiMAE

MultiMAE: Multi-modal Multi-task Masked Autoencoders, ECCV 2022

Language:PythonLicense:NOASSERTIONStargazers:535Issues:13Issues:33

Ego4d

Ego4d dataset repository. Download the dataset, visualize, extract features & example usage of the dataset

Language:Jupyter NotebookLicense:MITStargazers:330Issues:23Issues:153

MAE

PyTorch implementation of Masked Autoencoder

Language:PythonLicense:MITStargazers:206Issues:2Issues:21

SPT_LSA_ViT

Implementation of Visual Transformer for Small-size Datasets

VidIL

Pytorch code for Language Models with Image Descriptors are Strong Few-Shot Video-Language Learners

Language:PythonLicense:MITStargazers:110Issues:5Issues:11

imix

ICLR 2021 i-Mix: A Domain-Agnostic Strategy for Contrastive Representation Learning

Language:PythonLicense:MITStargazers:77Issues:3Issues:8

LIV

Official repository for "LIV: Language-Image Representations and Rewards for Robotic Control" (ICML 2023)

Language:PythonLicense:MITStargazers:77Issues:3Issues:7

MS-TCT

[CVPR2022] MS-TCT

PathLDM

Official Code for PathLDM: Text conditioned Latent Diffusion Model for Histopathology (WACV 2024)

Language:Jupyter NotebookStargazers:26Issues:7Issues:24

Limited-data-vits

[WACV 2024] Code for "Limited Data, Unlimited Potential: A Study on ViTs Augmented by Masked Autoencoders"

PoseAwareVT

Code for the paper Seeing the Pose in the Pixels: Learning Pose-Aware Representations in Vision Transformers

SI-MIL

SI-MIL

Language:PythonStargazers:18Issues:0Issues:0

3DTRL

Code for NeurIPS 2022 paper "Learning Viewpoint-Agnostic Visual Representations by Recovering Tokens in 3D Space"

Language:PythonLicense:MITStargazers:18Issues:6Issues:1

2s-AGCN-For-Daily-Living

2s-AGCN on Smarthome (dataset for daily living)

Fibottention

Inceptive Visual Representation Learning with Diverse Attention Across Heads

Language:PythonLicense:CC-BY-4.0Stargazers:14Issues:0Issues:0

LLAVIDAL

This is the offical repository of LLAVIDAL

Language:PythonLicense:CC-BY-4.0Stargazers:12Issues:1Issues:2

pi-vit

[CVPR 2024] Code and models for pi-ViT, a video transformer for understanding activities of daily living

Language:PythonLicense:NOASSERTIONStargazers:10Issues:3Issues:0

Toyota_Smarthome

Tools for Toyota Smarthome datasets

mavrec-code

This code is provided for reproducibility of results in the paper: Multiview Aerial Visual Recognition (MAVREC): Can Multi-view Improve Aerial Visual Perception?

Language:Jupyter NotebookLicense:CC-BY-4.0Stargazers:7Issues:1Issues:0

FreqMixFormer

[ACM MM 2024] Frequency Guidance Matters: Skeletal Action Recognition by Frequency-Aware Mixed Transformer

Stargazers:6Issues:0Issues:0

improved_HAR_on_Toyota

Improved action recognition with Separable spatio-temporal attention using alternative Skeletal and Video pre-processing

Language:PythonStargazers:4Issues:3Issues:0

separable_STA

Implementation of Separable Spatio-temporal attention (STA) netowork

Language:PythonStargazers:2Issues:3Issues:0

synchronization-is-all-you-need

Synchronization is All You Need: Exocentric-to-Egocentric Transfer for Temporal Action Segmentation with Unlabeled Synchronized Video Pairs [ECCV, 2024]

Language:PythonLicense:MITStargazers:1Issues:4Issues:1

Pyvideoresearch_new

The master PyVideoresearch is committed with few changes in this repository in order to make use of the pre-trained models.

Language:PythonLicense:GPL-3.0Stargazers:1Issues:2Issues:1