Piyush Bagad (bpiyush)

bpiyush

Geek Repo

Company:University of Oxford

Location:Oxford

Home Page:bpiyush.github.io

Twitter:@bagad_piyush

Github PK Tool:Github PK Tool

Piyush Bagad's repositories

TestOfTime

Official code for our CVPR 2023 paper: Test of Time: Instilling Video-Language Models with a Sense of Time

Language:PythonLicense:MITStargazers:45Issues:3Issues:5

rotation-equivariant-lfm

Rotation equivariance meets local feature matching

Language:Jupyter NotebookLicense:MITStargazers:18Issues:2Issues:0

dino-local

PyTorch code for Vision Transformers training with the Self-Supervised learning method DINO

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:1Issues:0Issues:0

FastSAM

Fast Segment Anything

Language:PythonLicense:AGPL-3.0Stargazers:1Issues:0Issues:0

FCN-f0

Fully-Convolutional Network for Pitch Estimation of Speech Signals

Language:PythonLicense:MITStargazers:1Issues:0Issues:0

new-machine-setup-scripts

Bunch of scripts useful to add when starting on a new machine

Language:ShellStargazers:1Issues:3Issues:0

NLP-CS671A

Course files for CS671A - Natural Language Processing

Language:PythonStargazers:1Issues:2Issues:0

sam-pt

SAM-PT: Extending SAM to zero-shot video segmentation with point-based tracking.

Language:PythonLicense:Apache-2.0Stargazers:1Issues:0Issues:0

sound-guided-semantic-image-manipulation

Sound-guided Semantic Image Manipulation - Official Pytorch Code (CVPR 2022)

Language:PythonLicense:NOASSERTIONStargazers:1Issues:0Issues:0

Sound2Scene

Clone of the Sound2Scene repo. Need to train on pouring water images.

Language:PythonStargazers:1Issues:0Issues:0

TempCompass

[ACL 2024 Findings] "TempCompass: Do Video LLMs Really Understand Videos?", Yuanxin Liu, Shicheng Li, Yi Liu, Yuxiang Wang, Shuhuai Ren, Lei Li, Sishuo Chen, Xu Sun, Lu Hou

Stargazers:1Issues:0Issues:0

transparent-liquid-segmentation

We build a novel self-supervised segmentation pipeline to segment transparent liquids (clear water) placed inside transparent containers.

Language:Jupyter NotebookLicense:MITStargazers:1Issues:0Issues:0
Stargazers:1Issues:0Issues:0

audio_codec_tests

Tests for codec artefacts in stored audio samples.

Language:PythonLicense:MITStargazers:0Issues:1Issues:0

bpiyush

My personal introductory repository

Stargazers:0Issues:2Issues:0

bpiyush.github.io

A portfolio page

Language:JavaScriptLicense:MITStargazers:0Issues:2Issues:0

ddsp-pytorch

Implementation of DDSP (PyTorch), Differentiable Digital Signal Processing (ICLR 2020)

Language:Jupyter NotebookLicense:MITStargazers:0Issues:0Issues:0

digan

Official PyTorch implementation of Generating Videos with Dynamics-aware Implicit Generative Adversarial Networks (ICLR 2022).

Language:PythonStargazers:0Issues:0Issues:0

InternVideo

Video Foundation Models & Data for Multimodal Understanding

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

LanguageBind

【ICLR 2024🔥】 Extending Video-Language Pretraining to N-modality by Language-based Semantic Alignment

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

LAVIS

LAVIS - A One-stop Library for Language-Vision Intelligence

License:BSD-3-ClauseStargazers:0Issues:0Issues:0

PhysParamInference

Clone of the WACV2023 paper. Adaptation on pouring water.

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

TimeChat

[CVPR 2024] TimeChat: A Time-sensitive Multimodal Large Language Model for Long Video Understanding

Language:Jupyter NotebookLicense:BSD-3-ClauseStargazers:0Issues:0Issues:0

unmasked_teacher

[ICCV2023 Oral] Unmasked Teacher: Towards Training-Efficient Video Foundation Models

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

Video-LLaMA

[EMNLP 2023 Demo] Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding

Language:PythonLicense:BSD-3-ClauseStargazers:0Issues:0Issues:0

VideoLLaMA2

VideoLLaMA 2: Advancing Spatial-Temporal Modeling and Audio Understanding in Video-LLMs

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

VideoMAE-ssl

[NeurIPS 2022 Spotlight] VideoMAE: Masked Autoencoders are Data-Efficient Learners for Self-Supervised Video Pre-Training

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0

ViLMA

ViLMA: A Zero-Shot Benchmark for Linguistic and Temporal Grounding in Video-Language Models (ICLR 2024, Official Implementation)

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

VTimeLLM

[CVPR'2024 Highlight] Official PyTorch implementation of the paper "VTimeLLM: Empower LLM to Grasp Video Moments".

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0

YouTube-scrapper-tutorial

Tutorial to scrape YouTube video for research purposes.

Language:Jupyter NotebookLicense:MITStargazers:0Issues:2Issues:0