Hanoona Rasheed (hanoonaR)

hanoonaR

Geek Repo

Company:MBZUAI

Location:Dubai, UAE

Home Page:https://www.hanoonarasheed.com/

Github PK Tool:Github PK Tool


Organizations
mbzuai-oryx

Hanoona Rasheed's starred repositories

baple

[MICCAI 2024] Official code repository of paper titled "BAPLe: Backdoor Attacks on Medical Foundation Models using Prompt Learning" accepted in MICCAI 2024 conference.

Language:PythonLicense:MITStargazers:43Issues:0Issues:0

segment-anything-2

The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:9976Issues:0Issues:0

GroupMamba

Official implementation of paper titled "GroupMamba: Parameter-Efficient and Accurate Group Visual State Space Model"

Language:PythonLicense:MITStargazers:54Issues:0Issues:0

VideoGPT-plus

Official Repository of paper VideoGPT+: Integrating Image and Video Encoders for Enhanced Video Understanding

Language:PythonLicense:CC-BY-4.0Stargazers:177Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:2118Issues:0Issues:0

corenet

CoreNet: A library for training deep neural networks

Language:PythonLicense:NOASSERTIONStargazers:6898Issues:0Issues:0

MobiLlama

MobiLlama : Small Language Model tailored for edge devices

Language:PythonLicense:Apache-2.0Stargazers:577Issues:0Issues:0

LLaVA-pp

🔥🔥 LLaVA++: Extending LLaVA with Phi-3 and LLaMA-3 (LLaVA LLaMA-3, LLaVA Phi-3)

Language:PythonStargazers:783Issues:0Issues:0

MAVOS

Efficient Video Object Segmentation via Modulated Cross-Attention Memory

License:BSD-3-ClauseStargazers:45Issues:0Issues:0

ovsam

[ECCV 2024] The official code of paper "Open-Vocabulary SAM".

Language:PythonLicense:NOASSERTIONStargazers:887Issues:0Issues:0

Video-LLaVA

PG-Video-LLaVA: Pixel Grounding in Large Multimodal Video Models

Language:PythonStargazers:233Issues:0Issues:0

groundingLMM

[CVPR 2024 🔥] Grounding Large Multimodal Model (GLaMM), the first-of-its-kind model capable of generating natural language responses that are seamlessly integrated with object segmentation masks.

Language:PythonStargazers:730Issues:0Issues:0

XM-GAN

[MICCAI 2023][Early Accept] Official code repository of paper titled "Cross-modulated Few-shot Image Generation for Colorectal Tissue Classification"

Language:PythonStargazers:44Issues:0Issues:0

GoogleBard-VisUnderstand

How Good is Google Bard's Visual Understanding? An Empirical Study on Open Challenges

Stargazers:30Issues:0Issues:0

vafa

[MICCAI 2023] Official code repository of paper titled "Frequency Domain Adversarial Training for Robust Volumetric Medical Segmentation" accepted in MICCAI 2023 conference.

Language:PythonLicense:MITStargazers:47Issues:0Issues:0

XrayGPT

[BIONLP@ACL 2024] XrayGPT: Chest Radiographs Summarization using Medical Vision-Language Models.

Language:PythonStargazers:455Issues:0Issues:0

ClimateGPT

[EMNLP'23] ClimateGPT: a specialized LLM for conversations related to Climate Change and Sustainability topics in both English and Arabic languages.

Language:PythonStargazers:73Issues:0Issues:0

Video-ChatGPT

[ACL 2024 🔥] Video-ChatGPT is a video conversation model capable of generating meaningful conversation about videos. It combines the capabilities of LLMs with a pretrained visual encoder adapted for spatiotemporal video representation. We also introduce a rigorous 'Quantitative Evaluation Benchmarking' for video-based conversational models.

Language:PythonLicense:CC-BY-4.0Stargazers:1130Issues:0Issues:0

XPretrain

Multi-modality pre-training

Language:PythonLicense:NOASSERTIONStargazers:466Issues:0Issues:0

SwiftFormer

[ICCV'23] Official repository of paper SwiftFormer: Efficient Additive Attention for Transformer-based Real-time Mobile Vision Applications

Language:PythonStargazers:239Issues:0Issues:0

Transformer-MM-Explainability

[ICCV 2021- Oral] Official PyTorch implementation for Generic Attention-model Explainability for Interpreting Bi-Modal and Encoder-Decoder Transformers, a novel method to visualize any Transformer-based network. Including examples for DETR, VQA.

Language:Jupyter NotebookLicense:MITStargazers:765Issues:0Issues:0

pointnet2

PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space

Language:PythonLicense:NOASSERTIONStargazers:3048Issues:0Issues:0

LanguageGroundedSemseg

Implementation for ECCV 2022 paper Language-Grounded Indoor 3D Semantic Segmentation in the Wild

Language:PythonStargazers:97Issues:0Issues:0

DenseCLIP

[CVPR 2022] DenseCLIP: Language-Guided Dense Prediction with Context-Aware Prompting

Language:PythonStargazers:505Issues:0Issues:0

CoOp

Prompt Learning for Vision-Language Models (IJCV'22, CVPR'22)

Language:PythonLicense:MITStargazers:1651Issues:0Issues:0

object-centric-ovd

[NeurIPS 2022] Official repository of paper titled "Bridging the Gap between Object and Image-level Representations for Open-Vocabulary Detection".

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:285Issues:0Issues:0

PaLM-rlhf-pytorch

Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the PaLM architecture. Basically ChatGPT but with PaLM

Language:PythonLicense:MITStargazers:7667Issues:0Issues:0

awesome-chatgpt-prompts

This repo includes ChatGPT prompt curation to use ChatGPT better.

Language:HTMLLicense:CC0-1.0Stargazers:108430Issues:0Issues:0