Lily (Xianling) Zhang's repositories
unpuzzle_dataset_bias
unpuzzle_dataset_bias
stablediffusion
High-Resolution Image Synthesis with Latent Diffusion Models
al-folio
A beautiful, simple, clean, and responsive Jekyll theme for academics
BEVFormer
[ECCV 2022] This is the official implementation of BEVFormer, a camera-only framework for autonomous driving perception, e.g., 3D object detection and semantic map segmentation.
bevfusion
BEVFusion: Multi-Task Multi-Sensor Fusion with Unified Bird's-Eye View Representation
BEVPerception-Survey-Recipe
Awesome BEV perception papers and cookbook for achieving SOTA results
deep-learning
Repo for the Deep Learning Nanodegree Foundations program.
deepdrive-sim
Unreal project and plugin, assets, and Python extension
llama
Inference code for LLaMA models
Multimodal-GPT
Multimodal-GPT
neurad-studio
[CVPR2024] NeuRAD: Neural Rendering for Autonomous Driving
neural-dependency-parser-pytorch
build a neural dependency parser using PyTorch
nuscenes-instance-videos
Create videos from NuScene data of instances across multiple frames
open_flamingo
An open-source framework for training large multimodal models.
Research-Paper-Study-Domain-Adaptation-with-GAN
Show the status of the research paper study
Segment-and-Track-Anything
An open-source project dedicated to tracking and segmenting any objects in videos, either automatically or interactively. The primary algorithms utilized include the Segment Anything Model (SAM) for key-frame segmentation and Associating Objects with Transformers (AOT) for efficient tracking and propagation purposes.
SRP-DNN
A python implementation of “SRP-DNN: Learning Direct-Path Phase Difference for Multiple Moving Sound Source Localization” [ICASSP 2022]
Track-Anything
Track-Anything is a flexible and interactive tool for video object tracking and segmentation, based on Segment Anything, XMem, and E2FGVI.
WebVR_Playground
For experimental demo
YOLOP
You Only Look Once for Panopitic Driving Perception.(MIR2022)