UCLWilson / DSify

Boosting Driving Scene Understanding with Advanced Vision-Language Models

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

๐ŸšฆDriveScenify: Boosting Driving Scene Understanding with Advanced Vision-Language Models

Introduction ๐Ÿ“š

The increasing complexity of traffic situations, coupled with the rapid growth of urban populations, necessitates the development of innovative solutions that can mitigate congestion, reduce traffic-related accidents, and facilitate smoother transportation systems. Recognizing the significant impact of ChatGPT and computer vision technologies on various domains, it is timely to investigate how these advancements can be harnessed to address the critical challenges in urban transportation safety and efficiency.

With this motivation, we introduce DriveScenify, an approach that aims to boost driving scene understanding by leveraging advanced vision-language models. Our research focuses on developing a tailored version of MiniGPT-4, called DSify, which is specifically designed to process and generate contextually relevant responses based on driving scene videos. DriveScenify's integration of advanced vision-language models into the realm of transportation aims to unlock new possibilities for improving urban mobility, reducing traffic-related accidents, and enhancing the overall driving experience.

Furthermore, our unique combination among various encoders enables DSify to provide accurate and context-aware insights, which can be applied to various transportation applications, especially for traffic management, and road safety analysis.

image

Features ๐ŸŒŸ

  • Spatial-temporal Safe Driving Scene Comprehension: DriveScenify is meticulously developed to accurately interpret diverse driving scenarios, encompassing traffic patterns, vehicle classifications, road conditions and temporal information, with a particular emphasis on promoting driving safety.
  • Contextual Response Formulation: The model is capable of generating context-sensitive responses and recommendations derived from the driving scene, offering valuable guidance to users.
  • While our central focus lies in training DSify using driving scenario videos, the model also exhibits a degree of competence in understanding and processing general video content. This versatility enhances its potential applications across a broader range of domains while maintaining its primary objective of improving driving safety and scene understanding.

Example ๐Ÿ’ฌ

demo

Usage ๐Ÿ’ป

DriveScenify was initially designed to comprehend corner cases and potentially hazardous situations within driving scenes. Our aim was to leverage the capabilities of Large Language Models (LLMs) to enhance the reasoning process for video understanding, providing a more comprehensive analysis of complex and challenging driving scenarios.

If you want to try the demo of this repo, you only need to refer to the installation process of MiniGPT-4, prepare the environment and Vicuna weights.

Then change the ckpt path in eval_configs/minigpt4_eval.yaml. You can download our weight here. Checkpoint Aligned with Vicuna 13B.

Launching Demo Locally

Try out our demo demo_video.py on your local machine by running

python demo_video.py --cfg-path eval_configs/minigpt4_eval.yaml

In fact, the demo supports both image and video inputs, so feel free to give it a try, even though the file is named "demo_video". Have fun exploring! ๐Ÿ˜„๐ŸŽ‰๐Ÿ“ท๐ŸŽฅ

Upcoming Tasks ๐Ÿค–

  • Strong video foundation model.
  • Training with dialogue datasets.
  • Expanding data generation capabilities.
  • ...

Contributors:

Xiaowei Gao* (SpacetimeLab, University College London, UK) email

Pengxiang Li* (IIAU-Lab, Dalian University of Technology, China) email

Ming Li (National Heart & Lung Institute, Imperial College London, UK) email

Xinke Jiang (Key Lab of High Confidence Software Technologies,Peking Universtiy, China) email

(*Equal Contribution)

Supervisors:

Dr. James Haworth (SpacetimeLab, University College London, UK)

Dr. Jonathan Cardoso-Silva (Data Science Institute, London School of Economics and Political Science, UK)

Contributing ๐Ÿค

At present, DriveScenify is in its initial stages, and in many cases, the performance may not be as ideal as expected. Data generation is still ongoing, and we are continuously working to improve the model. We highly appreciate and welcome contributions from the community to help enhance DriveScenify's capabilities and performance.

License ๐Ÿ“„

This repository is under BSD 3-Clause License. Many codes are based on Lavis with BSD 3-Clause License here.

Acknowledgments ๐Ÿค

We would like to thank the developers of MiniGPT-4, LLaVA, InternVideo, Ask-Anything, Image2Paragraph and Vicuna for their incredible work and providing the foundation for DriveScenify.

About

Boosting Driving Scene Understanding with Advanced Vision-Language Models

License:BSD 3-Clause "New" or "Revised" License


Languages

Language:Python 100.0%