tfgbestneal's repositories

MobileVLM

Strong and Open Vision Language Assistant for Mobile Devices

Language:PythonLicense:Apache-2.0Stargazers:1Issues:0Issues:0

agentsflow

Drag & drop UI to build and run a flow of autogen AI agents

Language:TypeScriptLicense:GPL-3.0Stargazers:0Issues:0Issues:0

ComfyUI

The most powerful and modular stable diffusion GUI with a graph/nodes interface.

Language:PythonLicense:GPL-3.0Stargazers:0Issues:0Issues:0

DetectSegPlatform

YoloWorld & Flask

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

donkeycar

Open source hardware and software platform to build a small scale self driving car.

License:MITStargazers:0Issues:0Issues:0

DriveLM

DriveLM: Driving with Graph Visual Question Answering

License:Apache-2.0Stargazers:0Issues:0Issues:0

Focal_TSMP

Deep learning for vegetation health prediction and agricultural drought assessment from a regional climate simulation

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

Gemini

Google Gemini AI model w/speech recognition and voice.

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

gpt-assistant-android

免费的ChatGPT API的安卓语音助手,可用音量键唤起并进行语音交流,支持联网、Vision拍照识图、提问模板等功能 | A free ChatGPT API voice assistant for Android, activated via volume keys for voice interaction, supporting features such as network connectivity, Vision photo recognition, and question templates.

Language:JavaLicense:GPL-3.0Stargazers:0Issues:0Issues:0

gpt-pilot

Dev tool that writes scalable apps from scratch while the developer oversees the implementation

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

groundingLMM

Grounding Large Multimodal Model (GLaMM), the first-of-its-kind model capable of generating natural language responses that are seamlessly integrated with object segmentation masks.

Stargazers:0Issues:0Issues:0

InternLM-XComposer

InternLM-XComposer2 is a groundbreaking vision-language large model (VLLM) excelling in free-form text-image composition and comprehension.

Language:PythonStargazers:0Issues:0Issues:0

InternVL

[CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4V. 接近GPT-4V表现的可商用开源模型

License:MITStargazers:0Issues:0Issues:0

jepa

PyTorch code and models for V-JEPA self-supervised learning from video.

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0

MetaGPT

🌟 The Multi-Agent Framework: Given one line Requirement, return PRD, Design, Tasks, Repo

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

MiniGPT4Qwen

Cleaned Lavis + DeepSpeed Support! Align MiniGPT4 with Qwen-Chat LLM. I just use 18.8k high-quality instruction-tuning data(Bi-lingual, from minigpt4 and llava). Just fine-tune the projection layer.

Stargazers:0Issues:0Issues:0

Open-Sora

Open-Sora: Democratizing Efficient Video Production for All

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

PaddleSpeech

Easy-to-use Speech Toolkit including Self-Supervised Learning model, SOTA/Streaming ASR with punctuation, Streaming TTS with text frontend, Speaker Verification System, End-to-End Speech Translation and Keyword Spotting. Won NAACL2022 Best Demo Award.

License:Apache-2.0Stargazers:0Issues:0Issues:0

RoadVision

Revolutionizing navigation with AR and MapKit integration, this iOS app offers immersive, real-time directions and customizable UI for an intuitive experience. #iOSDevelopment #AugmentedReality #MapKit #SwiftUI #Innovation

License:MITStargazers:0Issues:0Issues:0

SAM_gDINO_AutoLabeling

Auto Segmentation label generation with SAM (Segment Anything) + Grounding DINO

License:Apache-2.0Stargazers:0Issues:0Issues:0
Language:TypeScriptLicense:MITStargazers:0Issues:0Issues:0

screenshot-to-code

Drop in a screenshot and convert it to clean HTML/Tailwind/JS code

Language:TypeScriptLicense:MITStargazers:0Issues:1Issues:0

Segment-and-Track-Anything

An open-source project dedicated to tracking and segmenting any objects in videos, either automatically or interactively. The primary algorithms utilized include the Segment Anything Model (SAM) for key-frame segmentation and Associating Objects with Transformers (AOT) for efficient tracking and propagation purposes.

Language:Jupyter NotebookLicense:AGPL-3.0Stargazers:0Issues:0Issues:0

Segment-Everything-Everywhere-All-At-Once

[NeurIPS 2023] Official implementation of the paper "Segment Everything Everywhere All at Once"

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

SegmentAnything3D

SAM3D: Segment Anything in 3D Scenes

License:MITStargazers:0Issues:0Issues:0

SpeechAgents

SpeechAgents: Human-Communication Simulation with Multi-Modal Multi-Agent Systems

Stargazers:0Issues:0Issues:0

tracking_ros

ROS compatible package for object tracking based on SAM, Cutie, GroundingDINO, YOLO-World, VLPart and DEVA

Stargazers:0Issues:0Issues:0

U-2-Net

The code for our newly accepted paper in Pattern Recognition 2020: "U^2-Net: Going Deeper with Nested U-Structure for Salient Object Detection."

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

X-AnyLabeling

Effortless data labeling with AI support from Segment Anything and other awesome models.

Language:PythonLicense:GPL-3.0Stargazers:0Issues:0Issues:0

YOLOV8_SAM

yolov8 model with SAM meta

Language:Jupyter NotebookLicense:MITStargazers:0Issues:0Issues:0