There are 18 repositories under stt topic.
Your AI second brain. Self-hostable. Get answers from the web or your docs. Build custom agents, schedule automations, do deep research. Turn any online or local LLM into your personal, autonomous AI (gpt, claude, gemini, llama, qwen, mistral). Get started - free.
Offline speech recognition API for Android, iOS, Raspberry Pi and servers with Python, Java, C# and Node
Silero Models: pre-trained speech-to-text, text-to-speech and text-enhancement models made embarrassingly simple
Voice Recognition to Text Tool / 一个离线运行的本地音视频转字幕工具,输出json、srt字幕、纯文字格式
Transcribe any audio to text, translate and edit subtitles 100% locally with a web UI. Powered by whisper models!
🐸STT - The deep learning toolkit for Speech-to-Text. Training and deploying STT models has never been so easy.
🎙Speech recognition using the tensorflow deep learning framework, sequence-to-sequence neural networks
💎 A list of accessible speech corpora for ASR, TTS, and other Speech Technologies
🍦 Speech-AI-Forge is a project developed around TTS generation model, implementing an API Server and a Gradio-based WebUI.
Gp.nvim (GPT prompt) Neovim AI plugin: ChatGPT sessions & Instructable text/code operations & Speech to text [OpenAI, Ollama, Anthropic, ..]
Synchronized Translation for Videos. Video dubbing
Meet Ava, the WhatsApp Agent
Speech Note Linux app. Note taking, reading and translating with offline Speech to Text, Text to Speech and Machine translation.
Speech to Text to Speech. Song now playing. Sends text as OSC messages to VRChat to display on avatar. (STTTS) (Speech to TTS) (VRC STT System) (VTuber TTS)
:speech_balloon: /so.nus/ STT (speech to text) for Node with offline hotword detection
A React component to make correcting automated transcriptions of audio and video easier and faster. By BBC News Labs. - Work in progress
MLX Omni Server is a local inference server powered by Apple's MLX framework, specifically designed for Apple Silicon (M-series) chips. It implements OpenAI-compatible API endpoints, enabling seamless integration with existing OpenAI SDK clients while leveraging the power of local ML inference.
Running speech to text model (whisper.cpp) in Unity3d on your local machine.
A conversational, AI device + software framework for companionship, entertainment, education, healthcare, IoT applications, and DIY robotics. Built with Python, NextJS, Arduino, ESP32, LLMs (GPT-4o), Deepgram STT and Azure TTS 🤖
A speech recognition library running in the browser thanks to a WebAssembly build of Vosk
A collection of resources to make a smart speaker
Fast text based video editing, node Electron Os X desktop app, with Backbone front end.
A Conversational Assistant equipped with synthetic voices including J.A.R.V.I.S's. Powered by OpenAI and IBM Watson APIs and a Tacotron model for voice generation.
Deepgram Conversational AI demo
实时STT,连接OpenAI接口/智谱AI(流式LLM)和GPT-SOVITS/Edge-TTS,通过网页的方式,进行跨网络的服务调用,实现实时对话的效果
Striving to create a great Application with full functions of learning languages by ChatGPT, TTS, STT and other awesome AI models, supports talking, speaking assessment, memorizing words with contexts, Listening test, so on.
Open source speech to text models for Indic Languages
Speech-to-text in Obsidian using OpenAI Whisper
🗣 An overlay that gets your user’s voice permission and input as text in a customizable UI
Talk to ChatGPT in real time using LiveKit
Get started using Deepgram's Live Transcription with this Next.js demo app
Live-Transcription (STT) with Whisper PoC