There are 7 repositories under speech-translation topic.
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
Easy-to-use Speech Toolkit including Self-Supervised Learning model, SOTA/Streaming ASR with punctuation, Streaming TTS with text frontend, Speaker Verification System, End-to-End Speech Translation and Keyword Spotting. Won NAACL2022 Best Demo Award.
Speech To Speech: an effort for an open-sourced and modular GPT4-o
StreamSpeech is an โAll in Oneโ seamless model for offline and simultaneous speech recognition, speech translation and speech synthesis.
Paper list of simultaneous translation / streaming translation, including text-to-text machine translation and speech-to-text translation.
A realtime speech transcription and translation application using Whisper OpenAI and free translation API. Interface made using Tkinter. Code written fully in Python.
The dataset of Speech Recognition
Cross-platform speech toolset, used from the command-line or as a Node.js library. Includes a variety of engines for speech synthesis, speech recognition, forced alignment, speech translation, voice isolation, language detection and more.
Tracking the progress in end-to-end speech translation
MooER: Moore-threads Open Omni model for speech-to-speech intERaction. MooER-omni includes a series of end-to-end speech interaction models along with training and inference code, covering but not limited to end-to-end speech interaction, end-to-end speech translation and speech recognition.
code for paper "Cross-modal Contrastive Learning for Speech Translation" (NAACL 2022)
Code for NeurIPS 2023 paper "DASpeech: Directed Acyclic Transformer for Fast and High-quality Speech-to-Speech Translation".
Repository containing the open source code of works published at the FBK MT unit.
List of direct speech-to-speech translation papers.
Code for ACL 2022 main conference paper "STEMM: Self-learning with Speech-text Manifold Mixup for Speech Translation".
A fast parallel PyTorch implementation of the "CIF: Continuous Integrate-and-Fire for End-to-End Speech Recognition" https://arxiv.org/abs/1905.11235.
Source code for ACL 2023 paper "End-to-End Simultaneous Speech Translation with Differentiable Segmentation"
Code for the INTERSPEECH 2023 paper "Learning When to Speak: Latency and Quality Trade-offs for Simultaneous Speech-to-Speech Translation with Offline Models"
PyTorch toolkit for streaming speech recognition, speech translation and simultaneous translation based on fairseq.
๐ Framework for seamless fine-tuning of Whisper model on a multi-lingual dataset and deployment to prod.
Code for ACL 2024 main conference paper "Can We Achieve High-quality Direct Speech-to-Speech Translation Without Parallel Speech Data?".
This is an implementation of paper "End-to-end Speech Translation via Cross-modal Progressive Training" (Interspeech2021)
A High-Quality and Large-Scale Dataset for English-Vietnamese Speech Translation (INTERSPEECH 2022)
Code for ACL 2023 main conference paper "Understanding and Bridging the Modality Gap for Speech Translation".
Code for EMNLP 2022 main conference paper "Information-Transport-based Policy for Simultaneous Translation"
Code for ACL 2023 main conference paper "Back Translation for Speech-to-text Translation Without Transcripts".
Revisiting End-to-End Speech-to-Text Translation From Scratch
A corpus that can be used to train English-to-Italian End-to-End Speech-to-Text Machine Translation models
๐๏ธ This project combines multiple operations in Microsoft Azure Cognitive Services into one GUI, including QnA Maker, LUIS, Computer Vision, Custom Vision, Face, Form Recognizer, Text To Speech, Speech To Text and Speech Translation. It's very user-friendly for users to implement any operation mentioned above.