MMMMichaelzhang

MMMMichaelzhang

Geek Repo

Github PK Tool:Github PK Tool

MMMMichaelzhang's repositories

Language:JavaScriptStargazers:1Issues:0Issues:0

StarGANv2-VC

StarGANv2-VC: A Diverse, Unsupervised, Non-parallel Framework for Natural-Sounding Voice Conversion

Language:PythonLicense:MITStargazers:1Issues:0Issues:0

assem-vc

Official Code for Assem-VC @ICASSP2022

Language:Jupyter NotebookLicense:BSD-3-ClauseStargazers:0Issues:0Issues:0

asteroid

The PyTorch-based audio source separation toolkit for researchers || Pretrained models available

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

AudioMass

Free full-featured web-based audio & waveform editing tool

Language:JavaScriptStargazers:0Issues:0Issues:0

AuxiliaryASR

Phoneme-level ASR for Voice Conversion and TTS (Text-Mel Alignment)

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

bark

🔊 Text-Prompted Generative Audio Model

Language:Jupyter NotebookLicense:MITStargazers:0Issues:0Issues:0

CMGAN

Conformer-based Metric GAN for speech enhancement

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

Cross-Lingual-Voice-Cloning

Tacotron 2 - PyTorch implementation with faster-than-realtime inference modified to enable cross lingual voice cloning.

Language:Jupyter NotebookLicense:BSD-3-ClauseStargazers:0Issues:0Issues:0

DeepFilterNet

Noise supression using deep filtering

License:NOASSERTIONStargazers:0Issues:0Issues:0

demucs

Code for the paper Hybrid Spectrogram and Waveform Source Separation

License:MITStargazers:0Issues:0Issues:0

FACIAL

FACIAL: Synthesizing Dynamic Talking Face With Implicit Attribute Learning. ICCV, 2021.

License:AGPL-3.0Stargazers:0Issues:0Issues:0

fastVC

A simple voice conversion tool

Stargazers:0Issues:0Issues:0

FullSubNet-plus

The official PyTorch implementation of "FullSubNet+: Channel Attention FullSubNet with Complex Spectrograms for Speech Enhancement".

License:Apache-2.0Stargazers:0Issues:0Issues:0

GPT-SoVITS

1 min voice data can also be used to train a good TTS model! (few shot voice cloning)

License:MITStargazers:0Issues:0Issues:0

matchering

🎚️ Open Source Audio Matching and Mastering

License:GPL-3.0Stargazers:0Issues:0Issues:0

mellotron

Mellotron: a multispeaker voice synthesis model based on Tacotron 2 GST that can make a voice emote and sing without emotive or singing training data

License:BSD-3-ClauseStargazers:0Issues:0Issues:0

mir-svc

Unsupervised WaveNet-based Singing Voice Conversion Using Pitch Augmentation and Two-phase Approach

License:NOASSERTIONStargazers:0Issues:0Issues:0

MockingBird

🚀AI拟声: 5秒内克隆您的声音并生成任意语音内容 Clone a voice in 5 seconds to generate arbitrary speech in real-time

License:NOASSERTIONStargazers:0Issues:0Issues:0

NeuralSVB

Learning the Beauty in Songs: Neural Singing Voice Beautifier; ACL 2022 (Main conference); Official code

Language:PythonLicense:GPL-3.0Stargazers:0Issues:0Issues:0

nonparaSeq2seqVC_code

Implementation code of non-parallel sequence-to-sequence VC

License:MITStargazers:0Issues:0Issues:0
License:Apache-2.0Stargazers:0Issues:0Issues:0

ParallelWaveGAN

Unofficial Parallel WaveGAN (+ MelGAN & Multi-band MelGAN & HiFi-GAN & StyleMelGAN) with Pytorch

License:MITStargazers:0Issues:0Issues:0

PitchExtractor

Deep Neural Pitch Extractor for Voice Conversion and TTS Training

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

Retrieval-based-Voice-Conversion-WebUI

Voice data <= 10 mins can also be used to train a good VC model!

License:MITStargazers:0Issues:0Issues:0

so-vits-svc-fork

so-vits-svc fork with realtime support, improved interface and more features.

License:NOASSERTIONStargazers:0Issues:0Issues:0

ssr_eval

Evaluation and Benchmarking of Speech Super-resolution Methods

Stargazers:0Issues:0Issues:0

StyleTTS

Official Implementation of StyleTTS

License:MITStargazers:0Issues:0Issues:0

svoice

We provide a PyTorch implementation of the paper Voice Separation with an Unknown Number of Multiple Speakers In which, we present a new method for separating a mixed audio sequence, in which multiple voices speak simultaneously. The new method employs gated neural networks that are trained to separate the voices at multiple processing steps, while maintaining the speaker in each output channel fixed. A different model is trained for every number of possible speakers, and the model with the largest number of speakers is employed to select the actual number of speakers in a given sample. Our method greatly outperforms the current state of the art, which, as we show, is not competitive for more than two speakers.

License:NOASSERTIONStargazers:0Issues:0Issues:0

ultimatevocalremovergui

GUI for a Vocal Remover that uses Deep Neural Networks.

License:MITStargazers:0Issues:0Issues:0