Duo MA (shanguanma)

shanguanma

Geek Repo

Company:The Chinese University of Hong Kong, Shenzhen(CUHK-SZ); Shenzhen Research Institute of Big Data(SRIBD)

Location:Shenzhen

Home Page:nanr9544@gmail.com

Github PK Tool:Github PK Tool

Duo MA's repositories

Aligners

HMM, CTC, RNN-Transducer, forward-backward algorithm

Language:Jupyter NotebookStargazers:19Issues:3Issues:0

audiolm-pytorch

Implementation of AudioLM, a Language Modeling Approach to Audio Generation out of Google Research, in Pytorch

Language:PythonLicense:MITStargazers:0Issues:0Issues:0
Language:PythonStargazers:0Issues:1Issues:0

fairseq_speechtext

fairseq_speechtext project focus on dataset and model part of multi-modual pretraining(i.e: speech and text) for research.

Language:PythonLicense:MITStargazers:0Issues:2Issues:0
Language:PythonStargazers:0Issues:1Issues:0

s3prl

Audio Foundation Models (Self-Supervised Speech/Sound Pre-training and Representation Learning Toolkit)

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

bytepiece

更纯粹、更高压缩率的Tokenizer

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

diarizer

Clustering-based methods for overlapping diarization

Stargazers:0Issues:0Issues:0

fairseq2

FAIR Sequence Modeling Toolkit 2

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

flash-attention

Fast and memory-efficient exact attention

Language:PythonLicense:BSD-3-ClauseStargazers:0Issues:0Issues:0
Language:C++License:MITStargazers:0Issues:1Issues:0
Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

jsalt2020_simulate

Training data simulation

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

lit-gpt

Hackable implementation of state-of-the-art open-source LLMs based on nanoGPT. Supports flash attention, 4-bit and 8-bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

lit-llama

Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

llm.c

LLM training in simple, raw C/CUDA

Language:CudaLicense:MITStargazers:0Issues:0Issues:0

modern-cpp-tutorial

📚 Modern C++ Tutorial: C++11/14/17/20 On the Fly | https://changkun.de/modern-cpp/

Language:C++License:MITStargazers:0Issues:0Issues:0

nanoGPT

The simplest, fastest repository for training/finetuning medium-sized GPTs.

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

rir-generator

Room Impulse Response Generator

Language:PythonLicense:GPL-3.0Stargazers:0Issues:0Issues:0

scikit-learn

scikit-learn: machine learning in Python

Language:PythonLicense:BSD-3-ClauseStargazers:0Issues:0Issues:0
Stargazers:0Issues:1Issues:0
Language:CSSStargazers:0Issues:1Issues:0

sherpa-onnx

Speech-to-text and text-to-speech using next-gen Kaldi with onnxruntime without Internet connection. Support embedded systems, Android, iOS, Raspberry Pi, x86_64 servers, websocket server/client, C/C++, Python, Kotlin, C#, Go

Language:C++License:Apache-2.0Stargazers:0Issues:0Issues:0

Slurm_tools

My tools for the Slurm HPC workload manager

Language:ShellLicense:GPL-3.0Stargazers:0Issues:0Issues:0
Language:PythonStargazers:0Issues:1Issues:0

transfusion-asr

Transcribing Speech with Multinomial Diffusion, training code and models.

Language:PythonStargazers:0Issues:0Issues:0

tts

微软 tts 文本转语音 音频下载

Language:JavaScriptStargazers:0Issues:0Issues:0

voice-activity-detection

Pytorch implementation of SELF-ATTENTIVE VAD, ICASSP 2021

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

wenet

Production First and Production Ready End-to-End Speech Recognition Toolkit

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

wespeaker

Research and Production Oriented Speaker Recognition Toolkit

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0