There are 8 repositories under librosa topic.
Building and training Speech Emotion Recognizer that predicts human emotions using Python, Sci-kit learn and Keras
AudioMuse-AI is an Open Source Dockerized environment that brings automatic playlist generation to Jellyfin, Navidrome, LMS and Lyrion. Using powerful tools like Librosa and ONNX, it performs sonic analysis on your audio files locally, allowing you to curate the perfect playlist for any mood or occasion without relying on external APIs.
Lightweight and Interpretable ML Model for Speech Emotion Recognition and Ambiguity Resolution (trained on IEMOCAP dataset)
Understanding emotions from audio files using neural networks and multiple datasets.
A Machine Learning Approach of Emotional Model
A command-line music video generator based on rhythm
An open-source Python library for audio time-scale modification.
LibrosaCpp is a c++ implemention of librosa to compute short-time fourier transform coefficients,mel spectrogram or mfcc
Speech Emotion Recognition (SER) in real-time, using Deep Neural Networks (DNN) of Long Short Memory Term (LSTM).
(monophonic) audio to midi converter using Python and librosa
NineSong aims to provide Cloud native and AI extended solutions for data sharing in various ToC businesses
Python framework for Speech and Music Detection using Keras.
Human Emotion Understanding using multimodal dataset.
In this project is presented a simple method to train an MLP neural network for audio signals. The trained model can be exported on a Raspberry Pi (2 or superior suggested) to classify audio signal registered with USB microphone
基于PaddlePaddle实现的音频分类,支持EcapaTdnn、PANNS、TDNN、Res2Net、ResNetSE等各种模型,还有多种预处理方法
一个简单的小网页,录入人声哼唱,转化成钢琴音及钢琴谱输出。灵感稍纵即逝,本项目的目标是能够记录下一段小调,以音频形式输入,读取识别其曲调,并制成谱子,最终以钢琴弹奏的形式输出,依此将一些日常生活中的小灵感保存起来,以便日后回忆甚至再创作。
Music Synthesis with Python talk, originally given at PyGotham 2017.
Music genre classification model using CRNN
Environmental sound classification with Convolutional neural networks and the UrbanSound8K dataset.
Image Processing, Speech Processing, Encoder Decoder, Research Paper implementation
Advanced ML Project : An Orca Call classifier using mel-spectrograms as audio representations to detect Killer whales
Music genre classification from audio spectrograms using deep learning
Sound Classification using Neural Networks
Jellyfin plugin for AudioMuse-AI
We'll look into audio categorization using deep learning principles like Artificial Neural Networks (ANN), 1D Convolutional Neural Networks (CNN1D), and CNN2D in this repository. We undertake some basic data preprocessing and feature extraction on audio sources before developing models. As a result, the accuracy, training time, and prediction time of each model are compared. This is explained by model deployment, which allows users to load the desired sound output for each model that is successfully deployed, as will be addressed in more depth later.
Predicting various emotion in human speech signal by detecting different speech components affected by human emotion.
MIMII Sound Anomaly Detection with AutoEncoders
Binary classification problem that aims to classify human voices from audio recordings. Implemented using PyTorch and Librosa.
Classifying English Music (.mp3) files using Music Information Retrieval (MIR), Digital/Audio Signal Processing (DIP) and Machine Learning (ML) Strategies
Artificial intelligence bot for live voice improvisation
In this project we use RAVDESS Dataset to classify Speech Emotion using Multi Layer Perceptron Classifier
Methods to compute various chroma audio features and audio similarity measures particularly for the task of cover song identification