fakufaku / interspeech2023-moving-iva-samples

Repository containing samples produced by the method proposed in "Multi-channel separation of dynamic speech and sound events" and presented at Interspeech 2023.

Home Page:http://www.robinscheibler.org/interspeech2023-moving-iva-samples/

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Multi-channel separation of dynamic speech and sound events

This repo provides the supplementary materials for the following paper.

Takuya Fujimura and Robin Scheibler, "Multi-channel separation of dynamic speech and sound events," Interspeech 2023.

There are two experiments: one for speech signals, and one for sound events.

Speech separation

In the paper, we evaluated the separation performance for two speakers, including the cases of one or two moving speakers. As supplementary materials, we provide some samples of separated signals and the visualization of the attention weights. For the details, please refer to the speech/README.txt

Sound event detection (SED) with separation

In the paper, we evaluated the SED performance using separation. As supplementary materials, we provide some results of SED and separated signals. For the details, please refer to the sed/README.txt

Authors

Takuya Fujimura @ Nagoya University, Japan

Robin Scheibler @ LINE corporation, Japan

About

Repository containing samples produced by the method proposed in "Multi-channel separation of dynamic speech and sound events" and presented at Interspeech 2023.

http://www.robinscheibler.org/interspeech2023-moving-iva-samples/


Languages

Language:HTML 100.0%