GithubHelp home page GithubHelp logo

dxmsjtu / 2020_interspeech_gmdp Goto Github PK

View Code? Open in Web Editor NEW

This project forked from fakufaku/2020_interspeech_gmdp

0.0 0.0 0.0 13.64 MB

Generalized Minimal Distortion Principle for Blind Source Separation

Home Page: https://arxiv.org/abs/2009.05288

License: MIT License

Python 100.00%

2020_interspeech_gmdp's Introduction

Generalized Minimal Distortion Principle for Blind Source Separation

Abstract

We revisit the source image estimation problem from blind source separation (BSS). We generalize the traditional minimum distortion principle to maximum likelihood estimation with a model for the residual spectrograms. Because residual spectrograms typically contain other sources, we propose to use a mixed-norm model that lets us finely tune sparsity in time and frequency. We propose to carry out the minimization of the mixed-norm via majorization-maximization optimization, leading to an iteratively reweighted least-squares algorithm. The algorithm balances well efficiency and ease of implementation. We assess the performance of the proposed method as applied to two well-known determined BSS and one joint BSS-dereverberation algorithms. We find out that it is possible to tune the parameters to improve separation by up to 2 dB, with no increase in distortion, and at little computational cost. The method thus provides a cheap and easy way to boost the performance of blind source separation.

Author

Robin Scheibler ([email protected])

Run the experiments

The easiest is to rely on anaconda or miniconda for the installation.

We use ipyparallel to parallelize the experiment.

Setup

# prepare the environment
git clone --recursive https://github.com/fakufaku/2020_interspeech_gdmp.git
cd 2020_interspeech_gdmp
conda env create -f environment.yml
conda activate gmdp

# generate the dataset
cd bss_speech_dataset
python ../config_dataset.json
cd ..

Run the Experiments

# start the engines
ipcluster start --daemonize

# run experiment for AuxIVA and ILRMA
python ./paper_simulation.py ./experiment1_config.json

# run experiment for ILRMA-T
python ./paper_simulation.py ./experiment2_config.json

# stop the engines
ipcluster stop

Create the Tables and Figures

In general, do the following

python ./analysis.py ./sim_results/<results_folder>

To recreate the figures with the simulation results used in the paper do

python ./analysis.py ./sim_results/20200511-112906_experiment1_config_102af93240
python ./analysis.py ./sim_results/20200507-012736_experiment2_config_102af93240

License

The code is released under MIT License.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.