This repository is a wrapper around several freely available implementations of objective metrics for estimating the quality of speech signals. It includes both absolute and relative metrics, which means metrics that do or do not need a reference signal, respectively.
If you find speechmetrics useful, you are welcome to cite the original papers for the corresponding metrics, since this is just a wrapper around the implementations that were kindly provided by the original authors.
As of our recent tests, installation goes smoothly on ubuntu, but there may be some compiler errors for pypesq
on iOs.
For cpu usage:
pip install numpy
pip install git+https://github.com/aliutkus/speechmetrics#egg=speechmetrics[cpu]
For gpu usage (on the MOSNet)
pip install numpy
pip install git+https://github.com/aliutkus/speechmetrics#egg=speechmetrics[gpu]
speechmetrics
has been designed to be easily used in a modular way. All you need to do is to specify the actual metrics you want to use and it will load them.
This behaviour is encapsulated in the load
function from the root of the package, that takes two arguments:
- metrics: str or list of str
the available metrics that match this argument will be automatically loaded. This matching is relative to the structure of the speechmetrics package.
For instance:
- 'absolute' will match all absolute metrics
- 'absolute.srmr' or 'srmr' will only match SRMR
- '' will match all
- window: float or None gives the length in seconds of the windows on which to compute the actual scores. If None, the whole signals will be considered.
import speechmetrics
window_length = 5 # seconds
metrics = speechmetrics.load('absolute', window_length)
scores = metrics(path_to_audio_file)
As provided by the authors of MOSNet: Deep Learning based Objective Assessment for Voice Conversion. Original github here
@article{lo2019mosnet,
title={MOSNet: Deep Learning based Objective Assessment for Voice Conversion},
author={Lo, Chen-Chou and Fu, Szu-Wei and Huang, Wen-Chin and Wang, Xin and Yamagishi, Junichi and Tsao, Yu and Wang, Hsin-Min},
journal={arXiv preprint arXiv:1904.08352},
year={2019} }
As provided by the SRMR Toolbox, implemented by @jfsantos.
-
@article{falk2010non,
title={A non-intrusive quality and intelligibility measure of reverberant and dereverberated speech},
author={Falk, Tiago H and Zheng, Chenxi and Chan, Wai-Yip},
journal={IEEE Transactions on Audio, Speech, and Language Processing},
volume={18},
number={7},
pages={1766--1774},
year={2010},
} -
@inproceedings{santos2014updated, title={An updated objective intelligibility estimation metric for normal hearing listeners under noise and reverberation},
author={Santos, Joo F and Senoussaoui, Mohammed and Falk, Tiago H},
booktitle={Proc. Int. Workshop Acoust. Signal Enhancement},
pages={55--59},
year={2014}
} -
@article{santos2014updating,
title={Updating the SRMR-CI metric for improved intelligibility prediction for cochlear implant users},
author={Santos, Jo{~a}o F and Falk, Tiago H},
journal={IEEE/ACM Transactions on Audio, Speech and Language Processing (TASLP)},
volume={22},
number={12},
pages={2197--2206},
year={2014},
}
As presented in this paper and freely available in the official museval page, corresponds to BSSEval v4. There are 3 submetrics handled here: SDR, SAR, ISR.
@InProceedings{SiSEC18,
author="St{"o}ter, Fabian-Robert and Liutkus, Antoine and Ito, Nobutaka",
title="The 2018 Signal Separation Evaluation Campaign",
booktitle="Latent Variable Analysis and Signal Separation: 14th International Conference, LVA/ICA 2018, Surrey, UK",
year="2018",
pages="293--305"
}
As implemented there by @vBaiCai.
As implemented by @mpariente here
-
@inproceedings{taal2010short,
title={A short-time objective intelligibility measure for time-frequency weighted noisy speech},
author={Taal, Cees H and Hendriks, Richard C and Heusdens, Richard and Jensen, Jesper},
booktitle={2010 IEEE International Conference on Acoustics, Speech and Signal Processing},
pages={4214--4217},
year={2010},
organization={IEEE}
} -
@article{taal2011algorithm,
title={An algorithm for intelligibility prediction of time--frequency weighted noisy speech},
author={Taal, Cees H and Hendriks, Richard C and Heusdens, Richard and Jensen, Jesper},
journal={IEEE Transactions on Audio, Speech, and Language Processing},
volume={19},
number={7},
pages={2125--2136},
year={2011},
publisher={IEEE}
} -
@article{jensen2016algorithm,
title={An algorithm for predicting the intelligibility of speech masked by modulated noise maskers},
author={Jensen, Jesper and Taal, Cees H},
journal={IEEE/ACM Transactions on Audio, Speech, and Language Processing},
volume={24},
number={11},
pages={2009--2022},
year={2016},
publisher={IEEE}
}