Skip to content

Questions about evaluating recognizers #1079

Answered by sammlapp
ctmittelstaedt asked this question in Q&A
Discussion options

You must be logged in to vote

Hi Charlotte,

The default reported score metric is not the F1 score, it is the mean average precision. Note that F1 score depends on choosing a specific score threshold. You can use your model to predict on the validation set, threshold the predictions using a score, then compute the F1 score at a specific threshold using sklearn.metrics.precision_recall_f1_support (this is from memory the function name might be slightly different )

As for the confusion matrix, the idea of a confusion matrix is suited to single target classification but not multi target problems where a sample can have 0,1, or >1 correct labels. Most bioacoustics problems are framed as multi target classification of audio…

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@ctmittelstaedt
Comment options

Answer selected by ctmittelstaedt
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants