Solved – How to control trade-off between precision and recall

classificationinterpretationmodel selectionprecision-recall

I applied different classification algorithms in combination with different sampling techniques to a dataset and I get > 100 different models with different performances.

I can choose a model for high precision or for high recall, but obviously not both at the same time.

Is there an approach/method/function out there where I can penalize either false positives or false negatives more – based on what is more/less important to me – so I can choose the perfect model out of all the ones I calculated?

Best Answer

Sure. You can use Fbeta score.

enter image description here

Beta = 1 means you value precision and recall equally, higher beta (beta > 1) means you value precision more then recall.

More on wiki: F1 score