Hey everyone, I've stumbled upon an inconsistency with the F1 score and I can't seem to get around it. I have two lists y_true = [0, 1, 2, 2, 2] and y_pred = [0, 0, 2, 2, 1]. sklearn tells me that the macro-averaged F1 score is 0.488888... If I understand correctly the macro-average F1 score is the harmonic mean of the macro-average precision score and the macro-average recall score. sklearn tells me that the macro-average precision is 0.5 whilst the macro-average recall is 0.555555... If use the statistics.harmonic_mean function from Python's standard library this gives me around 0.526315.
So which is correct: 0.488888 or 0.526315? I apologize in advance if I've overlooked something silly. Best regards. -- Max Halford +336 28 25 13 38
_______________________________________________ scikit-learn mailing list scikit-learn@python.org https://mail.python.org/mailman/listinfo/scikit-learn