from functools import partial from itertools import product from itertools import chain from itertools import permutations import warnings import re import numpy as np from scipy import linalg import pytest from sklearn import datasets from sklearn import svm from sklearn.datasets import make_multilabel_classification from sklearn.preprocessing import label_binarize, LabelBinarizer from sklearn.utils.validation import check_random_state from sklearn.utils._testing import assert_almost_equal from sklearn.utils._testing import assert_array_equal from sklearn.utils._testing import assert_array_almost_equal from sklearn.utils._testing import assert_allclose from sklearn.utils._testing import assert_warns_div0 from sklearn.utils._testing import assert_no_warnings from sklearn.utils._testing import assert_warns_message from sklearn.utils._testing import ignore_warnings from sklearn.utils._mocking import MockDataFrame from sklearn.metrics import accuracy_score from sklearn.metrics import average_precision_score from sklearn.metrics import balanced_accuracy_score from sklearn.metrics import classification_report from sklearn.metrics import cohen_kappa_score from sklearn.metrics import confusion_matrix from sklearn.metrics import f1_score from sklearn.metrics import fbeta_score from sklearn.metrics import hamming_loss from sklearn.metrics import hinge_loss from sklearn.metrics import jaccard_score from sklearn.metrics import log_loss from sklearn.metrics import matthews_corrcoef from sklearn.metrics import precision_recall_fscore_support from sklearn.metrics import precision_score from sklearn.metrics import recall_score from sklearn.metrics import zero_one_loss from sklearn.metrics import brier_score_loss from sklearn.metrics import multilabel_confusion_matrix from sklearn.metrics._classification import _check_targets from sklearn.exceptions import UndefinedMetricWarning from scipy.spatial.distance import hamming as sp_hamming ############################################################################### # Utilities for testing def make_prediction(dataset=None, binary=False): """Make some classification predictions on a toy dataset using a SVC If binary is True restrict to a binary classification problem instead of a multiclass classification problem """ if dataset is None: # import some data to play with dataset = datasets.load_iris() X = dataset.data y = dataset.target if binary: # restrict to a binary classification task X, y = X[y < 2], y[y < 2] n_samples, n_features = X.shape p = np.arange(n_samples) rng = check_random_state(37) rng.shuffle(p) X, y = X[p], y[p] half = int(n_samples / 2) # add noisy features to make the problem harder and avoid perfect results rng = np.random.RandomState(0) X = np.c_[X, rng.randn(n_samples, 200 * n_features)] # run classifier, get class probabilities and label predictions clf = svm.SVC(kernel='linear', probability=True, random_state=0) probas_pred = clf.fit(X[:half], y[:half]).predict_proba(X[half:]) if binary: # only interested in probabilities of the positive case # XXX: do we really want a special API for the binary case? probas_pred = probas_pred[:, 1] y_pred = clf.predict(X[half:]) y_true = y[half:] return y_true, y_pred, probas_pred ############################################################################### # Tests def test_classification_report_dictionary_output(): # Test performance report with dictionary output iris = datasets.load_iris() y_true, y_pred, _ = make_prediction(dataset=iris, binary=False) # print classification report with class names expected_report = {'setosa': {'precision': 0.82608695652173914, 'recall': 0.79166666666666663, 'f1-score': 0.8085106382978724, 'support': 24}, 'versicolor': {'precision': 0.33333333333333331, 'recall': 0.096774193548387094, 'f1-score': 0.15000000000000002, 'support': 31}, 'virginica': {'precision': 0.41860465116279072, 'recall': 0.90000000000000002, 'f1-score': 0.57142857142857151, 'support': 20}, 'macro avg': {'f1-score': 0.5099797365754813, 'precision': 0.5260083136726211, 'recall': 0.596146953405018, 'support': 75}, 'accuracy': 0.5333333333333333, 'weighted avg': {'f1-score': 0.47310435663627154, 'precision': 0.5137535108414785, 'recall': 0.5333333333333333, 'support': 75}} report = classification_report( y_true, y_pred, labels=np.arange(len(iris.target_names)), target_names=iris.target_names, output_dict=True) # assert the 2 dicts are equal. assert(report.keys() == expected_report.keys()) for key in expected_report: if key == 'accuracy': assert isinstance(report[key], float) assert report[key] == expected_report[key] else: assert report[key].keys() == expected_report[key].keys() for metric in expected_report[key]: assert_almost_equal(expected_report[key][metric], report[key][metric]) assert type(expected_report['setosa']['precision']) == float assert type(expected_report['macro avg']['precision']) == float assert type(expected_report['setosa']['support']) == int assert type(expected_report['macro avg']['support']) == int @pytest.mark.parametrize('zero_division', ["warn", 0, 1]) def test_classification_report_zero_division_warning(zero_division): y_true, y_pred = ["a", "b", "c"], ["a", "b", "d"] with warnings.catch_warnings(record=True) as record: classification_report( y_true, y_pred, zero_division=zero_division, output_dict=True) if zero_division == "warn": assert len(record) > 1 for item in record: msg = ("Use `zero_division` parameter to control this " "behavior.") assert msg in str(item.message) else: assert not record def test_multilabel_accuracy_score_subset_accuracy(): # Dense label indicator matrix format y1 = np.array([[0, 1, 1], [1, 0, 1]]) y2 = np.array([[0, 0, 1], [1, 0, 1]]) assert accuracy_score(y1, y2) == 0.5 assert accuracy_score(y1, y1) == 1 assert accuracy_score(y2, y2) == 1 assert accuracy_score(y2, np.logical_not(y2)) == 0 assert accuracy_score(y1, np.logical_not(y1)) == 0 assert accuracy_score(y1, np.zeros(y1.shape)) == 0 assert accuracy_score(y2, np.zeros(y1.shape)) == 0 def test_precision_recall_f1_score_binary(): # Test Precision Recall and F1 Score for binary classification task y_true, y_pred, _ = make_prediction(binary=True) # detailed measures for each class p, r, f, s = precision_recall_fscore_support(y_true, y_pred, average=None) assert_array_almost_equal(p, [0.73, 0.85], 2) assert_array_almost_equal(r, [0.88, 0.68], 2) assert_array_almost_equal(f, [0.80, 0.76], 2) assert_array_equal(s, [25, 25]) # individual scoring function that can be used for grid search: in the # binary class case the score is the value of the measure for the positive # class (e.g. label == 1). This is deprecated for average != 'binary'. for kwargs, my_assert in [({}, assert_no_warnings), ({'average': 'binary'}, assert_no_warnings)]: ps = my_assert(precision_score, y_true, y_pred, **kwargs) assert_array_almost_equal(ps, 0.85, 2) rs = my_assert(recall_score, y_true, y_pred, **kwargs) assert_array_almost_equal(rs, 0.68, 2) fs = my_assert(f1_score, y_true, y_pred, **kwargs) assert_array_almost_equal(fs, 0.76, 2) assert_almost_equal(my_assert(fbeta_score, y_true, y_pred, beta=2, **kwargs), (1 + 2 ** 2) * ps * rs / (2 ** 2 * ps + rs), 2) @ignore_warnings def test_precision_recall_f_binary_single_class(): # Test precision, recall and F-scores behave with a single positive or # negative class # Such a case may occur with non-stratified cross-validation assert 1. == precision_score([1, 1], [1, 1]) assert 1. == recall_score([1, 1], [1, 1]) assert 1. == f1_score([1, 1], [1, 1]) assert 1. == fbeta_score([1, 1], [1, 1], 0) assert 0. == precision_score([-1, -1], [-1, -1]) assert 0. == recall_score([-1, -1], [-1, -1]) assert 0. == f1_score([-1, -1], [-1, -1]) assert 0. == fbeta_score([-1, -1], [-1, -1], float('inf')) assert fbeta_score([-1, -1], [-1, -1], float('inf')) == pytest.approx( fbeta_score([-1, -1], [-1, -1], beta=1e5)) @ignore_warnings def test_precision_recall_f_extra_labels(): # Test handling of explicit additional (not in input) labels to PRF y_true = [1, 3, 3, 2] y_pred = [1, 1, 3, 2] y_true_bin = label_binarize(y_true, classes=np.arange(5)) y_pred_bin = label_binarize(y_pred, classes=np.arange(5)) data = [(y_true, y_pred), (y_true_bin, y_pred_bin)] for i, (y_true, y_pred) in enumerate(data): # No average: zeros in array actual = recall_score(y_true, y_pred, labels=[0, 1, 2, 3, 4], average=None) assert_array_almost_equal([0., 1., 1., .5, 0.], actual) # Macro average is changed actual = recall_score(y_true, y_pred, labels=[0, 1, 2, 3, 4], average='macro') assert_array_almost_equal(np.mean([0., 1., 1., .5, 0.]), actual) # No effect otheriwse for average in ['micro', 'weighted', 'samples']: if average == 'samples' and i == 0: continue assert_almost_equal(recall_score(y_true, y_pred, labels=[0, 1, 2, 3, 4], average=average), recall_score(y_true, y_pred, labels=None, average=average)) # Error when introducing invalid label in multilabel case # (although it would only affect performance if average='macro'/None) for average in [None, 'macro', 'micro', 'samples']: with pytest.raises(ValueError): recall_score(y_true_bin, y_pred_bin, labels=np.arange(6), average=average) with pytest.raises(ValueError): recall_score(y_true_bin, y_pred_bin, labels=np.arange(-1, 4), average=average) # tests non-regression on issue #10307 y_true = np.array([[0, 1, 1], [1, 0, 0]]) y_pred = np.array([[1, 1, 1], [1, 0, 1]]) p, r, f, _ = precision_recall_fscore_support(y_true, y_pred, average='samples', labels=[0, 1]) assert_almost_equal(np.array([p, r, f]), np.array([3 / 4, 1, 5 / 6])) @ignore_warnings def test_precision_recall_f_ignored_labels(): # Test a subset of labels may be requested for PRF y_true = [1, 1, 2, 3] y_pred = [1, 3, 3, 3] y_true_bin = label_binarize(y_true, classes=np.arange(5)) y_pred_bin = label_binarize(y_pred, classes=np.arange(5)) data = [(y_true, y_pred), (y_true_bin, y_pred_bin)] for i, (y_true, y_pred) in enumerate(data): recall_13 = partial(recall_score, y_true, y_pred, labels=[1, 3]) recall_all = partial(recall_score, y_true, y_pred, labels=None) assert_array_almost_equal([.5, 1.], recall_13(average=None)) assert_almost_equal((.5 + 1.) / 2, recall_13(average='macro')) assert_almost_equal((.5 * 2 + 1. * 1) / 3, recall_13(average='weighted')) assert_almost_equal(2. / 3, recall_13(average='micro')) # ensure the above were meaningful tests: for average in ['macro', 'weighted', 'micro']: assert (recall_13(average=average) != recall_all(average=average)) def test_average_precision_score_score_non_binary_class(): # Test that average_precision_score function returns an error when trying # to compute average_precision_score for multiclass task. rng = check_random_state(404) y_pred = rng.rand(10) # y_true contains three different class values y_true = rng.randint(0, 3, size=10) err_msg = "multiclass format is not supported" with pytest.raises(ValueError, match=err_msg): average_precision_score(y_true, y_pred) def test_average_precision_score_duplicate_values(): # Duplicate values with precision-recall require a different # processing than when computing the AUC of a ROC, because the # precision-recall curve is a decreasing curve # The following situation corresponds to a perfect # test statistic, the average_precision_score should be 1 y_true = [0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1] y_score = [0, .1, .1, .4, .5, .6, .6, .9, .9, 1, 1] assert average_precision_score(y_true, y_score) == 1 def test_average_precision_score_tied_values(): # Here if we go from left to right in y_true, the 0 values are # are separated from the 1 values, so it appears that we've # Correctly sorted our classifications. But in fact the first two # values have the same score (0.5) and so the first two values # could be swapped around, creating an imperfect sorting. This # imperfection should come through in the end score, making it less # than one. y_true = [0, 1, 1] y_score = [.5, .5, .6] assert average_precision_score(y_true, y_score) != 1. @ignore_warnings def test_precision_recall_fscore_support_errors(): y_true, y_pred, _ = make_prediction(binary=True) # Bad beta with pytest.raises(ValueError): precision_recall_fscore_support(y_true, y_pred, beta=-0.1) # Bad pos_label with pytest.raises(ValueError): precision_recall_fscore_support(y_true, y_pred, pos_label=2, average='binary') # Bad average option with pytest.raises(ValueError): precision_recall_fscore_support([0, 1, 2], [1, 2, 0], average='mega') def test_precision_recall_f_unused_pos_label(): # Check warning that pos_label unused when set to non-default value # but average != 'binary'; even if data is binary. assert_warns_message(UserWarning, "Note that pos_label (set to 2) is " "ignored when average != 'binary' (got 'macro'). You " "may use labels=[pos_label] to specify a single " "positive class.", precision_recall_fscore_support, [1, 2, 1], [1, 2, 2], pos_label=2, average='macro') def test_confusion_matrix_binary(): # Test confusion matrix - binary classification case y_true, y_pred, _ = make_prediction(binary=True) def test(y_true, y_pred): cm = confusion_matrix(y_true, y_pred) assert_array_equal(cm, [[22, 3], [8, 17]]) tp, fp, fn, tn = cm.flatten() num = (tp * tn - fp * fn) den = np.sqrt((tp + fp) * (tp + fn) * (tn + fp) * (tn + fn)) true_mcc = 0 if den == 0 else num / den mcc = matthews_corrcoef(y_true, y_pred) assert_array_almost_equal(mcc, true_mcc, decimal=2) assert_array_almost_equal(mcc, 0.57, decimal=2) test(y_true, y_pred) test([str(y) for y in y_true], [str(y) for y in y_pred]) def test_multilabel_confusion_matrix_binary(): # Test multilabel confusion matrix - binary classification case y_true, y_pred, _ = make_prediction(binary=True) def test(y_true, y_pred): cm = multilabel_confusion_matrix(y_true, y_pred) assert_array_equal(cm, [[[17, 8], [3, 22]], [[22, 3], [8, 17]]]) test(y_true, y_pred) test([str(y) for y in y_true], [str(y) for y in y_pred]) def test_multilabel_confusion_matrix_multiclass(): # Test multilabel confusion matrix - multi-class case y_true, y_pred, _ = make_prediction(binary=False) def test(y_true, y_pred, string_type=False): # compute confusion matrix with default labels introspection cm = multilabel_confusion_matrix(y_true, y_pred) assert_array_equal(cm, [[[47, 4], [5, 19]], [[38, 6], [28, 3]], [[30, 25], [2, 18]]]) # compute confusion matrix with explicit label ordering labels = ['0', '2', '1'] if string_type else [0, 2, 1] cm = multilabel_confusion_matrix(y_true, y_pred, labels=labels) assert_array_equal(cm, [[[47, 4], [5, 19]], [[30, 25], [2, 18]], [[38, 6], [28, 3]]]) # compute confusion matrix with super set of present labels labels = ['0', '2', '1', '3'] if string_type else [0, 2, 1, 3] cm = multilabel_confusion_matrix(y_true, y_pred, labels=labels) assert_array_equal(cm, [[[47, 4], [5, 19]], [[30, 25], [2, 18]], [[38, 6], [28, 3]], [[75, 0], [0, 0]]]) test(y_true, y_pred) test(list(str(y) for y in y_true), list(str(y) for y in y_pred), string_type=True) def test_multilabel_confusion_matrix_multilabel(): # Test multilabel confusion matrix - multilabel-indicator case from scipy.sparse import csc_matrix, csr_matrix y_true = np.array([[1, 0, 1], [0, 1, 0], [1, 1, 0]]) y_pred = np.array([[1, 0, 0], [0, 1, 1], [0, 0, 1]]) y_true_csr = csr_matrix(y_true) y_pred_csr = csr_matrix(y_pred) y_true_csc = csc_matrix(y_true) y_pred_csc = csc_matrix(y_pred) # cross test different types sample_weight = np.array([2, 1, 3]) real_cm = [[[1, 0], [1, 1]], [[1, 0], [1, 1]], [[0, 2], [1, 0]]] trues = [y_true, y_true_csr, y_true_csc] preds = [y_pred, y_pred_csr, y_pred_csc] for y_true_tmp in trues: for y_pred_tmp in preds: cm = multilabel_confusion_matrix(y_true_tmp, y_pred_tmp) assert_array_equal(cm, real_cm) # test support for samplewise cm = multilabel_confusion_matrix(y_true, y_pred, samplewise=True) assert_array_equal(cm, [[[1, 0], [1, 1]], [[1, 1], [0, 1]], [[0, 1], [2, 0]]]) # test support for labels cm = multilabel_confusion_matrix(y_true, y_pred, labels=[2, 0]) assert_array_equal(cm, [[[0, 2], [1, 0]], [[1, 0], [1, 1]]]) # test support for labels with samplewise cm = multilabel_confusion_matrix(y_true, y_pred, labels=[2, 0], samplewise=True) assert_array_equal(cm, [[[0, 0], [1, 1]], [[1, 1], [0, 0]], [[0, 1], [1, 0]]]) # test support for sample_weight with sample_wise cm = multilabel_confusion_matrix(y_true, y_pred, sample_weight=sample_weight, samplewise=True) assert_array_equal(cm, [[[2, 0], [2, 2]], [[1, 1], [0, 1]], [[0, 3], [6, 0]]]) def test_multilabel_confusion_matrix_errors(): y_true = np.array([[1, 0, 1], [0, 1, 0], [1, 1, 0]]) y_pred = np.array([[1, 0, 0], [0, 1, 1], [0, 0, 1]]) # Bad sample_weight with pytest.raises(ValueError, match="inconsistent numbers of samples"): multilabel_confusion_matrix(y_true, y_pred, sample_weight=[1, 2]) with pytest.raises(ValueError, match="should be a 1d array"): multilabel_confusion_matrix(y_true, y_pred, sample_weight=[[1, 2, 3], [2, 3, 4], [3, 4, 5]]) # Bad labels err_msg = r"All labels must be in \[0, n labels\)" with pytest.raises(ValueError, match=err_msg): multilabel_confusion_matrix(y_true, y_pred, labels=[-1]) err_msg = r"All labels must be in \[0, n labels\)" with pytest.raises(ValueError, match=err_msg): multilabel_confusion_matrix(y_true, y_pred, labels=[3]) # Using samplewise outside multilabel with pytest.raises(ValueError, match="Samplewise metrics"): multilabel_confusion_matrix([0, 1, 2], [1, 2, 0], samplewise=True) # Bad y_type err_msg = "multiclass-multioutput is not supported" with pytest.raises(ValueError, match=err_msg): multilabel_confusion_matrix([[0, 1, 2], [2, 1, 0]], [[1, 2, 0], [1, 0, 2]]) @pytest.mark.parametrize( "normalize, cm_dtype, expected_results", [('true', 'f', 0.333333333), ('pred', 'f', 0.333333333), ('all', 'f', 0.1111111111), (None, 'i', 2)] ) def test_confusion_matrix_normalize(normalize, cm_dtype, expected_results): y_test = [0, 1, 2] * 6 y_pred = list(chain(*permutations([0, 1, 2]))) cm = confusion_matrix(y_test, y_pred, normalize=normalize) assert_allclose(cm, expected_results) assert cm.dtype.kind == cm_dtype def test_confusion_matrix_normalize_wrong_option(): y_test = [0, 0, 0, 0, 1, 1, 1, 1] y_pred = [0, 0, 0, 0, 0, 0, 0, 0] with pytest.raises(ValueError, match='normalize must be one of'): confusion_matrix(y_test, y_pred, normalize=True) def test_confusion_matrix_normalize_single_class(): y_test = [0, 0, 0, 0, 1, 1, 1, 1] y_pred = [0, 0, 0, 0, 0, 0, 0, 0] cm_true = confusion_matrix(y_test, y_pred, normalize='true') assert cm_true.sum() == pytest.approx(2.0) # additionally check that no warnings are raised due to a division by zero with pytest.warns(None) as rec: cm_pred = confusion_matrix(y_test, y_pred, normalize='pred') assert not rec assert cm_pred.sum() == pytest.approx(1.0) with pytest.warns(None) as rec: cm_pred = confusion_matrix(y_pred, y_test, normalize='true') assert not rec def test_cohen_kappa(): # These label vectors reproduce the contingency matrix from Artstein and # Poesio (2008), Table 1: np.array([[20, 20], [10, 50]]). y1 = np.array([0] * 40 + [1] * 60) y2 = np.array([0] * 20 + [1] * 20 + [0] * 10 + [1] * 50) kappa = cohen_kappa_score(y1, y2) assert_almost_equal(kappa, .348, decimal=3) assert kappa == cohen_kappa_score(y2, y1) # Add spurious labels and ignore them. y1 = np.append(y1, [2] * 4) y2 = np.append(y2, [2] * 4) assert cohen_kappa_score(y1, y2, labels=[0, 1]) == kappa assert_almost_equal(cohen_kappa_score(y1, y1), 1.) # Multiclass example: Artstein and Poesio, Table 4. y1 = np.array([0] * 46 + [1] * 44 + [2] * 10) y2 = np.array([0] * 52 + [1] * 32 + [2] * 16) assert_almost_equal(cohen_kappa_score(y1, y2), .8013, decimal=4) # Weighting example: none, linear, quadratic. y1 = np.array([0] * 46 + [1] * 44 + [2] * 10) y2 = np.array([0] * 50 + [1] * 40 + [2] * 10) assert_almost_equal(cohen_kappa_score(y1, y2), .9315, decimal=4) assert_almost_equal(cohen_kappa_score(y1, y2, weights="linear"), 0.9412, decimal=4) assert_almost_equal(cohen_kappa_score(y1, y2, weights="quadratic"), 0.9541, decimal=4) @ignore_warnings def test_matthews_corrcoef_nan(): assert matthews_corrcoef([0], [1]) == 0.0 assert matthews_corrcoef([0, 0], [0, 1]) == 0.0 def test_matthews_corrcoef_against_numpy_corrcoef(): rng = np.random.RandomState(0) y_true = rng.randint(0, 2, size=20) y_pred = rng.randint(0, 2, size=20) assert_almost_equal(matthews_corrcoef(y_true, y_pred), np.corrcoef(y_true, y_pred)[0, 1], 10) def test_matthews_corrcoef_against_jurman(): # Check that the multiclass matthews_corrcoef agrees with the definition # presented in Jurman, Riccadonna, Furlanello, (2012). A Comparison of MCC # and CEN Error Measures in MultiClass Prediction rng = np.random.RandomState(0) y_true = rng.randint(0, 2, size=20) y_pred = rng.randint(0, 2, size=20) sample_weight = rng.rand(20) C = confusion_matrix(y_true, y_pred, sample_weight=sample_weight) N = len(C) cov_ytyp = sum([ C[k, k] * C[m, l] - C[l, k] * C[k, m] for k in range(N) for m in range(N) for l in range(N) ]) cov_ytyt = sum([ C[:, k].sum() * np.sum([C[g, f] for f in range(N) for g in range(N) if f != k]) for k in range(N) ]) cov_ypyp = np.sum([ C[k, :].sum() * np.sum([C[f, g] for f in range(N) for g in range(N) if f != k]) for k in range(N) ]) mcc_jurman = cov_ytyp / np.sqrt(cov_ytyt * cov_ypyp) mcc_ours = matthews_corrcoef(y_true, y_pred, sample_weight=sample_weight) assert_almost_equal(mcc_ours, mcc_jurman, 10) def test_matthews_corrcoef(): rng = np.random.RandomState(0) y_true = ["a" if i == 0 else "b" for i in rng.randint(0, 2, size=20)] # corrcoef of same vectors must be 1 assert_almost_equal(matthews_corrcoef(y_true, y_true), 1.0) # corrcoef, when the two vectors are opposites of each other, should be -1 y_true_inv = ["b" if i == "a" else "a" for i in y_true] assert_almost_equal(matthews_corrcoef(y_true, y_true_inv), -1) y_true_inv2 = label_binarize(y_true, classes=["a", "b"]) y_true_inv2 = np.where(y_true_inv2, 'a', 'b') assert_almost_equal(matthews_corrcoef(y_true, y_true_inv2), -1) # For the zero vector case, the corrcoef cannot be calculated and should # result in a RuntimeWarning mcc = assert_warns_div0(matthews_corrcoef, [0, 0, 0, 0], [0, 0, 0, 0]) # But will output 0 assert_almost_equal(mcc, 0.) # And also for any other vector with 0 variance mcc = assert_warns_div0(matthews_corrcoef, y_true, ['a'] * len(y_true)) # But will output 0 assert_almost_equal(mcc, 0.) # These two vectors have 0 correlation and hence mcc should be 0 y_1 = [1, 0, 1, 1, 0, 1, 1, 1, 0, 1, 1, 1, 1, 1, 1, 1, 0, 1, 1, 1] y_2 = [1, 1, 1, 0, 0, 1, 1, 1, 1, 0, 1, 1, 1, 0, 1, 1, 1, 0, 1, 1] assert_almost_equal(matthews_corrcoef(y_1, y_2), 0.) # Check that sample weight is able to selectively exclude mask = [1] * 10 + [0] * 10 # Now the first half of the vector elements are alone given a weight of 1 # and hence the mcc will not be a perfect 0 as in the previous case with pytest.raises(AssertionError): assert_almost_equal(matthews_corrcoef(y_1, y_2, sample_weight=mask), 0.) def test_matthews_corrcoef_multiclass(): rng = np.random.RandomState(0) ord_a = ord('a') n_classes = 4 y_true = [chr(ord_a + i) for i in rng.randint(0, n_classes, size=20)] # corrcoef of same vectors must be 1 assert_almost_equal(matthews_corrcoef(y_true, y_true), 1.0) # with multiclass > 2 it is not possible to achieve -1 y_true = [0, 0, 1, 1, 2, 2] y_pred_bad = [2, 2, 0, 0, 1, 1] assert_almost_equal(matthews_corrcoef(y_true, y_pred_bad), -.5) # Maximizing false positives and negatives minimizes the MCC # The minimum will be different for depending on the input y_true = [0, 0, 1, 1, 2, 2] y_pred_min = [1, 1, 0, 0, 0, 0] assert_almost_equal(matthews_corrcoef(y_true, y_pred_min), -12 / np.sqrt(24 * 16)) # Zero variance will result in an mcc of zero and a Runtime Warning y_true = [0, 1, 2] y_pred = [3, 3, 3] mcc = assert_warns_message(RuntimeWarning, 'invalid value encountered', matthews_corrcoef, y_true, y_pred) assert_almost_equal(mcc, 0.0) # These two vectors have 0 correlation and hence mcc should be 0 y_1 = [0, 1, 2, 0, 1, 2, 0, 1, 2] y_2 = [1, 1, 1, 2, 2, 2, 0, 0, 0] assert_almost_equal(matthews_corrcoef(y_1, y_2), 0.) # We can test that binary assumptions hold using the multiclass computation # by masking the weight of samples not in the first two classes # Masking the last label should let us get an MCC of -1 y_true = [0, 0, 1, 1, 2] y_pred = [1, 1, 0, 0, 2] sample_weight = [1, 1, 1, 1, 0] assert_almost_equal(matthews_corrcoef(y_true, y_pred, sample_weight=sample_weight), -1) # For the zero vector case, the corrcoef cannot be calculated and should # result in a RuntimeWarning y_true = [0, 0, 1, 2] y_pred = [0, 0, 1, 2] sample_weight = [1, 1, 0, 0] mcc = assert_warns_message(RuntimeWarning, 'invalid value encountered', matthews_corrcoef, y_true, y_pred, sample_weight=sample_weight) # But will output 0 assert_almost_equal(mcc, 0.) @pytest.mark.parametrize('n_points', [100, 10000]) def test_matthews_corrcoef_overflow(n_points): # https://github.com/scikit-learn/scikit-learn/issues/9622 rng = np.random.RandomState(20170906) def mcc_safe(y_true, y_pred): conf_matrix = confusion_matrix(y_true, y_pred) true_pos = conf_matrix[1, 1] false_pos = conf_matrix[1, 0] false_neg = conf_matrix[0, 1] n_points = len(y_true) pos_rate = (true_pos + false_neg) / n_points activity = (true_pos + false_pos) / n_points mcc_numerator = true_pos / n_points - pos_rate * activity mcc_denominator = activity * pos_rate * (1 - activity) * (1 - pos_rate) return mcc_numerator / np.sqrt(mcc_denominator) def random_ys(n_points): # binary x_true = rng.random_sample(n_points) x_pred = x_true + 0.2 * (rng.random_sample(n_points) - 0.5) y_true = (x_true > 0.5) y_pred = (x_pred > 0.5) return y_true, y_pred arr = np.repeat([0., 1.], n_points) # binary assert_almost_equal(matthews_corrcoef(arr, arr), 1.0) arr = np.repeat([0., 1., 2.], n_points) # multiclass assert_almost_equal(matthews_corrcoef(arr, arr), 1.0) y_true, y_pred = random_ys(n_points) assert_almost_equal(matthews_corrcoef(y_true, y_true), 1.0) assert_almost_equal(matthews_corrcoef(y_true, y_pred), mcc_safe(y_true, y_pred)) def test_precision_recall_f1_score_multiclass(): # Test Precision Recall and F1 Score for multiclass classification task y_true, y_pred, _ = make_prediction(binary=False) # compute scores with default labels introspection p, r, f, s = precision_recall_fscore_support(y_true, y_pred, average=None) assert_array_almost_equal(p, [0.83, 0.33, 0.42], 2) assert_array_almost_equal(r, [0.79, 0.09, 0.90], 2) assert_array_almost_equal(f, [0.81, 0.15, 0.57], 2) assert_array_equal(s, [24, 31, 20]) # averaging tests ps = precision_score(y_true, y_pred, pos_label=1, average='micro') assert_array_almost_equal(ps, 0.53, 2) rs = recall_score(y_true, y_pred, average='micro') assert_array_almost_equal(rs, 0.53, 2) fs = f1_score(y_true, y_pred, average='micro') assert_array_almost_equal(fs, 0.53, 2) ps = precision_score(y_true, y_pred, average='macro') assert_array_almost_equal(ps, 0.53, 2) rs = recall_score(y_true, y_pred, average='macro') assert_array_almost_equal(rs, 0.60, 2) fs = f1_score(y_true, y_pred, average='macro') assert_array_almost_equal(fs, 0.51, 2) ps = precision_score(y_true, y_pred, average='weighted') assert_array_almost_equal(ps, 0.51, 2) rs = recall_score(y_true, y_pred, average='weighted') assert_array_almost_equal(rs, 0.53, 2) fs = f1_score(y_true, y_pred, average='weighted') assert_array_almost_equal(fs, 0.47, 2) with pytest.raises(ValueError): precision_score(y_true, y_pred, average="samples") with pytest.raises(ValueError): recall_score(y_true, y_pred, average="samples") with pytest.raises(ValueError): f1_score(y_true, y_pred, average="samples") with pytest.raises(ValueError): fbeta_score(y_true, y_pred, average="samples", beta=0.5) # same prediction but with and explicit label ordering p, r, f, s = precision_recall_fscore_support( y_true, y_pred, labels=[0, 2, 1], average=None) assert_array_almost_equal(p, [0.83, 0.41, 0.33], 2) assert_array_almost_equal(r, [0.79, 0.90, 0.10], 2) assert_array_almost_equal(f, [0.81, 0.57, 0.15], 2) assert_array_equal(s, [24, 20, 31]) @pytest.mark.parametrize('average', ['samples', 'micro', 'macro', 'weighted', None]) def test_precision_refcall_f1_score_multilabel_unordered_labels(average): # test that labels need not be sorted in the multilabel case y_true = np.array([[1, 1, 0, 0]]) y_pred = np.array([[0, 0, 1, 1]]) p, r, f, s = precision_recall_fscore_support( y_true, y_pred, labels=[3, 0, 1, 2], warn_for=[], average=average) assert_array_equal(p, 0) assert_array_equal(r, 0) assert_array_equal(f, 0) if average is None: assert_array_equal(s, [0, 1, 1, 0]) def test_precision_recall_f1_score_binary_averaged(): y_true = np.array([0, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1]) y_pred = np.array([1, 1, 0, 1, 0, 1, 1, 1, 1, 0, 1, 0, 1, 0, 1]) # compute scores with default labels introspection ps, rs, fs, _ = precision_recall_fscore_support(y_true, y_pred, average=None) p, r, f, _ = precision_recall_fscore_support(y_true, y_pred, average='macro') assert p == np.mean(ps) assert r == np.mean(rs) assert f == np.mean(fs) p, r, f, _ = precision_recall_fscore_support(y_true, y_pred, average='weighted') support = np.bincount(y_true) assert p == np.average(ps, weights=support) assert r == np.average(rs, weights=support) assert f == np.average(fs, weights=support) def test_zero_precision_recall(): # Check that pathological cases do not bring NaNs old_error_settings = np.seterr(all='raise') try: y_true = np.array([0, 1, 2, 0, 1, 2]) y_pred = np.array([2, 0, 1, 1, 2, 0]) assert_almost_equal(precision_score(y_true, y_pred, average='macro'), 0.0, 2) assert_almost_equal(recall_score(y_true, y_pred, average='macro'), 0.0, 2) assert_almost_equal(f1_score(y_true, y_pred, average='macro'), 0.0, 2) finally: np.seterr(**old_error_settings) def test_confusion_matrix_multiclass_subset_labels(): # Test confusion matrix - multi-class case with subset of labels y_true, y_pred, _ = make_prediction(binary=False) # compute confusion matrix with only first two labels considered cm = confusion_matrix(y_true, y_pred, labels=[0, 1]) assert_array_equal(cm, [[19, 4], [4, 3]]) # compute confusion matrix with explicit label ordering for only subset # of labels cm = confusion_matrix(y_true, y_pred, labels=[2, 1]) assert_array_equal(cm, [[18, 2], [24, 3]]) # a label not in y_true should result in zeros for that row/column extra_label = np.max(y_true) + 1 cm = confusion_matrix(y_true, y_pred, labels=[2, extra_label]) assert_array_equal(cm, [[18, 0], [0, 0]]) @pytest.mark.parametrize( "labels, err_msg", [([], "'labels' should contains at least one label."), ([3, 4], "At least one label specified must be in y_true")], ids=["empty list", "unknown labels"] ) def test_confusion_matrix_error(labels, err_msg): y_true, y_pred, _ = make_prediction(binary=False) with pytest.raises(ValueError, match=err_msg): confusion_matrix(y_true, y_pred, labels=labels) @pytest.mark.parametrize( 'labels', (None, [0, 1], [0, 1, 2]), ids=['None', 'binary', 'multiclass'] ) def test_confusion_matrix_on_zero_length_input(labels): expected_n_classes = len(labels) if labels else 0 expected = np.zeros((expected_n_classes, expected_n_classes), dtype=np.int) cm = confusion_matrix([], [], labels=labels) assert_array_equal(cm, expected) def test_confusion_matrix_dtype(): y = [0, 1, 1] weight = np.ones(len(y)) # confusion_matrix returns int64 by default cm = confusion_matrix(y, y) assert cm.dtype == np.int64 # The dtype of confusion_matrix is always 64 bit for dtype in [np.bool_, np.int32, np.uint64]: cm = confusion_matrix(y, y, sample_weight=weight.astype(dtype, copy=False)) assert cm.dtype == np.int64 for dtype in [np.float32, np.float64, None, object]: cm = confusion_matrix(y, y, sample_weight=weight.astype(dtype, copy=False)) assert cm.dtype == np.float64 # np.iinfo(np.uint32).max should be accumulated correctly weight = np.full(len(y), 4294967295, dtype=np.uint32) cm = confusion_matrix(y, y, sample_weight=weight) assert cm[0, 0] == 4294967295 assert cm[1, 1] == 8589934590 # np.iinfo(np.int64).max should cause an overflow weight = np.full(len(y), 9223372036854775807, dtype=np.int64) cm = confusion_matrix(y, y, sample_weight=weight) assert cm[0, 0] == 9223372036854775807 assert cm[1, 1] == -2 def test_classification_report_multiclass(): # Test performance report iris = datasets.load_iris() y_true, y_pred, _ = make_prediction(dataset=iris, binary=False) # print classification report with class names expected_report = """\ precision recall f1-score support setosa 0.83 0.79 0.81 24 versicolor 0.33 0.10 0.15 31 virginica 0.42 0.90 0.57 20 accuracy 0.53 75 macro avg 0.53 0.60 0.51 75 weighted avg 0.51 0.53 0.47 75 """ report = classification_report( y_true, y_pred, labels=np.arange(len(iris.target_names)), target_names=iris.target_names) assert report == expected_report def test_classification_report_multiclass_balanced(): y_true, y_pred = [0, 0, 0, 1, 1, 1, 2, 2, 2], [0, 1, 2, 0, 1, 2, 0, 1, 2] expected_report = """\ precision recall f1-score support 0 0.33 0.33 0.33 3 1 0.33 0.33 0.33 3 2 0.33 0.33 0.33 3 accuracy 0.33 9 macro avg 0.33 0.33 0.33 9 weighted avg 0.33 0.33 0.33 9 """ report = classification_report(y_true, y_pred) assert report == expected_report def test_classification_report_multiclass_with_label_detection(): iris = datasets.load_iris() y_true, y_pred, _ = make_prediction(dataset=iris, binary=False) # print classification report with label detection expected_report = """\ precision recall f1-score support 0 0.83 0.79 0.81 24 1 0.33 0.10 0.15 31 2 0.42 0.90 0.57 20 accuracy 0.53 75 macro avg 0.53 0.60 0.51 75 weighted avg 0.51 0.53 0.47 75 """ report = classification_report(y_true, y_pred) assert report == expected_report def test_classification_report_multiclass_with_digits(): # Test performance report with added digits in floating point values iris = datasets.load_iris() y_true, y_pred, _ = make_prediction(dataset=iris, binary=False) # print classification report with class names expected_report = """\ precision recall f1-score support setosa 0.82609 0.79167 0.80851 24 versicolor 0.33333 0.09677 0.15000 31 virginica 0.41860 0.90000 0.57143 20 accuracy 0.53333 75 macro avg 0.52601 0.59615 0.50998 75 weighted avg 0.51375 0.53333 0.47310 75 """ report = classification_report( y_true, y_pred, labels=np.arange(len(iris.target_names)), target_names=iris.target_names, digits=5) assert report == expected_report def test_classification_report_multiclass_with_string_label(): y_true, y_pred, _ = make_prediction(binary=False) y_true = np.array(["blue", "green", "red"])[y_true] y_pred = np.array(["blue", "green", "red"])[y_pred] expected_report = """\ precision recall f1-score support blue 0.83 0.79 0.81 24 green 0.33 0.10 0.15 31 red 0.42 0.90 0.57 20 accuracy 0.53 75 macro avg 0.53 0.60 0.51 75 weighted avg 0.51 0.53 0.47 75 """ report = classification_report(y_true, y_pred) assert report == expected_report expected_report = """\ precision recall f1-score support a 0.83 0.79 0.81 24 b 0.33 0.10 0.15 31 c 0.42 0.90 0.57 20 accuracy 0.53 75 macro avg 0.53 0.60 0.51 75 weighted avg 0.51 0.53 0.47 75 """ report = classification_report(y_true, y_pred, target_names=["a", "b", "c"]) assert report == expected_report def test_classification_report_multiclass_with_unicode_label(): y_true, y_pred, _ = make_prediction(binary=False) labels = np.array(["blue\xa2", "green\xa2", "red\xa2"]) y_true = labels[y_true] y_pred = labels[y_pred] expected_report = """\ precision recall f1-score support blue\xa2 0.83 0.79 0.81 24 green\xa2 0.33 0.10 0.15 31 red\xa2 0.42 0.90 0.57 20 accuracy 0.53 75 macro avg 0.53 0.60 0.51 75 weighted avg 0.51 0.53 0.47 75 """ report = classification_report(y_true, y_pred) assert report == expected_report def test_classification_report_multiclass_with_long_string_label(): y_true, y_pred, _ = make_prediction(binary=False) labels = np.array(["blue", "green" * 5, "red"]) y_true = labels[y_true] y_pred = labels[y_pred] expected_report = """\ precision recall f1-score support blue 0.83 0.79 0.81 24 greengreengreengreengreen 0.33 0.10 0.15 31 red 0.42 0.90 0.57 20 accuracy 0.53 75 macro avg 0.53 0.60 0.51 75 weighted avg 0.51 0.53 0.47 75 """ report = classification_report(y_true, y_pred) assert report == expected_report def test_classification_report_labels_target_names_unequal_length(): y_true = [0, 0, 2, 0, 0] y_pred = [0, 2, 2, 0, 0] target_names = ['class 0', 'class 1', 'class 2'] assert_warns_message(UserWarning, "labels size, 2, does not " "match size of target_names, 3", classification_report, y_true, y_pred, labels=[0, 2], target_names=target_names) def test_classification_report_no_labels_target_names_unequal_length(): y_true = [0, 0, 2, 0, 0] y_pred = [0, 2, 2, 0, 0] target_names = ['class 0', 'class 1', 'class 2'] err_msg = ("Number of classes, 2, does not " "match size of target_names, 3. " "Try specifying the labels parameter") with pytest.raises(ValueError, match=err_msg): classification_report(y_true, y_pred, target_names=target_names) @ignore_warnings def test_multilabel_classification_report(): n_classes = 4 n_samples = 50 _, y_true = make_multilabel_classification(n_features=1, n_samples=n_samples, n_classes=n_classes, random_state=0) _, y_pred = make_multilabel_classification(n_features=1, n_samples=n_samples, n_classes=n_classes, random_state=1) expected_report = """\ precision recall f1-score support 0 0.50 0.67 0.57 24 1 0.51 0.74 0.61 27 2 0.29 0.08 0.12 26 3 0.52 0.56 0.54 27 micro avg 0.50 0.51 0.50 104 macro avg 0.45 0.51 0.46 104 weighted avg 0.45 0.51 0.46 104 samples avg 0.46 0.42 0.40 104 """ report = classification_report(y_true, y_pred) assert report == expected_report def test_multilabel_zero_one_loss_subset(): # Dense label indicator matrix format y1 = np.array([[0, 1, 1], [1, 0, 1]]) y2 = np.array([[0, 0, 1], [1, 0, 1]]) assert zero_one_loss(y1, y2) == 0.5 assert zero_one_loss(y1, y1) == 0 assert zero_one_loss(y2, y2) == 0 assert zero_one_loss(y2, np.logical_not(y2)) == 1 assert zero_one_loss(y1, np.logical_not(y1)) == 1 assert zero_one_loss(y1, np.zeros(y1.shape)) == 1 assert zero_one_loss(y2, np.zeros(y1.shape)) == 1 def test_multilabel_hamming_loss(): # Dense label indicator matrix format y1 = np.array([[0, 1, 1], [1, 0, 1]]) y2 = np.array([[0, 0, 1], [1, 0, 1]]) w = np.array([1, 3]) assert hamming_loss(y1, y2) == 1 / 6 assert hamming_loss(y1, y1) == 0 assert hamming_loss(y2, y2) == 0 assert hamming_loss(y2, 1 - y2) == 1 assert hamming_loss(y1, 1 - y1) == 1 assert hamming_loss(y1, np.zeros(y1.shape)) == 4 / 6 assert hamming_loss(y2, np.zeros(y1.shape)) == 0.5 assert hamming_loss(y1, y2, sample_weight=w) == 1. / 12 assert hamming_loss(y1, 1-y2, sample_weight=w) == 11. / 12 assert hamming_loss(y1, np.zeros_like(y1), sample_weight=w) == 2. / 3 # sp_hamming only works with 1-D arrays assert hamming_loss(y1[0], y2[0]) == sp_hamming(y1[0], y2[0]) def test_jaccard_score_validation(): y_true = np.array([0, 1, 0, 1, 1]) y_pred = np.array([0, 1, 0, 1, 1]) err_msg = r"pos_label=2 is not a valid label: array\(\[0, 1\]\)" with pytest.raises(ValueError, match=err_msg): jaccard_score(y_true, y_pred, average='binary', pos_label=2) y_true = np.array([[0, 1, 1], [1, 0, 0]]) y_pred = np.array([[1, 1, 1], [1, 0, 1]]) msg1 = (r"Target is multilabel-indicator but average='binary'. " r"Please choose another average setting, one of \[None, " r"'micro', 'macro', 'weighted', 'samples'\].") with pytest.raises(ValueError, match=msg1): jaccard_score(y_true, y_pred, average='binary', pos_label=-1) y_true = np.array([0, 1, 1, 0, 2]) y_pred = np.array([1, 1, 1, 1, 0]) msg2 = (r"Target is multiclass but average='binary'. Please choose " r"another average setting, one of \[None, 'micro', 'macro', " r"'weighted'\].") with pytest.raises(ValueError, match=msg2): jaccard_score(y_true, y_pred, average='binary') msg3 = ("Samplewise metrics are not available outside of multilabel " "classification.") with pytest.raises(ValueError, match=msg3): jaccard_score(y_true, y_pred, average='samples') assert_warns_message(UserWarning, "Note that pos_label (set to 3) is ignored when " "average != 'binary' (got 'micro'). You may use " "labels=[pos_label] to specify a single positive " "class.", jaccard_score, y_true, y_pred, average='micro', pos_label=3) def test_multilabel_jaccard_score(recwarn): # Dense label indicator matrix format y1 = np.array([[0, 1, 1], [1, 0, 1]]) y2 = np.array([[0, 0, 1], [1, 0, 1]]) # size(y1 \inter y2) = [1, 2] # size(y1 \union y2) = [2, 2] assert jaccard_score(y1, y2, average='samples') == 0.75 assert jaccard_score(y1, y1, average='samples') == 1 assert jaccard_score(y2, y2, average='samples') == 1 assert jaccard_score(y2, np.logical_not(y2), average='samples') == 0 assert jaccard_score(y1, np.logical_not(y1), average='samples') == 0 assert jaccard_score(y1, np.zeros(y1.shape), average='samples') == 0 assert jaccard_score(y2, np.zeros(y1.shape), average='samples') == 0 y_true = np.array([[0, 1, 1], [1, 0, 0]]) y_pred = np.array([[1, 1, 1], [1, 0, 1]]) # average='macro' assert_almost_equal(jaccard_score(y_true, y_pred, average='macro'), 2. / 3) # average='micro' assert_almost_equal(jaccard_score(y_true, y_pred, average='micro'), 3. / 5) # average='samples' assert_almost_equal(jaccard_score(y_true, y_pred, average='samples'), 7. / 12) assert_almost_equal(jaccard_score(y_true, y_pred, average='samples', labels=[0, 2]), 1. / 2) assert_almost_equal(jaccard_score(y_true, y_pred, average='samples', labels=[1, 2]), 1. / 2) # average=None assert_array_equal(jaccard_score(y_true, y_pred, average=None), np.array([1. / 2, 1., 1. / 2])) y_true = np.array([[0, 1, 1], [1, 0, 1]]) y_pred = np.array([[1, 1, 1], [1, 0, 1]]) assert_almost_equal(jaccard_score(y_true, y_pred, average='macro'), 5. / 6) # average='weighted' assert_almost_equal(jaccard_score(y_true, y_pred, average='weighted'), 7. / 8) msg2 = 'Got 4 > 2' with pytest.raises(ValueError, match=msg2): jaccard_score(y_true, y_pred, labels=[4], average='macro') msg3 = 'Got -1 < 0' with pytest.raises(ValueError, match=msg3): jaccard_score(y_true, y_pred, labels=[-1], average='macro') msg = ('Jaccard is ill-defined and being set to 0.0 in labels ' 'with no true or predicted samples.') assert assert_warns_message(UndefinedMetricWarning, msg, jaccard_score, np.array([[0, 1]]), np.array([[0, 1]]), average='macro') == 0.5 msg = ('Jaccard is ill-defined and being set to 0.0 in samples ' 'with no true or predicted labels.') assert assert_warns_message(UndefinedMetricWarning, msg, jaccard_score, np.array([[0, 0], [1, 1]]), np.array([[0, 0], [1, 1]]), average='samples') == 0.5 assert not list(recwarn) def test_multiclass_jaccard_score(recwarn): y_true = ['ant', 'ant', 'cat', 'cat', 'ant', 'cat', 'bird', 'bird'] y_pred = ['cat', 'ant', 'cat', 'cat', 'ant', 'bird', 'bird', 'cat'] labels = ['ant', 'bird', 'cat'] lb = LabelBinarizer() lb.fit(labels) y_true_bin = lb.transform(y_true) y_pred_bin = lb.transform(y_pred) multi_jaccard_score = partial(jaccard_score, y_true, y_pred) bin_jaccard_score = partial(jaccard_score, y_true_bin, y_pred_bin) multi_labels_list = [['ant', 'bird'], ['ant', 'cat'], ['cat', 'bird'], ['ant'], ['bird'], ['cat'], None] bin_labels_list = [[0, 1], [0, 2], [2, 1], [0], [1], [2], None] # other than average='samples'/'none-samples', test everything else here for average in ('macro', 'weighted', 'micro', None): for m_label, b_label in zip(multi_labels_list, bin_labels_list): assert_almost_equal(multi_jaccard_score(average=average, labels=m_label), bin_jaccard_score(average=average, labels=b_label)) y_true = np.array([[0, 0], [0, 0], [0, 0]]) y_pred = np.array([[0, 0], [0, 0], [0, 0]]) with ignore_warnings(): assert (jaccard_score(y_true, y_pred, average='weighted') == 0) assert not list(recwarn) def test_average_binary_jaccard_score(recwarn): # tp=0, fp=0, fn=1, tn=0 assert jaccard_score([1], [0], average='binary') == 0. # tp=0, fp=0, fn=0, tn=1 msg = ('Jaccard is ill-defined and being set to 0.0 due to ' 'no true or predicted samples') assert assert_warns_message(UndefinedMetricWarning, msg, jaccard_score, [0, 0], [0, 0], average='binary') == 0. # tp=1, fp=0, fn=0, tn=0 (pos_label=0) assert jaccard_score([0], [0], pos_label=0, average='binary') == 1. y_true = np.array([1, 0, 1, 1, 0]) y_pred = np.array([1, 0, 1, 1, 1]) assert_almost_equal(jaccard_score(y_true, y_pred, average='binary'), 3. / 4) assert_almost_equal(jaccard_score(y_true, y_pred, average='binary', pos_label=0), 1. / 2) assert not list(recwarn) @ignore_warnings def test_precision_recall_f1_score_multilabel_1(): # Test precision_recall_f1_score on a crafted multilabel example # First crafted example y_true = np.array([[1, 0, 0, 0], [0, 1, 0, 0], [0, 0, 1, 1]]) y_pred = np.array([[0, 1, 0, 0], [0, 1, 0, 0], [1, 0, 1, 0]]) p, r, f, s = precision_recall_fscore_support(y_true, y_pred, average=None) # tp = [0, 1, 1, 0] # fn = [1, 0, 0, 1] # fp = [1, 1, 0, 0] # Check per class assert_array_almost_equal(p, [0.0, 0.5, 1.0, 0.0], 2) assert_array_almost_equal(r, [0.0, 1.0, 1.0, 0.0], 2) assert_array_almost_equal(f, [0.0, 1 / 1.5, 1, 0.0], 2) assert_array_almost_equal(s, [1, 1, 1, 1], 2) f2 = fbeta_score(y_true, y_pred, beta=2, average=None) support = s assert_array_almost_equal(f2, [0, 0.83, 1, 0], 2) # Check macro p, r, f, s = precision_recall_fscore_support(y_true, y_pred, average="macro") assert_almost_equal(p, 1.5 / 4) assert_almost_equal(r, 0.5) assert_almost_equal(f, 2.5 / 1.5 * 0.25) assert s is None assert_almost_equal(fbeta_score(y_true, y_pred, beta=2, average="macro"), np.mean(f2)) # Check micro p, r, f, s = precision_recall_fscore_support(y_true, y_pred, average="micro") assert_almost_equal(p, 0.5) assert_almost_equal(r, 0.5) assert_almost_equal(f, 0.5) assert s is None assert_almost_equal(fbeta_score(y_true, y_pred, beta=2, average="micro"), (1 + 4) * p * r / (4 * p + r)) # Check weighted p, r, f, s = precision_recall_fscore_support(y_true, y_pred, average="weighted") assert_almost_equal(p, 1.5 / 4) assert_almost_equal(r, 0.5) assert_almost_equal(f, 2.5 / 1.5 * 0.25) assert s is None assert_almost_equal(fbeta_score(y_true, y_pred, beta=2, average="weighted"), np.average(f2, weights=support)) # Check samples # |h(x_i) inter y_i | = [0, 1, 1] # |y_i| = [1, 1, 2] # |h(x_i)| = [1, 1, 2] p, r, f, s = precision_recall_fscore_support(y_true, y_pred, average="samples") assert_almost_equal(p, 0.5) assert_almost_equal(r, 0.5) assert_almost_equal(f, 0.5) assert s is None assert_almost_equal(fbeta_score(y_true, y_pred, beta=2, average="samples"), 0.5) @ignore_warnings def test_precision_recall_f1_score_multilabel_2(): # Test precision_recall_f1_score on a crafted multilabel example 2 # Second crafted example y_true = np.array([[1, 0, 0, 0], [0, 1, 0, 0], [0, 1, 1, 0]]) y_pred = np.array([[0, 0, 0, 1], [0, 0, 0, 1], [1, 1, 0, 0]]) # tp = [ 0. 1. 0. 0.] # fp = [ 1. 0. 0. 2.] # fn = [ 1. 1. 1. 0.] p, r, f, s = precision_recall_fscore_support(y_true, y_pred, average=None) assert_array_almost_equal(p, [0.0, 1.0, 0.0, 0.0], 2) assert_array_almost_equal(r, [0.0, 0.5, 0.0, 0.0], 2) assert_array_almost_equal(f, [0.0, 0.66, 0.0, 0.0], 2) assert_array_almost_equal(s, [1, 2, 1, 0], 2) f2 = fbeta_score(y_true, y_pred, beta=2, average=None) support = s assert_array_almost_equal(f2, [0, 0.55, 0, 0], 2) p, r, f, s = precision_recall_fscore_support(y_true, y_pred, average="micro") assert_almost_equal(p, 0.25) assert_almost_equal(r, 0.25) assert_almost_equal(f, 2 * 0.25 * 0.25 / 0.5) assert s is None assert_almost_equal(fbeta_score(y_true, y_pred, beta=2, average="micro"), (1 + 4) * p * r / (4 * p + r)) p, r, f, s = precision_recall_fscore_support(y_true, y_pred, average="macro") assert_almost_equal(p, 0.25) assert_almost_equal(r, 0.125) assert_almost_equal(f, 2 / 12) assert s is None assert_almost_equal(fbeta_score(y_true, y_pred, beta=2, average="macro"), np.mean(f2)) p, r, f, s = precision_recall_fscore_support(y_true, y_pred, average="weighted") assert_almost_equal(p, 2 / 4) assert_almost_equal(r, 1 / 4) assert_almost_equal(f, 2 / 3 * 2 / 4) assert s is None assert_almost_equal(fbeta_score(y_true, y_pred, beta=2, average="weighted"), np.average(f2, weights=support)) p, r, f, s = precision_recall_fscore_support(y_true, y_pred, average="samples") # Check samples # |h(x_i) inter y_i | = [0, 0, 1] # |y_i| = [1, 1, 2] # |h(x_i)| = [1, 1, 2] assert_almost_equal(p, 1 / 6) assert_almost_equal(r, 1 / 6) assert_almost_equal(f, 2 / 4 * 1 / 3) assert s is None assert_almost_equal(fbeta_score(y_true, y_pred, beta=2, average="samples"), 0.1666, 2) @ignore_warnings @pytest.mark.parametrize('zero_division', ["warn", 0, 1]) def test_precision_recall_f1_score_with_an_empty_prediction(zero_division): y_true = np.array([[0, 1, 0, 0], [1, 0, 0, 0], [0, 1, 1, 0]]) y_pred = np.array([[0, 0, 0, 0], [0, 0, 0, 1], [0, 1, 1, 0]]) # true_pos = [ 0. 1. 1. 0.] # false_pos = [ 0. 0. 0. 1.] # false_neg = [ 1. 1. 0. 0.] zero_division = 1.0 if zero_division == 1.0 else 0.0 p, r, f, s = precision_recall_fscore_support(y_true, y_pred, average=None, zero_division=zero_division) assert_array_almost_equal(p, [zero_division, 1.0, 1.0, 0.0], 2) assert_array_almost_equal(r, [0.0, 0.5, 1.0, zero_division], 2) assert_array_almost_equal(f, [0.0, 1 / 1.5, 1, 0.0], 2) assert_array_almost_equal(s, [1, 2, 1, 0], 2) f2 = fbeta_score(y_true, y_pred, beta=2, average=None, zero_division=zero_division) support = s assert_array_almost_equal(f2, [0, 0.55, 1, 0], 2) p, r, f, s = precision_recall_fscore_support(y_true, y_pred, average="macro", zero_division=zero_division) assert_almost_equal(p, (2 + zero_division) / 4) assert_almost_equal(r, (1.5 + zero_division) / 4) assert_almost_equal(f, 2.5 / (4 * 1.5)) assert s is None assert_almost_equal(fbeta_score(y_true, y_pred, beta=2, average="macro"), np.mean(f2)) p, r, f, s = precision_recall_fscore_support(y_true, y_pred, average="micro", zero_division=zero_division) assert_almost_equal(p, 2 / 3) assert_almost_equal(r, 0.5) assert_almost_equal(f, 2 / 3 / (2 / 3 + 0.5)) assert s is None assert_almost_equal(fbeta_score(y_true, y_pred, beta=2, average="micro", zero_division=zero_division), (1 + 4) * p * r / (4 * p + r)) p, r, f, s = precision_recall_fscore_support(y_true, y_pred, average="weighted", zero_division=zero_division) assert_almost_equal(p, 3 / 4 if zero_division == 0 else 1.0) assert_almost_equal(r, 0.5) assert_almost_equal(f, (2 / 1.5 + 1) / 4) assert s is None assert_almost_equal(fbeta_score(y_true, y_pred, beta=2, average="weighted", zero_division=zero_division), np.average(f2, weights=support), ) p, r, f, s = precision_recall_fscore_support(y_true, y_pred, average="samples") # |h(x_i) inter y_i | = [0, 0, 2] # |y_i| = [1, 1, 2] # |h(x_i)| = [0, 1, 2] assert_almost_equal(p, 1 / 3) assert_almost_equal(r, 1 / 3) assert_almost_equal(f, 1 / 3) assert s is None assert_almost_equal(fbeta_score(y_true, y_pred, beta=2, average="samples", zero_division=zero_division), 0.333, 2) @pytest.mark.parametrize('beta', [1]) @pytest.mark.parametrize('average', ["macro", "micro", "weighted", "samples"]) @pytest.mark.parametrize('zero_division', [0, 1]) def test_precision_recall_f1_no_labels(beta, average, zero_division): y_true = np.zeros((20, 3)) y_pred = np.zeros_like(y_true) p, r, f, s = assert_no_warnings(precision_recall_fscore_support, y_true, y_pred, average=average, beta=beta, zero_division=zero_division) fbeta = assert_no_warnings(fbeta_score, y_true, y_pred, beta=beta, average=average, zero_division=zero_division) zero_division = float(zero_division) assert_almost_equal(p, zero_division) assert_almost_equal(r, zero_division) assert_almost_equal(f, zero_division) assert s is None assert_almost_equal(fbeta, float(zero_division)) @pytest.mark.parametrize('average', ["macro", "micro", "weighted", "samples"]) def test_precision_recall_f1_no_labels_check_warnings(average): y_true = np.zeros((20, 3)) y_pred = np.zeros_like(y_true) func = precision_recall_fscore_support with pytest.warns(UndefinedMetricWarning): p, r, f, s = func(y_true, y_pred, average=average, beta=1.0) assert_almost_equal(p, 0) assert_almost_equal(r, 0) assert_almost_equal(f, 0) assert s is None with pytest.warns(UndefinedMetricWarning): fbeta = fbeta_score(y_true, y_pred, average=average, beta=1.0) assert_almost_equal(fbeta, 0) @pytest.mark.parametrize('zero_division', [0, 1]) def test_precision_recall_f1_no_labels_average_none(zero_division): y_true = np.zeros((20, 3)) y_pred = np.zeros_like(y_true) # tp = [0, 0, 0] # fn = [0, 0, 0] # fp = [0, 0, 0] # support = [0, 0, 0] # |y_hat_i inter y_i | = [0, 0, 0] # |y_i| = [0, 0, 0] # |y_hat_i| = [0, 0, 0] p, r, f, s = assert_no_warnings(precision_recall_fscore_support, y_true, y_pred, average=None, beta=1.0, zero_division=zero_division) fbeta = assert_no_warnings(fbeta_score, y_true, y_pred, beta=1.0, average=None, zero_division=zero_division) zero_division = float(zero_division) assert_array_almost_equal( p, [zero_division, zero_division, zero_division], 2 ) assert_array_almost_equal( r, [zero_division, zero_division, zero_division], 2 ) assert_array_almost_equal( f, [zero_division, zero_division, zero_division], 2 ) assert_array_almost_equal(s, [0, 0, 0], 2) assert_array_almost_equal( fbeta, [zero_division, zero_division, zero_division], 2 ) def test_precision_recall_f1_no_labels_average_none_warn(): y_true = np.zeros((20, 3)) y_pred = np.zeros_like(y_true) # tp = [0, 0, 0] # fn = [0, 0, 0] # fp = [0, 0, 0] # support = [0, 0, 0] # |y_hat_i inter y_i | = [0, 0, 0] # |y_i| = [0, 0, 0] # |y_hat_i| = [0, 0, 0] with pytest.warns(UndefinedMetricWarning): p, r, f, s = precision_recall_fscore_support( y_true, y_pred, average=None, beta=1 ) assert_array_almost_equal(p, [0, 0, 0], 2) assert_array_almost_equal(r, [0, 0, 0], 2) assert_array_almost_equal(f, [0, 0, 0], 2) assert_array_almost_equal(s, [0, 0, 0], 2) with pytest.warns(UndefinedMetricWarning): fbeta = fbeta_score(y_true, y_pred, beta=1, average=None) assert_array_almost_equal(fbeta, [0, 0, 0], 2) def test_prf_warnings(): # average of per-label scores f, w = precision_recall_fscore_support, UndefinedMetricWarning for average in [None, 'weighted', 'macro']: msg = ('Precision and F-score are ill-defined and ' 'being set to 0.0 in labels with no predicted samples.' ' Use `zero_division` parameter to control' ' this behavior.') assert_warns_message(w, msg, f, [0, 1, 2], [1, 1, 2], average=average) msg = ('Recall and F-score are ill-defined and ' 'being set to 0.0 in labels with no true samples.' ' Use `zero_division` parameter to control' ' this behavior.') assert_warns_message(w, msg, f, [1, 1, 2], [0, 1, 2], average=average) # average of per-sample scores msg = ('Precision and F-score are ill-defined and ' 'being set to 0.0 in samples with no predicted labels.' ' Use `zero_division` parameter to control' ' this behavior.') assert_warns_message(w, msg, f, np.array([[1, 0], [1, 0]]), np.array([[1, 0], [0, 0]]), average='samples') msg = ('Recall and F-score are ill-defined and ' 'being set to 0.0 in samples with no true labels.' ' Use `zero_division` parameter to control' ' this behavior.') assert_warns_message(w, msg, f, np.array([[1, 0], [0, 0]]), np.array([[1, 0], [1, 0]]), average='samples') # single score: micro-average msg = ('Precision and F-score are ill-defined and ' 'being set to 0.0 due to no predicted samples.' ' Use `zero_division` parameter to control' ' this behavior.') assert_warns_message(w, msg, f, np.array([[1, 1], [1, 1]]), np.array([[0, 0], [0, 0]]), average='micro') msg = ('Recall and F-score are ill-defined and ' 'being set to 0.0 due to no true samples.' ' Use `zero_division` parameter to control' ' this behavior.') assert_warns_message(w, msg, f, np.array([[0, 0], [0, 0]]), np.array([[1, 1], [1, 1]]), average='micro') # single positive label msg = ('Precision and F-score are ill-defined and ' 'being set to 0.0 due to no predicted samples.' ' Use `zero_division` parameter to control' ' this behavior.') assert_warns_message(w, msg, f, [1, 1], [-1, -1], average='binary') msg = ('Recall and F-score are ill-defined and ' 'being set to 0.0 due to no true samples.' ' Use `zero_division` parameter to control' ' this behavior.') assert_warns_message(w, msg, f, [-1, -1], [1, 1], average='binary') with warnings.catch_warnings(record=True) as record: warnings.simplefilter('always') precision_recall_fscore_support([0, 0], [0, 0], average="binary") msg = ('Recall and F-score are ill-defined and ' 'being set to 0.0 due to no true samples.' ' Use `zero_division` parameter to control' ' this behavior.') assert str(record.pop().message) == msg msg = ('Precision and F-score are ill-defined and ' 'being set to 0.0 due to no predicted samples.' ' Use `zero_division` parameter to control' ' this behavior.') assert str(record.pop().message) == msg @pytest.mark.parametrize('zero_division', [0, 1]) def test_prf_no_warnings_if_zero_division_set(zero_division): # average of per-label scores f = precision_recall_fscore_support for average in [None, 'weighted', 'macro']: assert_no_warnings(f, [0, 1, 2], [1, 1, 2], average=average, zero_division=zero_division) assert_no_warnings(f, [1, 1, 2], [0, 1, 2], average=average, zero_division=zero_division) # average of per-sample scores assert_no_warnings(f, np.array([[1, 0], [1, 0]]), np.array([[1, 0], [0, 0]]), average='samples', zero_division=zero_division) assert_no_warnings(f, np.array([[1, 0], [0, 0]]), np.array([[1, 0], [1, 0]]), average='samples', zero_division=zero_division) # single score: micro-average assert_no_warnings(f, np.array([[1, 1], [1, 1]]), np.array([[0, 0], [0, 0]]), average='micro', zero_division=zero_division) assert_no_warnings(f, np.array([[0, 0], [0, 0]]), np.array([[1, 1], [1, 1]]), average='micro', zero_division=zero_division) # single positive label assert_no_warnings(f, [1, 1], [-1, -1], average='binary', zero_division=zero_division) assert_no_warnings(f, [-1, -1], [1, 1], average='binary', zero_division=zero_division) with warnings.catch_warnings(record=True) as record: warnings.simplefilter('always') precision_recall_fscore_support([0, 0], [0, 0], average="binary", zero_division=zero_division) assert len(record) == 0 @pytest.mark.parametrize('zero_division', ["warn", 0, 1]) def test_recall_warnings(zero_division): assert_no_warnings(recall_score, np.array([[1, 1], [1, 1]]), np.array([[0, 0], [0, 0]]), average='micro', zero_division=zero_division) with warnings.catch_warnings(record=True) as record: warnings.simplefilter('always') recall_score(np.array([[0, 0], [0, 0]]), np.array([[1, 1], [1, 1]]), average='micro', zero_division=zero_division) if zero_division == "warn": assert (str(record.pop().message) == 'Recall is ill-defined and ' 'being set to 0.0 due to no true samples.' ' Use `zero_division` parameter to control' ' this behavior.') else: assert len(record) == 0 recall_score([0, 0], [0, 0]) if zero_division == "warn": assert (str(record.pop().message) == 'Recall is ill-defined and ' 'being set to 0.0 due to no true samples.' ' Use `zero_division` parameter to control' ' this behavior.') @pytest.mark.parametrize('zero_division', ["warn", 0, 1]) def test_precision_warnings(zero_division): with warnings.catch_warnings(record=True) as record: warnings.simplefilter('always') precision_score(np.array([[1, 1], [1, 1]]), np.array([[0, 0], [0, 0]]), average='micro', zero_division=zero_division) if zero_division == "warn": assert (str(record.pop().message) == 'Precision is ill-defined and ' 'being set to 0.0 due to no predicted samples.' ' Use `zero_division` parameter to control' ' this behavior.') else: assert len(record) == 0 precision_score([0, 0], [0, 0]) if zero_division == "warn": assert (str(record.pop().message) == 'Precision is ill-defined and ' 'being set to 0.0 due to no predicted samples.' ' Use `zero_division` parameter to control' ' this behavior.') assert_no_warnings(precision_score, np.array([[0, 0], [0, 0]]), np.array([[1, 1], [1, 1]]), average='micro', zero_division=zero_division) @pytest.mark.parametrize('zero_division', ["warn", 0, 1]) def test_fscore_warnings(zero_division): with warnings.catch_warnings(record=True) as record: warnings.simplefilter('always') for score in [f1_score, partial(fbeta_score, beta=2)]: score(np.array([[1, 1], [1, 1]]), np.array([[0, 0], [0, 0]]), average='micro', zero_division=zero_division) assert len(record) == 0 score(np.array([[0, 0], [0, 0]]), np.array([[1, 1], [1, 1]]), average='micro', zero_division=zero_division) assert len(record) == 0 score(np.array([[0, 0], [0, 0]]), np.array([[0, 0], [0, 0]]), average='micro', zero_division=zero_division) if zero_division == "warn": assert (str(record.pop().message) == 'F-score is ill-defined and ' 'being set to 0.0 due to no true nor predicted ' 'samples. Use `zero_division` parameter to ' 'control this behavior.') else: assert len(record) == 0 def test_prf_average_binary_data_non_binary(): # Error if user does not explicitly set non-binary average mode y_true_mc = [1, 2, 3, 3] y_pred_mc = [1, 2, 3, 1] msg_mc = (r"Target is multiclass but average='binary'. Please " r"choose another average setting, one of \[" r"None, 'micro', 'macro', 'weighted'\].") y_true_ind = np.array([[0, 1, 1], [1, 0, 0], [0, 0, 1]]) y_pred_ind = np.array([[0, 1, 0], [1, 0, 0], [0, 0, 1]]) msg_ind = (r"Target is multilabel-indicator but average='binary'. Please " r"choose another average setting, one of \[" r"None, 'micro', 'macro', 'weighted', 'samples'\].") for y_true, y_pred, msg in [ (y_true_mc, y_pred_mc, msg_mc), (y_true_ind, y_pred_ind, msg_ind), ]: for metric in [precision_score, recall_score, f1_score, partial(fbeta_score, beta=2)]: with pytest.raises(ValueError, match=msg): metric(y_true, y_pred) def test__check_targets(): # Check that _check_targets correctly merges target types, squeezes # output and fails if input lengths differ. IND = 'multilabel-indicator' MC = 'multiclass' BIN = 'binary' CNT = 'continuous' MMC = 'multiclass-multioutput' MCN = 'continuous-multioutput' # all of length 3 EXAMPLES = [ (IND, np.array([[0, 1, 1], [1, 0, 0], [0, 0, 1]])), # must not be considered binary (IND, np.array([[0, 1], [1, 0], [1, 1]])), (MC, [2, 3, 1]), (BIN, [0, 1, 1]), (CNT, [0., 1.5, 1.]), (MC, np.array([[2], [3], [1]])), (BIN, np.array([[0], [1], [1]])), (CNT, np.array([[0.], [1.5], [1.]])), (MMC, np.array([[0, 2], [1, 3], [2, 3]])), (MCN, np.array([[0.5, 2.], [1.1, 3.], [2., 3.]])), ] # expected type given input types, or None for error # (types will be tried in either order) EXPECTED = { (IND, IND): IND, (MC, MC): MC, (BIN, BIN): BIN, (MC, IND): None, (BIN, IND): None, (BIN, MC): MC, # Disallowed types (CNT, CNT): None, (MMC, MMC): None, (MCN, MCN): None, (IND, CNT): None, (MC, CNT): None, (BIN, CNT): None, (MMC, CNT): None, (MCN, CNT): None, (IND, MMC): None, (MC, MMC): None, (BIN, MMC): None, (MCN, MMC): None, (IND, MCN): None, (MC, MCN): None, (BIN, MCN): None, } for (type1, y1), (type2, y2) in product(EXAMPLES, repeat=2): try: expected = EXPECTED[type1, type2] except KeyError: expected = EXPECTED[type2, type1] if expected is None: with pytest.raises(ValueError): _check_targets(y1, y2) if type1 != type2: err_msg = ("Classification metrics can't handle a mix " "of {0} and {1} targets".format(type1, type2)) with pytest.raises(ValueError, match=err_msg): _check_targets(y1, y2) else: if type1 not in (BIN, MC, IND): err_msg = "{0} is not supported".format(type1) with pytest.raises(ValueError, match=err_msg): _check_targets(y1, y2) else: merged_type, y1out, y2out = _check_targets(y1, y2) assert merged_type == expected if merged_type.startswith('multilabel'): assert y1out.format == 'csr' assert y2out.format == 'csr' else: assert_array_equal(y1out, np.squeeze(y1)) assert_array_equal(y2out, np.squeeze(y2)) with pytest.raises(ValueError): _check_targets(y1[:-1], y2) # Make sure seq of seq is not supported y1 = [(1, 2,), (0, 2, 3)] y2 = [(2,), (0, 2,)] msg = ('You appear to be using a legacy multi-label data representation. ' 'Sequence of sequences are no longer supported; use a binary array' ' or sparse matrix instead - the MultiLabelBinarizer' ' transformer can convert to this format.') with pytest.raises(ValueError, match=msg): _check_targets(y1, y2) def test__check_targets_multiclass_with_both_y_true_and_y_pred_binary(): # https://github.com/scikit-learn/scikit-learn/issues/8098 y_true = [0, 1] y_pred = [0, -1] assert _check_targets(y_true, y_pred)[0] == 'multiclass' def test_hinge_loss_binary(): y_true = np.array([-1, 1, 1, -1]) pred_decision = np.array([-8.5, 0.5, 1.5, -0.3]) assert hinge_loss(y_true, pred_decision) == 1.2 / 4 y_true = np.array([0, 2, 2, 0]) pred_decision = np.array([-8.5, 0.5, 1.5, -0.3]) assert hinge_loss(y_true, pred_decision) == 1.2 / 4 def test_hinge_loss_multiclass(): pred_decision = np.array([ [+0.36, -0.17, -0.58, -0.99], [-0.54, -0.37, -0.48, -0.58], [-1.45, -0.58, -0.38, -0.17], [-0.54, -0.38, -0.48, -0.58], [-2.36, -0.79, -0.27, +0.24], [-1.45, -0.58, -0.38, -0.17] ]) y_true = np.array([0, 1, 2, 1, 3, 2]) dummy_losses = np.array([ 1 - pred_decision[0][0] + pred_decision[0][1], 1 - pred_decision[1][1] + pred_decision[1][2], 1 - pred_decision[2][2] + pred_decision[2][3], 1 - pred_decision[3][1] + pred_decision[3][2], 1 - pred_decision[4][3] + pred_decision[4][2], 1 - pred_decision[5][2] + pred_decision[5][3] ]) np.clip(dummy_losses, 0, None, out=dummy_losses) dummy_hinge_loss = np.mean(dummy_losses) assert (hinge_loss(y_true, pred_decision) == dummy_hinge_loss) def test_hinge_loss_multiclass_missing_labels_with_labels_none(): y_true = np.array([0, 1, 2, 2]) pred_decision = np.array([ [+1.27, 0.034, -0.68, -1.40], [-1.45, -0.58, -0.38, -0.17], [-2.36, -0.79, -0.27, +0.24], [-2.36, -0.79, -0.27, +0.24] ]) error_message = ("Please include all labels in y_true " "or pass labels as third argument") with pytest.raises(ValueError, match=error_message): hinge_loss(y_true, pred_decision) def test_hinge_loss_multiclass_with_missing_labels(): pred_decision = np.array([ [+0.36, -0.17, -0.58, -0.99], [-0.55, -0.38, -0.48, -0.58], [-1.45, -0.58, -0.38, -0.17], [-0.55, -0.38, -0.48, -0.58], [-1.45, -0.58, -0.38, -0.17] ]) y_true = np.array([0, 1, 2, 1, 2]) labels = np.array([0, 1, 2, 3]) dummy_losses = np.array([ 1 - pred_decision[0][0] + pred_decision[0][1], 1 - pred_decision[1][1] + pred_decision[1][2], 1 - pred_decision[2][2] + pred_decision[2][3], 1 - pred_decision[3][1] + pred_decision[3][2], 1 - pred_decision[4][2] + pred_decision[4][3] ]) np.clip(dummy_losses, 0, None, out=dummy_losses) dummy_hinge_loss = np.mean(dummy_losses) assert (hinge_loss(y_true, pred_decision, labels=labels) == dummy_hinge_loss) def test_hinge_loss_multiclass_invariance_lists(): # Currently, invariance of string and integer labels cannot be tested # in common invariance tests because invariance tests for multiclass # decision functions is not implemented yet. y_true = ['blue', 'green', 'red', 'green', 'white', 'red'] pred_decision = [ [+0.36, -0.17, -0.58, -0.99], [-0.55, -0.38, -0.48, -0.58], [-1.45, -0.58, -0.38, -0.17], [-0.55, -0.38, -0.48, -0.58], [-2.36, -0.79, -0.27, +0.24], [-1.45, -0.58, -0.38, -0.17]] dummy_losses = np.array([ 1 - pred_decision[0][0] + pred_decision[0][1], 1 - pred_decision[1][1] + pred_decision[1][2], 1 - pred_decision[2][2] + pred_decision[2][3], 1 - pred_decision[3][1] + pred_decision[3][2], 1 - pred_decision[4][3] + pred_decision[4][2], 1 - pred_decision[5][2] + pred_decision[5][3] ]) np.clip(dummy_losses, 0, None, out=dummy_losses) dummy_hinge_loss = np.mean(dummy_losses) assert (hinge_loss(y_true, pred_decision) == dummy_hinge_loss) def test_log_loss(): # binary case with symbolic labels ("no" < "yes") y_true = ["no", "no", "no", "yes", "yes", "yes"] y_pred = np.array([[0.5, 0.5], [0.1, 0.9], [0.01, 0.99], [0.9, 0.1], [0.75, 0.25], [0.001, 0.999]]) loss = log_loss(y_true, y_pred) assert_almost_equal(loss, 1.8817971) # multiclass case; adapted from http://bit.ly/RJJHWA y_true = [1, 0, 2] y_pred = [[0.2, 0.7, 0.1], [0.6, 0.2, 0.2], [0.6, 0.1, 0.3]] loss = log_loss(y_true, y_pred, normalize=True) assert_almost_equal(loss, 0.6904911) # check that we got all the shapes and axes right # by doubling the length of y_true and y_pred y_true *= 2 y_pred *= 2 loss = log_loss(y_true, y_pred, normalize=False) assert_almost_equal(loss, 0.6904911 * 6, decimal=6) # check eps and handling of absolute zero and one probabilities y_pred = np.asarray(y_pred) > .5 loss = log_loss(y_true, y_pred, normalize=True, eps=.1) assert_almost_equal(loss, log_loss(y_true, np.clip(y_pred, .1, .9))) # raise error if number of classes are not equal. y_true = [1, 0, 2] y_pred = [[0.2, 0.7], [0.6, 0.5], [0.4, 0.1]] with pytest.raises(ValueError): log_loss(y_true, y_pred) # case when y_true is a string array object y_true = ["ham", "spam", "spam", "ham"] y_pred = [[0.2, 0.7], [0.6, 0.5], [0.4, 0.1], [0.7, 0.2]] loss = log_loss(y_true, y_pred) assert_almost_equal(loss, 1.0383217, decimal=6) # test labels option y_true = [2, 2] y_pred = [[0.2, 0.7], [0.6, 0.5]] y_score = np.array([[0.1, 0.9], [0.1, 0.9]]) error_str = (r'y_true contains only one label \(2\). Please provide ' r'the true labels explicitly through the labels argument.') with pytest.raises(ValueError, match=error_str): log_loss(y_true, y_pred) y_pred = [[0.2, 0.7], [0.6, 0.5], [0.2, 0.3]] error_str = ('Found input variables with inconsistent numbers of samples: ' '[3, 2]') (ValueError, error_str, log_loss, y_true, y_pred) # works when the labels argument is used true_log_loss = -np.mean(np.log(y_score[:, 1])) calculated_log_loss = log_loss(y_true, y_score, labels=[1, 2]) assert_almost_equal(calculated_log_loss, true_log_loss) # ensure labels work when len(np.unique(y_true)) != y_pred.shape[1] y_true = [1, 2, 2] y_score2 = [[0.2, 0.7, 0.3], [0.6, 0.5, 0.3], [0.3, 0.9, 0.1]] loss = log_loss(y_true, y_score2, labels=[1, 2, 3]) assert_almost_equal(loss, 1.0630345, decimal=6) def test_log_loss_pandas_input(): # case when input is a pandas series and dataframe gh-5715 y_tr = np.array(["ham", "spam", "spam", "ham"]) y_pr = np.array([[0.2, 0.7], [0.6, 0.5], [0.4, 0.1], [0.7, 0.2]]) types = [(MockDataFrame, MockDataFrame)] try: from pandas import Series, DataFrame types.append((Series, DataFrame)) except ImportError: pass for TrueInputType, PredInputType in types: # y_pred dataframe, y_true series y_true, y_pred = TrueInputType(y_tr), PredInputType(y_pr) loss = log_loss(y_true, y_pred) assert_almost_equal(loss, 1.0383217, decimal=6) def test_brier_score_loss(): # Check brier_score_loss function y_true = np.array([0, 1, 1, 0, 1, 1]) y_pred = np.array([0.1, 0.8, 0.9, 0.3, 1., 0.95]) true_score = linalg.norm(y_true - y_pred) ** 2 / len(y_true) assert_almost_equal(brier_score_loss(y_true, y_true), 0.0) assert_almost_equal(brier_score_loss(y_true, y_pred), true_score) assert_almost_equal(brier_score_loss(1. + y_true, y_pred), true_score) assert_almost_equal(brier_score_loss(2 * y_true - 1, y_pred), true_score) with pytest.raises(ValueError): brier_score_loss(y_true, y_pred[1:]) with pytest.raises(ValueError): brier_score_loss(y_true, y_pred + 1.) with pytest.raises(ValueError): brier_score_loss(y_true, y_pred - 1.) # ensure to raise an error for multiclass y_true y_true = np.array([0, 1, 2, 0]) y_pred = np.array([0.8, 0.6, 0.4, 0.2]) error_message = ("Only binary classification is supported. Labels " "in y_true: {}".format(np.array([0, 1, 2]))) with pytest.raises(ValueError, match=re.escape(error_message)): brier_score_loss(y_true, y_pred) # calculate correctly when there's only one class in y_true assert_almost_equal(brier_score_loss([-1], [0.4]), 0.16) assert_almost_equal(brier_score_loss([0], [0.4]), 0.16) assert_almost_equal(brier_score_loss([1], [0.4]), 0.36) assert_almost_equal( brier_score_loss(['foo'], [0.4], pos_label='bar'), 0.16) assert_almost_equal( brier_score_loss(['foo'], [0.4], pos_label='foo'), 0.36) def test_balanced_accuracy_score_unseen(): assert_warns_message(UserWarning, 'y_pred contains classes not in y_true', balanced_accuracy_score, [0, 0, 0], [0, 0, 1]) @pytest.mark.parametrize('y_true,y_pred', [ (['a', 'b', 'a', 'b'], ['a', 'a', 'a', 'b']), (['a', 'b', 'c', 'b'], ['a', 'a', 'a', 'b']), (['a', 'a', 'a', 'b'], ['a', 'b', 'c', 'b']), ]) def test_balanced_accuracy_score(y_true, y_pred): macro_recall = recall_score(y_true, y_pred, average='macro', labels=np.unique(y_true)) with ignore_warnings(): # Warnings are tested in test_balanced_accuracy_score_unseen balanced = balanced_accuracy_score(y_true, y_pred) assert balanced == pytest.approx(macro_recall) adjusted = balanced_accuracy_score(y_true, y_pred, adjusted=True) chance = balanced_accuracy_score(y_true, np.full_like(y_true, y_true[0])) assert adjusted == (balanced - chance) / (1 - chance)