1057 lines
38 KiB
Python
1057 lines
38 KiB
Python
# Authors: Alexandre Gramfort <alexandre.gramfort@telecom-paristech.fr>
|
|
# License: BSD 3 clause
|
|
|
|
import pytest
|
|
import numpy as np
|
|
from numpy.testing import assert_allclose
|
|
from scipy import sparse
|
|
|
|
from sklearn.base import BaseEstimator, clone
|
|
from sklearn.dummy import DummyClassifier
|
|
from sklearn.model_selection import LeaveOneOut, train_test_split
|
|
|
|
from sklearn.utils._testing import (
|
|
assert_array_almost_equal,
|
|
assert_almost_equal,
|
|
assert_array_equal,
|
|
)
|
|
from sklearn.utils.extmath import softmax
|
|
from sklearn.exceptions import NotFittedError
|
|
from sklearn.datasets import make_classification, make_blobs, load_iris
|
|
from sklearn.preprocessing import LabelEncoder
|
|
from sklearn.model_selection import KFold, cross_val_predict
|
|
from sklearn.naive_bayes import MultinomialNB
|
|
from sklearn.ensemble import (
|
|
RandomForestClassifier,
|
|
VotingClassifier,
|
|
)
|
|
from sklearn.linear_model import LogisticRegression, LinearRegression
|
|
from sklearn.tree import DecisionTreeClassifier
|
|
from sklearn.svm import LinearSVC
|
|
from sklearn.pipeline import Pipeline, make_pipeline
|
|
from sklearn.preprocessing import StandardScaler
|
|
from sklearn.isotonic import IsotonicRegression
|
|
from sklearn.feature_extraction import DictVectorizer
|
|
from sklearn.impute import SimpleImputer
|
|
from sklearn.metrics import brier_score_loss
|
|
from sklearn.calibration import (
|
|
_CalibratedClassifier,
|
|
_SigmoidCalibration,
|
|
_sigmoid_calibration,
|
|
CalibratedClassifierCV,
|
|
CalibrationDisplay,
|
|
calibration_curve,
|
|
)
|
|
from sklearn.utils._mocking import CheckingClassifier
|
|
from sklearn.utils._testing import _convert_container
|
|
|
|
|
|
N_SAMPLES = 200
|
|
|
|
|
|
@pytest.fixture(scope="module")
|
|
def data():
|
|
X, y = make_classification(n_samples=N_SAMPLES, n_features=6, random_state=42)
|
|
return X, y
|
|
|
|
|
|
@pytest.mark.parametrize("method", ["sigmoid", "isotonic"])
|
|
@pytest.mark.parametrize("ensemble", [True, False])
|
|
def test_calibration(data, method, ensemble):
|
|
# Test calibration objects with isotonic and sigmoid
|
|
n_samples = N_SAMPLES // 2
|
|
X, y = data
|
|
sample_weight = np.random.RandomState(seed=42).uniform(size=y.size)
|
|
|
|
X -= X.min() # MultinomialNB only allows positive X
|
|
|
|
# split train and test
|
|
X_train, y_train, sw_train = X[:n_samples], y[:n_samples], sample_weight[:n_samples]
|
|
X_test, y_test = X[n_samples:], y[n_samples:]
|
|
|
|
# Naive-Bayes
|
|
clf = MultinomialNB(force_alpha=True).fit(X_train, y_train, sample_weight=sw_train)
|
|
prob_pos_clf = clf.predict_proba(X_test)[:, 1]
|
|
|
|
cal_clf = CalibratedClassifierCV(clf, cv=y.size + 1, ensemble=ensemble)
|
|
with pytest.raises(ValueError):
|
|
cal_clf.fit(X, y)
|
|
|
|
# Naive Bayes with calibration
|
|
for this_X_train, this_X_test in [
|
|
(X_train, X_test),
|
|
(sparse.csr_matrix(X_train), sparse.csr_matrix(X_test)),
|
|
]:
|
|
cal_clf = CalibratedClassifierCV(clf, method=method, cv=5, ensemble=ensemble)
|
|
# Note that this fit overwrites the fit on the entire training
|
|
# set
|
|
cal_clf.fit(this_X_train, y_train, sample_weight=sw_train)
|
|
prob_pos_cal_clf = cal_clf.predict_proba(this_X_test)[:, 1]
|
|
|
|
# Check that brier score has improved after calibration
|
|
assert brier_score_loss(y_test, prob_pos_clf) > brier_score_loss(
|
|
y_test, prob_pos_cal_clf
|
|
)
|
|
|
|
# Check invariance against relabeling [0, 1] -> [1, 2]
|
|
cal_clf.fit(this_X_train, y_train + 1, sample_weight=sw_train)
|
|
prob_pos_cal_clf_relabeled = cal_clf.predict_proba(this_X_test)[:, 1]
|
|
assert_array_almost_equal(prob_pos_cal_clf, prob_pos_cal_clf_relabeled)
|
|
|
|
# Check invariance against relabeling [0, 1] -> [-1, 1]
|
|
cal_clf.fit(this_X_train, 2 * y_train - 1, sample_weight=sw_train)
|
|
prob_pos_cal_clf_relabeled = cal_clf.predict_proba(this_X_test)[:, 1]
|
|
assert_array_almost_equal(prob_pos_cal_clf, prob_pos_cal_clf_relabeled)
|
|
|
|
# Check invariance against relabeling [0, 1] -> [1, 0]
|
|
cal_clf.fit(this_X_train, (y_train + 1) % 2, sample_weight=sw_train)
|
|
prob_pos_cal_clf_relabeled = cal_clf.predict_proba(this_X_test)[:, 1]
|
|
if method == "sigmoid":
|
|
assert_array_almost_equal(prob_pos_cal_clf, 1 - prob_pos_cal_clf_relabeled)
|
|
else:
|
|
# Isotonic calibration is not invariant against relabeling
|
|
# but should improve in both cases
|
|
assert brier_score_loss(y_test, prob_pos_clf) > brier_score_loss(
|
|
(y_test + 1) % 2, prob_pos_cal_clf_relabeled
|
|
)
|
|
|
|
|
|
def test_calibration_default_estimator(data):
|
|
# Check estimator default is LinearSVC
|
|
X, y = data
|
|
calib_clf = CalibratedClassifierCV(cv=2)
|
|
calib_clf.fit(X, y)
|
|
|
|
base_est = calib_clf.calibrated_classifiers_[0].estimator
|
|
assert isinstance(base_est, LinearSVC)
|
|
|
|
|
|
@pytest.mark.parametrize("ensemble", [True, False])
|
|
def test_calibration_cv_splitter(data, ensemble):
|
|
# Check when `cv` is a CV splitter
|
|
X, y = data
|
|
|
|
splits = 5
|
|
kfold = KFold(n_splits=splits)
|
|
calib_clf = CalibratedClassifierCV(cv=kfold, ensemble=ensemble)
|
|
assert isinstance(calib_clf.cv, KFold)
|
|
assert calib_clf.cv.n_splits == splits
|
|
|
|
calib_clf.fit(X, y)
|
|
expected_n_clf = splits if ensemble else 1
|
|
assert len(calib_clf.calibrated_classifiers_) == expected_n_clf
|
|
|
|
|
|
@pytest.mark.parametrize("method", ["sigmoid", "isotonic"])
|
|
@pytest.mark.parametrize("ensemble", [True, False])
|
|
def test_sample_weight(data, method, ensemble):
|
|
n_samples = N_SAMPLES // 2
|
|
X, y = data
|
|
|
|
sample_weight = np.random.RandomState(seed=42).uniform(size=len(y))
|
|
X_train, y_train, sw_train = X[:n_samples], y[:n_samples], sample_weight[:n_samples]
|
|
X_test = X[n_samples:]
|
|
|
|
estimator = LinearSVC(random_state=42)
|
|
calibrated_clf = CalibratedClassifierCV(estimator, method=method, ensemble=ensemble)
|
|
calibrated_clf.fit(X_train, y_train, sample_weight=sw_train)
|
|
probs_with_sw = calibrated_clf.predict_proba(X_test)
|
|
|
|
# As the weights are used for the calibration, they should still yield
|
|
# different predictions
|
|
calibrated_clf.fit(X_train, y_train)
|
|
probs_without_sw = calibrated_clf.predict_proba(X_test)
|
|
|
|
diff = np.linalg.norm(probs_with_sw - probs_without_sw)
|
|
assert diff > 0.1
|
|
|
|
|
|
@pytest.mark.parametrize("method", ["sigmoid", "isotonic"])
|
|
@pytest.mark.parametrize("ensemble", [True, False])
|
|
def test_parallel_execution(data, method, ensemble):
|
|
"""Test parallel calibration"""
|
|
X, y = data
|
|
X_train, X_test, y_train, y_test = train_test_split(X, y, random_state=42)
|
|
|
|
estimator = LinearSVC(random_state=42)
|
|
|
|
cal_clf_parallel = CalibratedClassifierCV(
|
|
estimator, method=method, n_jobs=2, ensemble=ensemble
|
|
)
|
|
cal_clf_parallel.fit(X_train, y_train)
|
|
probs_parallel = cal_clf_parallel.predict_proba(X_test)
|
|
|
|
cal_clf_sequential = CalibratedClassifierCV(
|
|
estimator, method=method, n_jobs=1, ensemble=ensemble
|
|
)
|
|
cal_clf_sequential.fit(X_train, y_train)
|
|
probs_sequential = cal_clf_sequential.predict_proba(X_test)
|
|
|
|
assert_allclose(probs_parallel, probs_sequential)
|
|
|
|
|
|
@pytest.mark.parametrize("method", ["sigmoid", "isotonic"])
|
|
@pytest.mark.parametrize("ensemble", [True, False])
|
|
# increase the number of RNG seeds to assess the statistical stability of this
|
|
# test:
|
|
@pytest.mark.parametrize("seed", range(2))
|
|
def test_calibration_multiclass(method, ensemble, seed):
|
|
def multiclass_brier(y_true, proba_pred, n_classes):
|
|
Y_onehot = np.eye(n_classes)[y_true]
|
|
return np.sum((Y_onehot - proba_pred) ** 2) / Y_onehot.shape[0]
|
|
|
|
# Test calibration for multiclass with classifier that implements
|
|
# only decision function.
|
|
clf = LinearSVC(random_state=7)
|
|
X, y = make_blobs(
|
|
n_samples=500, n_features=100, random_state=seed, centers=10, cluster_std=15.0
|
|
)
|
|
|
|
# Use an unbalanced dataset by collapsing 8 clusters into one class
|
|
# to make the naive calibration based on a softmax more unlikely
|
|
# to work.
|
|
y[y > 2] = 2
|
|
n_classes = np.unique(y).shape[0]
|
|
X_train, y_train = X[::2], y[::2]
|
|
X_test, y_test = X[1::2], y[1::2]
|
|
|
|
clf.fit(X_train, y_train)
|
|
|
|
cal_clf = CalibratedClassifierCV(clf, method=method, cv=5, ensemble=ensemble)
|
|
cal_clf.fit(X_train, y_train)
|
|
probas = cal_clf.predict_proba(X_test)
|
|
# Check probabilities sum to 1
|
|
assert_allclose(np.sum(probas, axis=1), np.ones(len(X_test)))
|
|
|
|
# Check that the dataset is not too trivial, otherwise it's hard
|
|
# to get interesting calibration data during the internal
|
|
# cross-validation loop.
|
|
assert 0.65 < clf.score(X_test, y_test) < 0.95
|
|
|
|
# Check that the accuracy of the calibrated model is never degraded
|
|
# too much compared to the original classifier.
|
|
assert cal_clf.score(X_test, y_test) > 0.95 * clf.score(X_test, y_test)
|
|
|
|
# Check that Brier loss of calibrated classifier is smaller than
|
|
# loss obtained by naively turning OvR decision function to
|
|
# probabilities via a softmax
|
|
uncalibrated_brier = multiclass_brier(
|
|
y_test, softmax(clf.decision_function(X_test)), n_classes=n_classes
|
|
)
|
|
calibrated_brier = multiclass_brier(y_test, probas, n_classes=n_classes)
|
|
|
|
assert calibrated_brier < 1.1 * uncalibrated_brier
|
|
|
|
# Test that calibration of a multiclass classifier decreases log-loss
|
|
# for RandomForestClassifier
|
|
clf = RandomForestClassifier(n_estimators=30, random_state=42)
|
|
clf.fit(X_train, y_train)
|
|
clf_probs = clf.predict_proba(X_test)
|
|
uncalibrated_brier = multiclass_brier(y_test, clf_probs, n_classes=n_classes)
|
|
|
|
cal_clf = CalibratedClassifierCV(clf, method=method, cv=5, ensemble=ensemble)
|
|
cal_clf.fit(X_train, y_train)
|
|
cal_clf_probs = cal_clf.predict_proba(X_test)
|
|
calibrated_brier = multiclass_brier(y_test, cal_clf_probs, n_classes=n_classes)
|
|
assert calibrated_brier < 1.1 * uncalibrated_brier
|
|
|
|
|
|
def test_calibration_zero_probability():
|
|
# Test an edge case where _CalibratedClassifier avoids numerical errors
|
|
# in the multiclass normalization step if all the calibrators output
|
|
# are zero all at once for a given sample and instead fallback to uniform
|
|
# probabilities.
|
|
class ZeroCalibrator:
|
|
# This function is called from _CalibratedClassifier.predict_proba.
|
|
def predict(self, X):
|
|
return np.zeros(X.shape[0])
|
|
|
|
X, y = make_blobs(
|
|
n_samples=50, n_features=10, random_state=7, centers=10, cluster_std=15.0
|
|
)
|
|
clf = DummyClassifier().fit(X, y)
|
|
calibrator = ZeroCalibrator()
|
|
cal_clf = _CalibratedClassifier(
|
|
estimator=clf, calibrators=[calibrator], classes=clf.classes_
|
|
)
|
|
|
|
probas = cal_clf.predict_proba(X)
|
|
|
|
# Check that all probabilities are uniformly 1. / clf.n_classes_
|
|
assert_allclose(probas, 1.0 / clf.n_classes_)
|
|
|
|
|
|
def test_calibration_prefit():
|
|
"""Test calibration for prefitted classifiers"""
|
|
n_samples = 50
|
|
X, y = make_classification(n_samples=3 * n_samples, n_features=6, random_state=42)
|
|
sample_weight = np.random.RandomState(seed=42).uniform(size=y.size)
|
|
|
|
X -= X.min() # MultinomialNB only allows positive X
|
|
|
|
# split train and test
|
|
X_train, y_train, sw_train = X[:n_samples], y[:n_samples], sample_weight[:n_samples]
|
|
X_calib, y_calib, sw_calib = (
|
|
X[n_samples : 2 * n_samples],
|
|
y[n_samples : 2 * n_samples],
|
|
sample_weight[n_samples : 2 * n_samples],
|
|
)
|
|
X_test, y_test = X[2 * n_samples :], y[2 * n_samples :]
|
|
|
|
# Naive-Bayes
|
|
clf = MultinomialNB(force_alpha=True)
|
|
# Check error if clf not prefit
|
|
unfit_clf = CalibratedClassifierCV(clf, cv="prefit")
|
|
with pytest.raises(NotFittedError):
|
|
unfit_clf.fit(X_calib, y_calib)
|
|
|
|
clf.fit(X_train, y_train, sw_train)
|
|
prob_pos_clf = clf.predict_proba(X_test)[:, 1]
|
|
|
|
# Naive Bayes with calibration
|
|
for this_X_calib, this_X_test in [
|
|
(X_calib, X_test),
|
|
(sparse.csr_matrix(X_calib), sparse.csr_matrix(X_test)),
|
|
]:
|
|
for method in ["isotonic", "sigmoid"]:
|
|
cal_clf = CalibratedClassifierCV(clf, method=method, cv="prefit")
|
|
|
|
for sw in [sw_calib, None]:
|
|
cal_clf.fit(this_X_calib, y_calib, sample_weight=sw)
|
|
y_prob = cal_clf.predict_proba(this_X_test)
|
|
y_pred = cal_clf.predict(this_X_test)
|
|
prob_pos_cal_clf = y_prob[:, 1]
|
|
assert_array_equal(y_pred, np.array([0, 1])[np.argmax(y_prob, axis=1)])
|
|
|
|
assert brier_score_loss(y_test, prob_pos_clf) > brier_score_loss(
|
|
y_test, prob_pos_cal_clf
|
|
)
|
|
|
|
|
|
@pytest.mark.parametrize("method", ["sigmoid", "isotonic"])
|
|
def test_calibration_ensemble_false(data, method):
|
|
# Test that `ensemble=False` is the same as using predictions from
|
|
# `cross_val_predict` to train calibrator.
|
|
X, y = data
|
|
clf = LinearSVC(random_state=7)
|
|
|
|
cal_clf = CalibratedClassifierCV(clf, method=method, cv=3, ensemble=False)
|
|
cal_clf.fit(X, y)
|
|
cal_probas = cal_clf.predict_proba(X)
|
|
|
|
# Get probas manually
|
|
unbiased_preds = cross_val_predict(clf, X, y, cv=3, method="decision_function")
|
|
if method == "isotonic":
|
|
calibrator = IsotonicRegression(out_of_bounds="clip")
|
|
else:
|
|
calibrator = _SigmoidCalibration()
|
|
calibrator.fit(unbiased_preds, y)
|
|
# Use `clf` fit on all data
|
|
clf.fit(X, y)
|
|
clf_df = clf.decision_function(X)
|
|
manual_probas = calibrator.predict(clf_df)
|
|
assert_allclose(cal_probas[:, 1], manual_probas)
|
|
|
|
|
|
def test_sigmoid_calibration():
|
|
"""Test calibration values with Platt sigmoid model"""
|
|
exF = np.array([5, -4, 1.0])
|
|
exY = np.array([1, -1, -1])
|
|
# computed from my python port of the C++ code in LibSVM
|
|
AB_lin_libsvm = np.array([-0.20261354391187855, 0.65236314980010512])
|
|
assert_array_almost_equal(AB_lin_libsvm, _sigmoid_calibration(exF, exY), 3)
|
|
lin_prob = 1.0 / (1.0 + np.exp(AB_lin_libsvm[0] * exF + AB_lin_libsvm[1]))
|
|
sk_prob = _SigmoidCalibration().fit(exF, exY).predict(exF)
|
|
assert_array_almost_equal(lin_prob, sk_prob, 6)
|
|
|
|
# check that _SigmoidCalibration().fit only accepts 1d array or 2d column
|
|
# arrays
|
|
with pytest.raises(ValueError):
|
|
_SigmoidCalibration().fit(np.vstack((exF, exF)), exY)
|
|
|
|
|
|
def test_calibration_curve():
|
|
"""Check calibration_curve function"""
|
|
y_true = np.array([0, 0, 0, 1, 1, 1])
|
|
y_pred = np.array([0.0, 0.1, 0.2, 0.8, 0.9, 1.0])
|
|
prob_true, prob_pred = calibration_curve(y_true, y_pred, n_bins=2)
|
|
assert len(prob_true) == len(prob_pred)
|
|
assert len(prob_true) == 2
|
|
assert_almost_equal(prob_true, [0, 1])
|
|
assert_almost_equal(prob_pred, [0.1, 0.9])
|
|
|
|
# Probabilities outside [0, 1] should not be accepted at all.
|
|
with pytest.raises(ValueError):
|
|
calibration_curve([1], [-0.1])
|
|
|
|
# test that quantiles work as expected
|
|
y_true2 = np.array([0, 0, 0, 0, 1, 1])
|
|
y_pred2 = np.array([0.0, 0.1, 0.2, 0.5, 0.9, 1.0])
|
|
prob_true_quantile, prob_pred_quantile = calibration_curve(
|
|
y_true2, y_pred2, n_bins=2, strategy="quantile"
|
|
)
|
|
|
|
assert len(prob_true_quantile) == len(prob_pred_quantile)
|
|
assert len(prob_true_quantile) == 2
|
|
assert_almost_equal(prob_true_quantile, [0, 2 / 3])
|
|
assert_almost_equal(prob_pred_quantile, [0.1, 0.8])
|
|
|
|
# Check that error is raised when invalid strategy is selected
|
|
with pytest.raises(ValueError):
|
|
calibration_curve(y_true2, y_pred2, strategy="percentile")
|
|
|
|
|
|
# TODO(1.3): Remove this test.
|
|
def test_calibration_curve_with_unnormalized_proba():
|
|
"""Tests the `normalize` parameter of `calibration_curve`"""
|
|
y_true = np.array([0, 0, 0, 1, 1, 1])
|
|
y_pred = np.array([0.0, 0.1, 0.2, 0.8, 0.9, 1.0])
|
|
|
|
# Ensure `normalize` == False raises a FutureWarning.
|
|
with pytest.warns(FutureWarning):
|
|
calibration_curve(y_true, y_pred, n_bins=2, normalize=False)
|
|
|
|
# Ensure `normalize` == True raises a FutureWarning and behaves as expected.
|
|
with pytest.warns(FutureWarning):
|
|
prob_true_unnormalized, prob_pred_unnormalized = calibration_curve(
|
|
y_true, y_pred * 2, n_bins=2, normalize=True
|
|
)
|
|
prob_true, prob_pred = calibration_curve(y_true, y_pred, n_bins=2)
|
|
assert_almost_equal(prob_true, prob_true_unnormalized)
|
|
assert_almost_equal(prob_pred, prob_pred_unnormalized)
|
|
|
|
|
|
@pytest.mark.parametrize("ensemble", [True, False])
|
|
def test_calibration_nan_imputer(ensemble):
|
|
"""Test that calibration can accept nan"""
|
|
X, y = make_classification(
|
|
n_samples=10, n_features=2, n_informative=2, n_redundant=0, random_state=42
|
|
)
|
|
X[0, 0] = np.nan
|
|
clf = Pipeline(
|
|
[("imputer", SimpleImputer()), ("rf", RandomForestClassifier(n_estimators=1))]
|
|
)
|
|
clf_c = CalibratedClassifierCV(clf, cv=2, method="isotonic", ensemble=ensemble)
|
|
clf_c.fit(X, y)
|
|
clf_c.predict(X)
|
|
|
|
|
|
@pytest.mark.parametrize("ensemble", [True, False])
|
|
def test_calibration_prob_sum(ensemble):
|
|
# Test that sum of probabilities is 1. A non-regression test for
|
|
# issue #7796
|
|
num_classes = 2
|
|
X, y = make_classification(n_samples=10, n_features=5, n_classes=num_classes)
|
|
clf = LinearSVC(C=1.0, random_state=7)
|
|
clf_prob = CalibratedClassifierCV(
|
|
clf, method="sigmoid", cv=LeaveOneOut(), ensemble=ensemble
|
|
)
|
|
clf_prob.fit(X, y)
|
|
|
|
probs = clf_prob.predict_proba(X)
|
|
assert_array_almost_equal(probs.sum(axis=1), np.ones(probs.shape[0]))
|
|
|
|
|
|
@pytest.mark.parametrize("ensemble", [True, False])
|
|
def test_calibration_less_classes(ensemble):
|
|
# Test to check calibration works fine when train set in a test-train
|
|
# split does not contain all classes
|
|
# Since this test uses LOO, at each iteration train set will not contain a
|
|
# class label
|
|
X = np.random.randn(10, 5)
|
|
y = np.arange(10)
|
|
clf = LinearSVC(C=1.0, random_state=7)
|
|
cal_clf = CalibratedClassifierCV(
|
|
clf, method="sigmoid", cv=LeaveOneOut(), ensemble=ensemble
|
|
)
|
|
cal_clf.fit(X, y)
|
|
|
|
for i, calibrated_classifier in enumerate(cal_clf.calibrated_classifiers_):
|
|
proba = calibrated_classifier.predict_proba(X)
|
|
if ensemble:
|
|
# Check that the unobserved class has proba=0
|
|
assert_array_equal(proba[:, i], np.zeros(len(y)))
|
|
# Check for all other classes proba>0
|
|
assert np.all(proba[:, :i] > 0)
|
|
assert np.all(proba[:, i + 1 :] > 0)
|
|
else:
|
|
# Check `proba` are all 1/n_classes
|
|
assert np.allclose(proba, 1 / proba.shape[0])
|
|
|
|
|
|
@pytest.mark.parametrize(
|
|
"X",
|
|
[
|
|
np.random.RandomState(42).randn(15, 5, 2),
|
|
np.random.RandomState(42).randn(15, 5, 2, 6),
|
|
],
|
|
)
|
|
def test_calibration_accepts_ndarray(X):
|
|
"""Test that calibration accepts n-dimensional arrays as input"""
|
|
y = [1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0]
|
|
|
|
class MockTensorClassifier(BaseEstimator):
|
|
"""A toy estimator that accepts tensor inputs"""
|
|
|
|
def fit(self, X, y):
|
|
self.classes_ = np.unique(y)
|
|
return self
|
|
|
|
def decision_function(self, X):
|
|
# toy decision function that just needs to have the right shape:
|
|
return X.reshape(X.shape[0], -1).sum(axis=1)
|
|
|
|
calibrated_clf = CalibratedClassifierCV(MockTensorClassifier())
|
|
# we should be able to fit this classifier with no error
|
|
calibrated_clf.fit(X, y)
|
|
|
|
|
|
@pytest.fixture
|
|
def dict_data():
|
|
dict_data = [
|
|
{"state": "NY", "age": "adult"},
|
|
{"state": "TX", "age": "adult"},
|
|
{"state": "VT", "age": "child"},
|
|
]
|
|
text_labels = [1, 0, 1]
|
|
return dict_data, text_labels
|
|
|
|
|
|
@pytest.fixture
|
|
def dict_data_pipeline(dict_data):
|
|
X, y = dict_data
|
|
pipeline_prefit = Pipeline(
|
|
[("vectorizer", DictVectorizer()), ("clf", RandomForestClassifier())]
|
|
)
|
|
return pipeline_prefit.fit(X, y)
|
|
|
|
|
|
def test_calibration_dict_pipeline(dict_data, dict_data_pipeline):
|
|
"""Test that calibration works in prefit pipeline with transformer
|
|
|
|
`X` is not array-like, sparse matrix or dataframe at the start.
|
|
See https://github.com/scikit-learn/scikit-learn/issues/8710
|
|
|
|
Also test it can predict without running into validation errors.
|
|
See https://github.com/scikit-learn/scikit-learn/issues/19637
|
|
"""
|
|
X, y = dict_data
|
|
clf = dict_data_pipeline
|
|
calib_clf = CalibratedClassifierCV(clf, cv="prefit")
|
|
calib_clf.fit(X, y)
|
|
# Check attributes are obtained from fitted estimator
|
|
assert_array_equal(calib_clf.classes_, clf.classes_)
|
|
|
|
# Neither the pipeline nor the calibration meta-estimator
|
|
# expose the n_features_in_ check on this kind of data.
|
|
assert not hasattr(clf, "n_features_in_")
|
|
assert not hasattr(calib_clf, "n_features_in_")
|
|
|
|
# Ensure that no error is thrown with predict and predict_proba
|
|
calib_clf.predict(X)
|
|
calib_clf.predict_proba(X)
|
|
|
|
|
|
@pytest.mark.parametrize(
|
|
"clf, cv",
|
|
[
|
|
pytest.param(LinearSVC(C=1), 2),
|
|
pytest.param(LinearSVC(C=1), "prefit"),
|
|
],
|
|
)
|
|
def test_calibration_attributes(clf, cv):
|
|
# Check that `n_features_in_` and `classes_` attributes created properly
|
|
X, y = make_classification(n_samples=10, n_features=5, n_classes=2, random_state=7)
|
|
if cv == "prefit":
|
|
clf = clf.fit(X, y)
|
|
calib_clf = CalibratedClassifierCV(clf, cv=cv)
|
|
calib_clf.fit(X, y)
|
|
|
|
if cv == "prefit":
|
|
assert_array_equal(calib_clf.classes_, clf.classes_)
|
|
assert calib_clf.n_features_in_ == clf.n_features_in_
|
|
else:
|
|
classes = LabelEncoder().fit(y).classes_
|
|
assert_array_equal(calib_clf.classes_, classes)
|
|
assert calib_clf.n_features_in_ == X.shape[1]
|
|
|
|
|
|
def test_calibration_inconsistent_prefit_n_features_in():
|
|
# Check that `n_features_in_` from prefit base estimator
|
|
# is consistent with training set
|
|
X, y = make_classification(n_samples=10, n_features=5, n_classes=2, random_state=7)
|
|
clf = LinearSVC(C=1).fit(X, y)
|
|
calib_clf = CalibratedClassifierCV(clf, cv="prefit")
|
|
|
|
msg = "X has 3 features, but LinearSVC is expecting 5 features as input."
|
|
with pytest.raises(ValueError, match=msg):
|
|
calib_clf.fit(X[:, :3], y)
|
|
|
|
|
|
def test_calibration_votingclassifier():
|
|
# Check that `CalibratedClassifier` works with `VotingClassifier`.
|
|
# The method `predict_proba` from `VotingClassifier` is dynamically
|
|
# defined via a property that only works when voting="soft".
|
|
X, y = make_classification(n_samples=10, n_features=5, n_classes=2, random_state=7)
|
|
vote = VotingClassifier(
|
|
estimators=[("lr" + str(i), LogisticRegression()) for i in range(3)],
|
|
voting="soft",
|
|
)
|
|
vote.fit(X, y)
|
|
|
|
calib_clf = CalibratedClassifierCV(estimator=vote, cv="prefit")
|
|
# smoke test: should not raise an error
|
|
calib_clf.fit(X, y)
|
|
|
|
|
|
@pytest.fixture(scope="module")
|
|
def iris_data():
|
|
return load_iris(return_X_y=True)
|
|
|
|
|
|
@pytest.fixture(scope="module")
|
|
def iris_data_binary(iris_data):
|
|
X, y = iris_data
|
|
return X[y < 2], y[y < 2]
|
|
|
|
|
|
def test_calibration_display_validation(pyplot, iris_data, iris_data_binary):
|
|
X, y = iris_data
|
|
X_binary, y_binary = iris_data_binary
|
|
|
|
reg = LinearRegression().fit(X, y)
|
|
msg = "'estimator' should be a fitted classifier"
|
|
with pytest.raises(ValueError, match=msg):
|
|
CalibrationDisplay.from_estimator(reg, X, y)
|
|
|
|
clf = LinearSVC().fit(X, y)
|
|
msg = "response method predict_proba is not defined in"
|
|
with pytest.raises(ValueError, match=msg):
|
|
CalibrationDisplay.from_estimator(clf, X, y)
|
|
|
|
clf = LogisticRegression()
|
|
with pytest.raises(NotFittedError):
|
|
CalibrationDisplay.from_estimator(clf, X, y)
|
|
|
|
|
|
@pytest.mark.parametrize("constructor_name", ["from_estimator", "from_predictions"])
|
|
def test_calibration_display_non_binary(pyplot, iris_data, constructor_name):
|
|
X, y = iris_data
|
|
clf = DecisionTreeClassifier()
|
|
clf.fit(X, y)
|
|
y_prob = clf.predict_proba(X)
|
|
|
|
if constructor_name == "from_estimator":
|
|
msg = "to be a binary classifier, but got"
|
|
with pytest.raises(ValueError, match=msg):
|
|
CalibrationDisplay.from_estimator(clf, X, y)
|
|
else:
|
|
msg = "y should be a 1d array, got an array of shape"
|
|
with pytest.raises(ValueError, match=msg):
|
|
CalibrationDisplay.from_predictions(y, y_prob)
|
|
|
|
|
|
@pytest.mark.parametrize("n_bins", [5, 10])
|
|
@pytest.mark.parametrize("strategy", ["uniform", "quantile"])
|
|
def test_calibration_display_compute(pyplot, iris_data_binary, n_bins, strategy):
|
|
# Ensure `CalibrationDisplay.from_predictions` and `calibration_curve`
|
|
# compute the same results. Also checks attributes of the
|
|
# CalibrationDisplay object.
|
|
X, y = iris_data_binary
|
|
|
|
lr = LogisticRegression().fit(X, y)
|
|
|
|
viz = CalibrationDisplay.from_estimator(
|
|
lr, X, y, n_bins=n_bins, strategy=strategy, alpha=0.8
|
|
)
|
|
|
|
y_prob = lr.predict_proba(X)[:, 1]
|
|
prob_true, prob_pred = calibration_curve(
|
|
y, y_prob, n_bins=n_bins, strategy=strategy
|
|
)
|
|
|
|
assert_allclose(viz.prob_true, prob_true)
|
|
assert_allclose(viz.prob_pred, prob_pred)
|
|
assert_allclose(viz.y_prob, y_prob)
|
|
|
|
assert viz.estimator_name == "LogisticRegression"
|
|
|
|
# cannot fail thanks to pyplot fixture
|
|
import matplotlib as mpl # noqa
|
|
|
|
assert isinstance(viz.line_, mpl.lines.Line2D)
|
|
assert viz.line_.get_alpha() == 0.8
|
|
assert isinstance(viz.ax_, mpl.axes.Axes)
|
|
assert isinstance(viz.figure_, mpl.figure.Figure)
|
|
|
|
assert viz.ax_.get_xlabel() == "Mean predicted probability (Positive class: 1)"
|
|
assert viz.ax_.get_ylabel() == "Fraction of positives (Positive class: 1)"
|
|
|
|
expected_legend_labels = ["LogisticRegression", "Perfectly calibrated"]
|
|
legend_labels = viz.ax_.get_legend().get_texts()
|
|
assert len(legend_labels) == len(expected_legend_labels)
|
|
for labels in legend_labels:
|
|
assert labels.get_text() in expected_legend_labels
|
|
|
|
|
|
def test_plot_calibration_curve_pipeline(pyplot, iris_data_binary):
|
|
# Ensure pipelines are supported by CalibrationDisplay.from_estimator
|
|
X, y = iris_data_binary
|
|
clf = make_pipeline(StandardScaler(), LogisticRegression())
|
|
clf.fit(X, y)
|
|
viz = CalibrationDisplay.from_estimator(clf, X, y)
|
|
|
|
expected_legend_labels = [viz.estimator_name, "Perfectly calibrated"]
|
|
legend_labels = viz.ax_.get_legend().get_texts()
|
|
assert len(legend_labels) == len(expected_legend_labels)
|
|
for labels in legend_labels:
|
|
assert labels.get_text() in expected_legend_labels
|
|
|
|
|
|
@pytest.mark.parametrize(
|
|
"name, expected_label", [(None, "_line1"), ("my_est", "my_est")]
|
|
)
|
|
def test_calibration_display_default_labels(pyplot, name, expected_label):
|
|
prob_true = np.array([0, 1, 1, 0])
|
|
prob_pred = np.array([0.2, 0.8, 0.8, 0.4])
|
|
y_prob = np.array([])
|
|
|
|
viz = CalibrationDisplay(prob_true, prob_pred, y_prob, estimator_name=name)
|
|
viz.plot()
|
|
|
|
expected_legend_labels = [] if name is None else [name]
|
|
expected_legend_labels.append("Perfectly calibrated")
|
|
legend_labels = viz.ax_.get_legend().get_texts()
|
|
assert len(legend_labels) == len(expected_legend_labels)
|
|
for labels in legend_labels:
|
|
assert labels.get_text() in expected_legend_labels
|
|
|
|
|
|
def test_calibration_display_label_class_plot(pyplot):
|
|
# Checks that when instantiating `CalibrationDisplay` class then calling
|
|
# `plot`, `self.estimator_name` is the one given in `plot`
|
|
prob_true = np.array([0, 1, 1, 0])
|
|
prob_pred = np.array([0.2, 0.8, 0.8, 0.4])
|
|
y_prob = np.array([])
|
|
|
|
name = "name one"
|
|
viz = CalibrationDisplay(prob_true, prob_pred, y_prob, estimator_name=name)
|
|
assert viz.estimator_name == name
|
|
name = "name two"
|
|
viz.plot(name=name)
|
|
|
|
expected_legend_labels = [name, "Perfectly calibrated"]
|
|
legend_labels = viz.ax_.get_legend().get_texts()
|
|
assert len(legend_labels) == len(expected_legend_labels)
|
|
for labels in legend_labels:
|
|
assert labels.get_text() in expected_legend_labels
|
|
|
|
|
|
@pytest.mark.parametrize("constructor_name", ["from_estimator", "from_predictions"])
|
|
def test_calibration_display_name_multiple_calls(
|
|
constructor_name, pyplot, iris_data_binary
|
|
):
|
|
# Check that the `name` used when calling
|
|
# `CalibrationDisplay.from_predictions` or
|
|
# `CalibrationDisplay.from_estimator` is used when multiple
|
|
# `CalibrationDisplay.viz.plot()` calls are made.
|
|
X, y = iris_data_binary
|
|
clf_name = "my hand-crafted name"
|
|
clf = LogisticRegression().fit(X, y)
|
|
y_prob = clf.predict_proba(X)[:, 1]
|
|
|
|
constructor = getattr(CalibrationDisplay, constructor_name)
|
|
params = (clf, X, y) if constructor_name == "from_estimator" else (y, y_prob)
|
|
|
|
viz = constructor(*params, name=clf_name)
|
|
assert viz.estimator_name == clf_name
|
|
pyplot.close("all")
|
|
viz.plot()
|
|
|
|
expected_legend_labels = [clf_name, "Perfectly calibrated"]
|
|
legend_labels = viz.ax_.get_legend().get_texts()
|
|
assert len(legend_labels) == len(expected_legend_labels)
|
|
for labels in legend_labels:
|
|
assert labels.get_text() in expected_legend_labels
|
|
|
|
pyplot.close("all")
|
|
clf_name = "another_name"
|
|
viz.plot(name=clf_name)
|
|
assert len(legend_labels) == len(expected_legend_labels)
|
|
for labels in legend_labels:
|
|
assert labels.get_text() in expected_legend_labels
|
|
|
|
|
|
def test_calibration_display_ref_line(pyplot, iris_data_binary):
|
|
# Check that `ref_line` only appears once
|
|
X, y = iris_data_binary
|
|
lr = LogisticRegression().fit(X, y)
|
|
dt = DecisionTreeClassifier().fit(X, y)
|
|
|
|
viz = CalibrationDisplay.from_estimator(lr, X, y)
|
|
viz2 = CalibrationDisplay.from_estimator(dt, X, y, ax=viz.ax_)
|
|
|
|
labels = viz2.ax_.get_legend_handles_labels()[1]
|
|
assert labels.count("Perfectly calibrated") == 1
|
|
|
|
|
|
@pytest.mark.parametrize("dtype_y_str", [str, object])
|
|
def test_calibration_curve_pos_label_error_str(dtype_y_str):
|
|
"""Check error message when a `pos_label` is not specified with `str` targets."""
|
|
rng = np.random.RandomState(42)
|
|
y1 = np.array(["spam"] * 3 + ["eggs"] * 2, dtype=dtype_y_str)
|
|
y2 = rng.randint(0, 2, size=y1.size)
|
|
|
|
err_msg = (
|
|
"y_true takes value in {'eggs', 'spam'} and pos_label is not "
|
|
"specified: either make y_true take value in {0, 1} or {-1, 1} or "
|
|
"pass pos_label explicitly"
|
|
)
|
|
with pytest.raises(ValueError, match=err_msg):
|
|
calibration_curve(y1, y2)
|
|
|
|
|
|
@pytest.mark.parametrize("dtype_y_str", [str, object])
|
|
def test_calibration_curve_pos_label(dtype_y_str):
|
|
"""Check the behaviour when passing explicitly `pos_label`."""
|
|
y_true = np.array([0, 0, 0, 1, 1, 1, 1, 1, 1])
|
|
classes = np.array(["spam", "egg"], dtype=dtype_y_str)
|
|
y_true_str = classes[y_true]
|
|
y_pred = np.array([0.1, 0.2, 0.3, 0.4, 0.65, 0.7, 0.8, 0.9, 1.0])
|
|
|
|
# default case
|
|
prob_true, _ = calibration_curve(y_true, y_pred, n_bins=4)
|
|
assert_allclose(prob_true, [0, 0.5, 1, 1])
|
|
# if `y_true` contains `str`, then `pos_label` is required
|
|
prob_true, _ = calibration_curve(y_true_str, y_pred, n_bins=4, pos_label="egg")
|
|
assert_allclose(prob_true, [0, 0.5, 1, 1])
|
|
|
|
prob_true, _ = calibration_curve(y_true, 1 - y_pred, n_bins=4, pos_label=0)
|
|
assert_allclose(prob_true, [0, 0, 0.5, 1])
|
|
prob_true, _ = calibration_curve(y_true_str, 1 - y_pred, n_bins=4, pos_label="spam")
|
|
assert_allclose(prob_true, [0, 0, 0.5, 1])
|
|
|
|
|
|
@pytest.mark.parametrize("pos_label, expected_pos_label", [(None, 1), (0, 0), (1, 1)])
|
|
def test_calibration_display_pos_label(
|
|
pyplot, iris_data_binary, pos_label, expected_pos_label
|
|
):
|
|
"""Check the behaviour of `pos_label` in the `CalibrationDisplay`."""
|
|
X, y = iris_data_binary
|
|
|
|
lr = LogisticRegression().fit(X, y)
|
|
viz = CalibrationDisplay.from_estimator(lr, X, y, pos_label=pos_label)
|
|
|
|
y_prob = lr.predict_proba(X)[:, expected_pos_label]
|
|
prob_true, prob_pred = calibration_curve(y, y_prob, pos_label=pos_label)
|
|
|
|
assert_allclose(viz.prob_true, prob_true)
|
|
assert_allclose(viz.prob_pred, prob_pred)
|
|
assert_allclose(viz.y_prob, y_prob)
|
|
|
|
assert (
|
|
viz.ax_.get_xlabel()
|
|
== f"Mean predicted probability (Positive class: {expected_pos_label})"
|
|
)
|
|
assert (
|
|
viz.ax_.get_ylabel()
|
|
== f"Fraction of positives (Positive class: {expected_pos_label})"
|
|
)
|
|
|
|
expected_legend_labels = [lr.__class__.__name__, "Perfectly calibrated"]
|
|
legend_labels = viz.ax_.get_legend().get_texts()
|
|
assert len(legend_labels) == len(expected_legend_labels)
|
|
for labels in legend_labels:
|
|
assert labels.get_text() in expected_legend_labels
|
|
|
|
|
|
@pytest.mark.parametrize("method", ["sigmoid", "isotonic"])
|
|
@pytest.mark.parametrize("ensemble", [True, False])
|
|
def test_calibrated_classifier_cv_double_sample_weights_equivalence(method, ensemble):
|
|
"""Check that passing repeating twice the dataset `X` is equivalent to
|
|
passing a `sample_weight` with a factor 2."""
|
|
X, y = load_iris(return_X_y=True)
|
|
# Scale the data to avoid any convergence issue
|
|
X = StandardScaler().fit_transform(X)
|
|
# Only use 2 classes
|
|
X, y = X[:100], y[:100]
|
|
sample_weight = np.ones_like(y) * 2
|
|
|
|
# Interlace the data such that a 2-fold cross-validation will be equivalent
|
|
# to using the original dataset with a sample weights of 2
|
|
X_twice = np.zeros((X.shape[0] * 2, X.shape[1]), dtype=X.dtype)
|
|
X_twice[::2, :] = X
|
|
X_twice[1::2, :] = X
|
|
y_twice = np.zeros(y.shape[0] * 2, dtype=y.dtype)
|
|
y_twice[::2] = y
|
|
y_twice[1::2] = y
|
|
|
|
estimator = LogisticRegression()
|
|
calibrated_clf_without_weights = CalibratedClassifierCV(
|
|
estimator,
|
|
method=method,
|
|
ensemble=ensemble,
|
|
cv=2,
|
|
)
|
|
calibrated_clf_with_weights = clone(calibrated_clf_without_weights)
|
|
|
|
calibrated_clf_with_weights.fit(X, y, sample_weight=sample_weight)
|
|
calibrated_clf_without_weights.fit(X_twice, y_twice)
|
|
|
|
# Check that the underlying fitted estimators have the same coefficients
|
|
for est_with_weights, est_without_weights in zip(
|
|
calibrated_clf_with_weights.calibrated_classifiers_,
|
|
calibrated_clf_without_weights.calibrated_classifiers_,
|
|
):
|
|
assert_allclose(
|
|
est_with_weights.estimator.coef_,
|
|
est_without_weights.estimator.coef_,
|
|
)
|
|
|
|
# Check that the predictions are the same
|
|
y_pred_with_weights = calibrated_clf_with_weights.predict_proba(X)
|
|
y_pred_without_weights = calibrated_clf_without_weights.predict_proba(X)
|
|
|
|
assert_allclose(y_pred_with_weights, y_pred_without_weights)
|
|
|
|
|
|
@pytest.mark.parametrize("fit_params_type", ["list", "array"])
|
|
def test_calibration_with_fit_params(fit_params_type, data):
|
|
"""Tests that fit_params are passed to the underlying base estimator.
|
|
|
|
Non-regression test for:
|
|
https://github.com/scikit-learn/scikit-learn/issues/12384
|
|
"""
|
|
X, y = data
|
|
fit_params = {
|
|
"a": _convert_container(y, fit_params_type),
|
|
"b": _convert_container(y, fit_params_type),
|
|
}
|
|
|
|
clf = CheckingClassifier(expected_fit_params=["a", "b"])
|
|
pc_clf = CalibratedClassifierCV(clf)
|
|
|
|
pc_clf.fit(X, y, **fit_params)
|
|
|
|
|
|
@pytest.mark.parametrize(
|
|
"sample_weight",
|
|
[
|
|
[1.0] * N_SAMPLES,
|
|
np.ones(N_SAMPLES),
|
|
],
|
|
)
|
|
def test_calibration_with_sample_weight_base_estimator(sample_weight, data):
|
|
"""Tests that sample_weight is passed to the underlying base
|
|
estimator.
|
|
"""
|
|
X, y = data
|
|
clf = CheckingClassifier(expected_sample_weight=True)
|
|
pc_clf = CalibratedClassifierCV(clf)
|
|
|
|
pc_clf.fit(X, y, sample_weight=sample_weight)
|
|
|
|
|
|
def test_calibration_without_sample_weight_base_estimator(data):
|
|
"""Check that even if the estimator doesn't support
|
|
sample_weight, fitting with sample_weight still works.
|
|
|
|
There should be a warning, since the sample_weight is not passed
|
|
on to the estimator.
|
|
"""
|
|
X, y = data
|
|
sample_weight = np.ones_like(y)
|
|
|
|
class ClfWithoutSampleWeight(CheckingClassifier):
|
|
def fit(self, X, y, **fit_params):
|
|
assert "sample_weight" not in fit_params
|
|
return super().fit(X, y, **fit_params)
|
|
|
|
clf = ClfWithoutSampleWeight()
|
|
pc_clf = CalibratedClassifierCV(clf)
|
|
|
|
with pytest.warns(UserWarning):
|
|
pc_clf.fit(X, y, sample_weight=sample_weight)
|
|
|
|
|
|
def test_calibration_with_fit_params_inconsistent_length(data):
|
|
"""fit_params having different length than data should raise the
|
|
correct error message.
|
|
"""
|
|
X, y = data
|
|
fit_params = {"a": y[:5]}
|
|
clf = CheckingClassifier(expected_fit_params=fit_params)
|
|
pc_clf = CalibratedClassifierCV(clf)
|
|
|
|
msg = (
|
|
r"Found input variables with inconsistent numbers of "
|
|
r"samples: \[" + str(N_SAMPLES) + r", 5\]"
|
|
)
|
|
with pytest.raises(ValueError, match=msg):
|
|
pc_clf.fit(X, y, **fit_params)
|
|
|
|
|
|
@pytest.mark.parametrize("method", ["sigmoid", "isotonic"])
|
|
@pytest.mark.parametrize("ensemble", [True, False])
|
|
def test_calibrated_classifier_cv_zeros_sample_weights_equivalence(method, ensemble):
|
|
"""Check that passing removing some sample from the dataset `X` is
|
|
equivalent to passing a `sample_weight` with a factor 0."""
|
|
X, y = load_iris(return_X_y=True)
|
|
# Scale the data to avoid any convergence issue
|
|
X = StandardScaler().fit_transform(X)
|
|
# Only use 2 classes and select samples such that 2-fold cross-validation
|
|
# split will lead to an equivalence with a `sample_weight` of 0
|
|
X = np.vstack((X[:40], X[50:90]))
|
|
y = np.hstack((y[:40], y[50:90]))
|
|
sample_weight = np.zeros_like(y)
|
|
sample_weight[::2] = 1
|
|
|
|
estimator = LogisticRegression()
|
|
calibrated_clf_without_weights = CalibratedClassifierCV(
|
|
estimator,
|
|
method=method,
|
|
ensemble=ensemble,
|
|
cv=2,
|
|
)
|
|
calibrated_clf_with_weights = clone(calibrated_clf_without_weights)
|
|
|
|
calibrated_clf_with_weights.fit(X, y, sample_weight=sample_weight)
|
|
calibrated_clf_without_weights.fit(X[::2], y[::2])
|
|
|
|
# Check that the underlying fitted estimators have the same coefficients
|
|
for est_with_weights, est_without_weights in zip(
|
|
calibrated_clf_with_weights.calibrated_classifiers_,
|
|
calibrated_clf_without_weights.calibrated_classifiers_,
|
|
):
|
|
assert_allclose(
|
|
est_with_weights.estimator.coef_,
|
|
est_without_weights.estimator.coef_,
|
|
)
|
|
|
|
# Check that the predictions are the same
|
|
y_pred_with_weights = calibrated_clf_with_weights.predict_proba(X)
|
|
y_pred_without_weights = calibrated_clf_without_weights.predict_proba(X)
|
|
|
|
assert_allclose(y_pred_with_weights, y_pred_without_weights)
|
|
|
|
|
|
# TODO(1.4): Remove
|
|
def test_calibrated_classifier_error_base_estimator(data):
|
|
"""Check that we raise an error is a user set both `base_estimator` and
|
|
`estimator`."""
|
|
calibrated_classifier = CalibratedClassifierCV(
|
|
base_estimator=LogisticRegression(), estimator=LogisticRegression()
|
|
)
|
|
with pytest.raises(ValueError, match="Both `base_estimator` and `estimator`"):
|
|
calibrated_classifier.fit(*data)
|
|
|
|
|
|
# TODO(1.4): Remove
|
|
def test_calibrated_classifier_deprecation_base_estimator(data):
|
|
"""Check that we raise a warning regarding the deprecation of
|
|
`base_estimator`."""
|
|
calibrated_classifier = CalibratedClassifierCV(base_estimator=LogisticRegression())
|
|
warn_msg = "`base_estimator` was renamed to `estimator`"
|
|
with pytest.warns(FutureWarning, match=warn_msg):
|
|
calibrated_classifier.fit(*data)
|