Note
Go to the end to download the full example code.
Cache mechanism#
This example shows how EstimatorReport
and
CrossValidationReport
use caching to speed up computations.
We set some environment variables to avoid some spurious warnings related to parallelism.
import os
os.environ["POLARS_ALLOW_FORKING_THREAD"] = "1"
Loading some data#
First, we load a dataset from skrub
. Our goal is to predict if a company paid a
physician. The ultimate goal is to detect potential conflict of interest when it comes
to the actual problem that we want to solve.
from skrub.datasets import fetch_open_payments
dataset = fetch_open_payments()
df = dataset.X
y = dataset.y
from skrub import TableReport
TableReport(df)
import pandas as pd
TableReport(pd.DataFrame(y))
The dataset has over 70,000 records with only categorical features. Some categories are not well defined.
Caching with EstimatorReport
and CrossValidationReport
#
We use skrub
to create a simple predictive model that handles our dataset’s
challenges.
from skrub import tabular_learner
model = tabular_learner("classifier")
model
This model handles all types of data: numbers, categories, dates, and missing values. Let’s train it on part of our dataset.
from skore import train_test_split
X_train, X_test, y_train, y_test = train_test_split(df, y, random_state=42)
╭───────────────────────────── HighClassImbalanceWarning ──────────────────────────────╮
│ It seems that you have a classification problem with a high class imbalance. In this │
│ case, using train_test_split may not be a good idea because of high variability in │
│ the scores obtained on the test set. To tackle this challenge we suggest to use │
│ skore's cross_validate function. │
╰──────────────────────────────────────────────────────────────────────────────────────╯
╭───────────────────────────────── ShuffleTrueWarning ─────────────────────────────────╮
│ We detected that the `shuffle` parameter is set to `True` either explicitly or from │
│ its default value. In case of time-ordered events (even if they are independent), │
│ this will result in inflated model performance evaluation because natural drift will │
│ not be taken into account. We recommend setting the shuffle parameter to `False` in │
│ order to ensure the evaluation process is really representative of your production │
│ release process. │
╰──────────────────────────────────────────────────────────────────────────────────────╯
Caching the predictions for fast metric computation#
First, we focus on EstimatorReport
, as the same philosophy will
apply to CrossValidationReport
.
Let’s explore how EstimatorReport
uses caching to speed up
predictions. We start by training the model:
from skore import EstimatorReport
report = EstimatorReport(
model, X_train=X_train, y_train=y_train, X_test=X_test, y_test=y_test
)
report.help()
╭──────────── Tools to diagnose estimator HistGradientBoostingClassifier ─────────────╮
│ EstimatorReport │
│ ├── .metrics │
│ │ ├── .accuracy(...) (↗︎) - Compute the accuracy score. │
│ │ ├── .brier_score(...) (↘︎) - Compute the Brier score. │
│ │ ├── .log_loss(...) (↘︎) - Compute the log loss. │
│ │ ├── .precision(...) (↗︎) - Compute the precision score. │
│ │ ├── .precision_recall(...) - Plot the precision-recall curve. │
│ │ ├── .recall(...) (↗︎) - Compute the recall score. │
│ │ ├── .roc(...) - Plot the ROC curve. │
│ │ ├── .roc_auc(...) (↗︎) - Compute the ROC AUC score. │
│ │ ├── .custom_metric(...) - Compute a custom metric. │
│ │ └── .report_metrics(...) - Report a set of metrics for our estimator. │
│ ├── .cache_predictions(...) - Cache estimator's predictions. │
│ ├── .clear_cache(...) - Clear the cache. │
│ └── Attributes │
│ ├── .X_test │
│ ├── .X_train │
│ ├── .y_test │
│ ├── .y_train │
│ ├── .estimator_ │
│ └── .estimator_name_ │
│ │
│ │
│ Legend: │
│ (↗︎) higher is better (↘︎) lower is better │
╰─────────────────────────────────────────────────────────────────────────────────────╯
We compute the accuracy on our test set and measure how long it takes:
0.9528548123980424
Time taken: 1.56 seconds
For comparison, here’s how scikit-learn computes the same accuracy score:
from sklearn.metrics import accuracy_score
start = time.time()
result = accuracy_score(report.y_test, report.estimator_.predict(report.X_test))
end = time.time()
result
0.9528548123980424
Time taken: 1.55 seconds
Both approaches take similar time.
Now, watch what happens when we compute the accuracy again with our skore estimator report:
0.9528548123980424
Time taken: 0.00 seconds
The second calculation is instant! This happens because the report saves previous calculations in its cache. Let’s look inside the cache:
{(np.int64(785285855598900323), 'predict', 'test'): array(['disallowed', 'disallowed', 'disallowed', ..., 'disallowed',
'disallowed', 'disallowed'], shape=(18390,), dtype=object), (np.int64(785285855598900323), 'accuracy_score', 'test'): 0.9528548123980424}
The cache stores predictions by type and data source. This means that computing metrics that use the same type of predictions will be faster. Let’s try the precision metric:
{'allowed': np.float64(0.6906290115532734), 'disallowed': np.float64(0.9644540344103117)}
Time taken: 0.06 seconds
We observe that it takes only a few milliseconds to compute the precision because we don’t need to re-compute the predictions and only have to compute the precision metric itself. Since the predictions are the bottleneck in terms of computation time, we observe an interesting speedup.
Caching all the possible predictions at once#
We can pre-compute all predictions at once using parallel processing:
report.cache_predictions(n_jobs=4)
Now, all possible predictions are stored. Any metric calculation will be much faster, even on different data (like the training set):
0.09865494232505337
Time taken: 0.11 seconds
Caching external data#
The report can also work with external data. We use data_source="X_y"
to indicate
that we want to pass those external data.
0.12305206715107839
Time taken: 1.75 seconds
The first calculation of the above cell is slower than when using the internal train or test sets because it needs to compute a hash of the new data for later retrieval. Let’s calculate it again:
0.12305206715107839
Time taken: 0.18 seconds
It is much faster for the second time as the predictions are cached! The remaining time corresponds to the hash computation. Let’s compute the ROC AUC on the same data:
0.9439820500298637
Time taken: 0.21 seconds
We observe that the computation is already efficient because it boils down to two computations: the hash of the data and the ROC-AUC metric. We save a lot of time because we don’t need to re-compute the predictions.
Caching for plotting#
The cache also speeds up plots. Let’s create a ROC curve:
import matplotlib.pyplot as plt
start = time.time()
display = report.metrics.roc(pos_label="allowed")
display.plot()
end = time.time()
plt.tight_layout()

Time taken: 0.02 seconds
The second plot is instant because it uses cached data:
start = time.time()
display = report.metrics.roc(pos_label="allowed")
display.plot()
end = time.time()
plt.tight_layout()

Time taken: 0.01 seconds
We only use the cache to retrieve the display
object and not directly the matplotlib
figure. It means that we can still customize the cached plot before displaying it:
display.plot(roc_curve_kwargs={"color": "tab:orange"})
plt.tight_layout()

Be aware that we can clear the cache if we want to:
{}
It means that nothing is stored anymore in the cache.
Caching with CrossValidationReport
#
CrossValidationReport
uses the same caching system for each fold
in cross-validation by leveraging the previous EstimatorReport
:
from skore import CrossValidationReport
report = CrossValidationReport(model, X=df, y=y, cv_splitter=5, n_jobs=4)
report.help()
╭───────────── Tools to diagnose estimator HistGradientBoostingClassifier ─────────────╮
│ CrossValidationReport │
│ ├── .metrics │
│ │ ├── .accuracy(...) (↗︎) - Compute the accuracy score. │
│ │ ├── .brier_score(...) (↘︎) - Compute the Brier score. │
│ │ ├── .log_loss(...) (↘︎) - Compute the log loss. │
│ │ ├── .precision(...) (↗︎) - Compute the precision score. │
│ │ ├── .precision_recall(...) - Plot the precision-recall curve. │
│ │ ├── .recall(...) (↗︎) - Compute the recall score. │
│ │ ├── .roc(...) - Plot the ROC curve. │
│ │ ├── .roc_auc(...) (↗︎) - Compute the ROC AUC score. │
│ │ ├── .custom_metric(...) - Compute a custom metric. │
│ │ └── .report_metrics(...) - Report a set of metrics for our estimator. │
│ ├── .cache_predictions(...) - Cache the predictions for sub-estimators │
│ │ reports. │
│ ├── .clear_cache(...) - Clear the cache. │
│ └── Attributes │
│ ├── .X │
│ ├── .y │
│ ├── .estimator_ │
│ ├── .estimator_name_ │
│ ├── .estimator_reports_ │
│ └── .n_jobs │
│ │
│ │
│ Legend: │
│ (↗︎) higher is better (↘︎) lower is better │
╰──────────────────────────────────────────────────────────────────────────────────────╯
Since a CrossValidationReport
uses many
EstimatorReport
, we will observe the same behaviour as we previously
exposed.
The first call will be slow because it computes the predictions for each fold.
Time taken: 20.98 seconds
But the subsequent calls are fast because the predictions are cached.
Time taken: 0.00 seconds
Hence, we observe the same type of behaviour as we previously exposed.
Total running time of the script: (1 minutes 55.838 seconds)