Regression problems involve continuous outcomes rather than class labels.
This difference is not cosmetic.
It changes how performance is measured, how preprocessing affects results, and how evaluation errors arise.
Many applied workflows treat regression as a simpler case than classification.
In practice, regression is often more fragile under improper evaluation.
Why regression requires separate treatment
In regression:
outcomes have a natural scale
errors are measured in outcome units
preprocessing directly affects metric values
comparisons across datasets are rarely meaningful
Metrics such as RMSE and MAE are scale-dependent.
A “good” RMSE cannot be interpreted without context.
Leakage risks in regression workflows
Regression pipelines are particularly sensitive to leakage because:
scaling affects both predictors and outcome interpretation
imputation alters the outcome–predictor relationship
feature engineering often uses outcome-adjacent information
residual structure is easily distorted
As in earlier tutorials, leakage arises from when information is learned, not from model choice.
Data
We use a medical dataset with a continuous outcome.
As in previous tutorials, aggregated metrics hide variability.
fit$resampling_results$`linear_reg (lm)`$folds
# A tibble: 30 × 4
fold .metric .estimator .estimate
<chr> <chr> <chr> <dbl>
1 1 rmse standard 4.17
2 1 rsq standard 0.757
3 1 mae standard 3.09
4 2 rmse standard 3.41
5 2 rsq standard 0.822
6 2 mae standard 2.51
7 3 rmse standard 5.96
8 3 rsq standard 0.713
9 3 mae standard 3.66
10 4 rmse standard 4.59
# ℹ 20 more rows
Fold-level results reveal:
instability due to limited sample size,
sensitivity to data partitioning,
overlap between competing models.
Regression performance often varies more across folds than classification accuracy.
Model Comparison in Regression
Comparing regression models requires:
identical resampling splits,
identical preprocessing,
identical outcome scaling.
fastml enforces these conditions automatically.
Observed performance differences therefore reflect model behavior rather than evaluation artifacts.
What fastml Does Not Claim
fastml does not:
identify a “true” model,
guarantee optimal predictive accuracy,
justify causal interpretation of coefficients,
standardize outcomes for cross-study comparison.
Regression metrics describe predictive error under a specific evaluation design, nothing more.
Responsible Reporting
For regression analyses, a defensible report should include:
the outcome definition and scale,
the chosen evaluation metrics,
the resampling scheme,
aggregated performance estimates,
variability across folds.
Reporting a single RMSE without context is insufficient.
Summary
Regression introduces scale-dependent evaluation.
Metrics are sensitive to preprocessing and leakage.
Guarded resampling is essential for valid regression evaluation.
Variability across folds is informative.