BookmarkSubscribeRSS Feed
Zaid_Sanchez
Calcite | Level 5

I’m tuning a binary model with PROC GRADBOOST + AUTOTUNE and PARTITION (train/validation/test). The HISTORYTABLE= captures one objective per evaluation (Gini on validation or the k-fold average), but I need Gini per partition (train/valid/test) for every evaluated configuration. Right now, my workaround is to refit each configuration, score, and run PROC ASSESS BY _PartInd_ to compute AUC → Gini, which is costly.
Questions:

  1. Is there a built-in way to get partition-level metrics (train/valid/test) for all evaluated configurations without refitting each one?
  2. Is there a supported pattern to persist each candidate model during tuning (not just the champion) so I can score later and compute partition metrics without retraining?
  3. Any roadmap to add partition-level metrics (or an option to save all intermediate models) to AUTOTUNE outputs?

From the docs, AUTOTUNE records a single objective per evaluation (validation/k-fold) in the history and doesn’t expose per-partition metrics by default.

1 REPLY 1
sbxkoenk
SAS Super FREQ

Hello,

 

As far as I know, it's indeed not possible to get the tuner objective function (GINI for example or Misclassification Error Percentage) for all partitions. The objective metric you see (in autotune historytable=) is the one coming from validation calculations. 

 

Remark that a partition is always used by Autotune, specifically to avoid overfitting. This partition (separate table for scoring/validation) is either supplied by the user or created internally by Autotune.

 

 

Some references:


Ciao,

Koen

Catch up on SAS Innovate 2026

Dive into keynotes, announcements and breakthroughs on demand.

Explore Now →
How to choose a machine learning algorithm

Use this tutorial as a handy guide to weigh the pros and cons of these commonly used machine learning algorithms.

Find more tutorials on the SAS Users YouTube channel.

Discussion stats
  • 1 reply
  • 470 views
  • 1 like
  • 2 in conversation