<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: SAS EMiner Machine Learning Models - Stability Check in SAS Data Science</title>
    <link>https://communities.sas.com/t5/SAS-Data-Science/SAS-EMiner-Machine-Learning-Models-Stability-Check/m-p/386111#M5705</link>
    <description>&lt;BLOCKQUOTE&gt;
&lt;P&gt;&lt;SPAN&gt;Should we expect rank ordering to hold good on out of time samples for machine learning classification models?&amp;nbsp;&lt;/SPAN&gt;&lt;/P&gt;
&lt;/BLOCKQUOTE&gt;
&lt;P&gt;&lt;SPAN&gt;&lt;BR /&gt;If the model didn't perform reasonably well on out-of-time samples, it would not be a particularly useful model. &amp;nbsp;The expectation, of course, is that model performance on out-of-time samples will not be as good as the data which was used to train it but it does provide a benchmark for using the model going forward. &amp;nbsp;SAS Model Manager is designed to apply previously fit models to future data and evaluate the performance. &amp;nbsp; Over time, the model performance is likely to degrade and require a refit. &amp;nbsp; How quickly it degrades, though, is a function of many factors including how well the training data reflected the population at the time of modeling, how the population and/or external factors has changed, and how well the model actually fit. &amp;nbsp;Should the modeler desire to refit the model, SAS Model Manager can perform that task as well in most situations. &amp;nbsp; In general, performance the out-of-time sample provide the best evaluation about how useful a particular model is at that time. &amp;nbsp;&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;Hope this helps!&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;Doug&lt;/SPAN&gt;&lt;/P&gt;</description>
    <pubDate>Mon, 07 Aug 2017 19:06:26 GMT</pubDate>
    <dc:creator>DougWielenga</dc:creator>
    <dc:date>2017-08-07T19:06:26Z</dc:date>
    <item>
      <title>SAS EMiner Machine Learning Models - Stability Check</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/SAS-EMiner-Machine-Learning-Models-Stability-Check/m-p/343230#M5129</link>
      <description>&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Hi, I have built a Gradient Boosting Model using SAS EMiner 13.1 on data with 1.8% event rate { target is binary variable }.&lt;/P&gt;&lt;P&gt;Model results are good and I wanted to test the model on Out of time.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Hence applied the GB scoring code on data set generated on a different time frame. After running scoring code, I wanted to check rank ordering if it still holds good.{ not sure if this is expected on machine learning models,&amp;nbsp;It's done on traditional logit models for stability}&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;data temp1;&lt;BR /&gt;set Scored_gb_dataset ;&lt;/P&gt;&lt;P&gt;proc sort data=temp1; by descending EM_EVENTPROBABILITY; run;&lt;BR /&gt;data temp2 (drop = i count);&lt;BR /&gt;set temp1 nobs = size;&lt;/P&gt;&lt;P&gt;count + 1;&lt;BR /&gt;do i = 1 to 10;&lt;BR /&gt;&amp;nbsp; if (i-1) * (size/10) &amp;lt; count &amp;lt;= i * (size/10) then decile = i;&lt;BR /&gt;end;&lt;BR /&gt;run;&lt;/P&gt;&lt;P&gt;proc freq data = temp2 formchar = '&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; ';&lt;BR /&gt;tables decile * actual_target /nocum norow nocol nopercent;&lt;BR /&gt;run;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I sorted the data based on EM_EVENT PROBABILITY and created deciles based on number of observations and I have checked # of actual responders by&amp;nbsp;decile to see rank ordering and it breaks on 4th decile. However, it does capture ~75% events on top 3 deciles.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Usually for classification models like decision tree's, they would be classified as High/Medium/Low risk segments and events captured&amp;nbsp; by these H/M/L segments could indicate validity on out of time validation. But here probability is assigned for each observation or ID I think.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Should we expect rank ordering to hold good on out of time samples for machine learning classification models? Appreciate your help/thoughts on the same&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 22 Mar 2017 11:32:08 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/SAS-EMiner-Machine-Learning-Models-Stability-Check/m-p/343230#M5129</guid>
      <dc:creator>VMadhav</dc:creator>
      <dc:date>2017-03-22T11:32:08Z</dc:date>
    </item>
    <item>
      <title>Re: SAS EMiner Machine Learning Models - Stability Check</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/SAS-EMiner-Machine-Learning-Models-Stability-Check/m-p/386111#M5705</link>
      <description>&lt;BLOCKQUOTE&gt;
&lt;P&gt;&lt;SPAN&gt;Should we expect rank ordering to hold good on out of time samples for machine learning classification models?&amp;nbsp;&lt;/SPAN&gt;&lt;/P&gt;
&lt;/BLOCKQUOTE&gt;
&lt;P&gt;&lt;SPAN&gt;&lt;BR /&gt;If the model didn't perform reasonably well on out-of-time samples, it would not be a particularly useful model. &amp;nbsp;The expectation, of course, is that model performance on out-of-time samples will not be as good as the data which was used to train it but it does provide a benchmark for using the model going forward. &amp;nbsp;SAS Model Manager is designed to apply previously fit models to future data and evaluate the performance. &amp;nbsp; Over time, the model performance is likely to degrade and require a refit. &amp;nbsp; How quickly it degrades, though, is a function of many factors including how well the training data reflected the population at the time of modeling, how the population and/or external factors has changed, and how well the model actually fit. &amp;nbsp;Should the modeler desire to refit the model, SAS Model Manager can perform that task as well in most situations. &amp;nbsp; In general, performance the out-of-time sample provide the best evaluation about how useful a particular model is at that time. &amp;nbsp;&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;Hope this helps!&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;Doug&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Mon, 07 Aug 2017 19:06:26 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/SAS-EMiner-Machine-Learning-Models-Stability-Check/m-p/386111#M5705</guid>
      <dc:creator>DougWielenga</dc:creator>
      <dc:date>2017-08-07T19:06:26Z</dc:date>
    </item>
  </channel>
</rss>

