<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Cross Validation in regression and decision trees in SAS Data Science</title>
    <link>https://communities.sas.com/t5/SAS-Data-Science/Cross-Validation-in-regression-and-decision-trees/m-p/14155#M56</link>
    <description>In SAS decision trees, ’10 repeats’ means 10-fold cross-validation 10 times, for a total of 101 trees, including the original tree.&lt;BR /&gt;
&lt;BR /&gt;
'Leave-one-out' cross-validation has been available in the EM Regression Node. In leave-one-out CV,  n = the total # of observations in your data set.&lt;BR /&gt;
 &lt;BR /&gt;
Re: Using CV, do you still partition your data into training and validation subsets?&lt;BR /&gt;
Not for a single EM modelling node.  However, partitioning into data-available-for-CV vs test-hold-out is still useful, and if comparing models from several EM modeling nodes, using a single validation data set for the comparison may be useful.  It's up to the analyst.&lt;BR /&gt;
&lt;BR /&gt;
Re: primarily used when small data sets are not large enough for partitioning&lt;BR /&gt;
That is my belief.  Partitioning applies hold-out data directly to the model being deployed, providing a transparently unbiased estimate of accuracy.  CV validates the model construction process.  People disagree as to whether leave-one-out cross-validation provides unbiased or overrly optimistic estimates of prediction.&lt;BR /&gt;
&lt;BR /&gt;
However, many people prefer to CV anything, regardless of size.</description>
    <pubDate>Tue, 31 May 2011 17:26:58 GMT</pubDate>
    <dc:creator>PadraicGNeville</dc:creator>
    <dc:date>2011-05-31T17:26:58Z</dc:date>
    <item>
      <title>Cross Validation in regression and decision trees</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/Cross-Validation-in-regression-and-decision-trees/m-p/14153#M54</link>
      <description>I have 2 questions:&lt;BR /&gt;
&lt;BR /&gt;
1)  Cross validated Decision trees:  Under the panel for cross validation, if you select 'yes' and for number of subsets ='10' and number of repeats ='10' , are these results equivalent to 10-fold cross validation? &lt;BR /&gt;
&lt;BR /&gt;
Cross validated  regression: When you choose 'cross validation misclassification' as your selection criteria for the logistic regression node,  it seems that this is similar to an n-fold cross validation where n = the total # of observations in your data set? Is that correct? &lt;BR /&gt;
&lt;BR /&gt;
2) With cross validation techniques, do you still partition your data into training and validation subsets? I'm thinking, based on sas help documentation, since it is primarily used when small data sets are not large enough for partitioning, you wouldn't generally use a cross validation technique with partitioned data.</description>
      <pubDate>Mon, 21 Feb 2011 17:01:20 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/Cross-Validation-in-regression-and-decision-trees/m-p/14153#M54</guid>
      <dc:creator>SlutskyFan</dc:creator>
      <dc:date>2011-02-21T17:01:20Z</dc:date>
    </item>
    <item>
      <title>Re: Cross Validation in regression and decision trees</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/Cross-Validation-in-regression-and-decision-trees/m-p/14154#M55</link>
      <description>I'm not sure of the Cross validation in regression. I agree with the Decision Tree method and your second point.</description>
      <pubDate>Mon, 21 Mar 2011 05:30:43 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/Cross-Validation-in-regression-and-decision-trees/m-p/14154#M55</guid>
      <dc:creator>SatishG</dc:creator>
      <dc:date>2011-03-21T05:30:43Z</dc:date>
    </item>
    <item>
      <title>Re: Cross Validation in regression and decision trees</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/Cross-Validation-in-regression-and-decision-trees/m-p/14155#M56</link>
      <description>In SAS decision trees, ’10 repeats’ means 10-fold cross-validation 10 times, for a total of 101 trees, including the original tree.&lt;BR /&gt;
&lt;BR /&gt;
'Leave-one-out' cross-validation has been available in the EM Regression Node. In leave-one-out CV,  n = the total # of observations in your data set.&lt;BR /&gt;
 &lt;BR /&gt;
Re: Using CV, do you still partition your data into training and validation subsets?&lt;BR /&gt;
Not for a single EM modelling node.  However, partitioning into data-available-for-CV vs test-hold-out is still useful, and if comparing models from several EM modeling nodes, using a single validation data set for the comparison may be useful.  It's up to the analyst.&lt;BR /&gt;
&lt;BR /&gt;
Re: primarily used when small data sets are not large enough for partitioning&lt;BR /&gt;
That is my belief.  Partitioning applies hold-out data directly to the model being deployed, providing a transparently unbiased estimate of accuracy.  CV validates the model construction process.  People disagree as to whether leave-one-out cross-validation provides unbiased or overrly optimistic estimates of prediction.&lt;BR /&gt;
&lt;BR /&gt;
However, many people prefer to CV anything, regardless of size.</description>
      <pubDate>Tue, 31 May 2011 17:26:58 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/Cross-Validation-in-regression-and-decision-trees/m-p/14155#M56</guid>
      <dc:creator>PadraicGNeville</dc:creator>
      <dc:date>2011-05-31T17:26:58Z</dc:date>
    </item>
  </channel>
</rss>

