<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>WayneThompson Tracker</title>
    <link>https://communities.sas.com/kntur85557/tracker</link>
    <description>WayneThompson Tracker</description>
    <pubDate>Sun, 10 May 2026 14:29:24 GMT</pubDate>
    <dc:date>2026-05-10T14:29:24Z</dc:date>
    <item>
      <title>Re: Boosting, bagging and Random Forest</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/Boosting-bagging-and-Random-Forest/m-p/355625#M5275</link>
      <description>&lt;P&gt;Very thorough reply by Neville who implemented all of these methods in SAS. &amp;nbsp;A couple of supporting comments: &amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;1. Outliers The base model used in RF is a large decision tree. Decision trees are robust to outliers, because they isolate them in small regions of the feature space. Then, since the prediction for each leaf is the average (for regression) or the majority class (for classification), being isolated in separate leaves, outliers won't influence the rest of the predictions (in the case of regression for instance, they would not impact the mean of the other leaves)&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;2, Validation Data -- Yes please use in teh common case of rare events where the OOB might not be sufficient.&amp;nbsp;&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;3. Transforms - consider continous Y as with many algorithms.&amp;nbsp;&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 03 May 2017 14:37:51 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/Boosting-bagging-and-Random-Forest/m-p/355625#M5275</guid>
      <dc:creator>WayneThompson</dc:creator>
      <dc:date>2017-05-03T14:37:51Z</dc:date>
    </item>
    <item>
      <title>Re: What is SAS Visual Statistics?</title>
      <link>https://communities.sas.com/t5/SAS-Communities-Library/What-is-SAS-Visual-Statistics/tac-p/223461#M630</link>
      <description>&lt;HTML&gt;&lt;HEAD&gt;&lt;/HEAD&gt;&lt;BODY&gt;&lt;P&gt;Visual Statistics (VS) is a web client based product that supports interactive ad-hoc data analysis whereas Enterprise Miner (EM) is a rich client with&amp;nbsp; more of a&amp;nbsp; batch type interface for developing repeatable analysis through the process flow diagram. VS is also targeted to both novice analysts with some very fundamental understanding of regression modeling and to a lesser extent advanced users. EM tends to be more targeted more towards advanced data miners and scientists with lots more algorithmic depth.&amp;nbsp;&amp;nbsp;&amp;nbsp; One common use case I am seeing is use VS prior to modeling for feature reduction through the interactivity the product provides using all of the data and then fine tune the model in EM.&amp;nbsp; VS generates SAS score code so you can compare models in EM using the Model Import node.&amp;nbsp; Once use case I am not seeing enough is using VS to do post ad hoc validation of models – say evaluating score band cutoffs interactivity and&amp;nbsp; evaluating high leverage and influence points even with real big data.&amp;nbsp; You may also want to use VS to derive segments interactively and the use the Group Processing feature of EM to do stratified modeling.&amp;nbsp; It is a great question – I hope this helps some.&amp;nbsp;&amp;nbsp; I was the EM product manager for over a decade and now am working on VS.&amp;nbsp;&amp;nbsp; I still love both products and use them all the time. &lt;/P&gt;&lt;/BODY&gt;&lt;/HTML&gt;</description>
      <pubDate>Mon, 29 Jun 2015 18:53:03 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Communities-Library/What-is-SAS-Visual-Statistics/tac-p/223461#M630</guid>
      <dc:creator>WayneThompson</dc:creator>
      <dc:date>2015-06-29T18:53:03Z</dc:date>
    </item>
    <item>
      <title>Re: Difference between boosting through start groups and gradient boosting</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/Difference-between-boosting-through-start-groups-and-gradient/m-p/189933#M2363</link>
      <description>&lt;HTML&gt;&lt;HEAD&gt;&lt;/HEAD&gt;&lt;BODY&gt;&lt;P&gt;&lt;EM&gt;Bagging&lt;/EM&gt; (Breiman 1996) is a common ensemble algorithm, in which you do the following:&lt;/P&gt;&lt;OL style="list-style-type: decimal;"&gt;&lt;LI&gt;Develop separate models on &lt;EM&gt;k&lt;/EM&gt; random samples of the data of about the same size. &lt;/LI&gt;&lt;/OL&gt;&lt;P style="margin-left: .5in;"&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 2. Fit a classification or regression tree to each sample. I tend to bag only trees but the start and end group nodes.&amp;nbsp; allow other algorithms. &lt;/P&gt;&lt;P style="margin-left: .5in;"&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 3. Average or vote to derive the final predictions or classifications.&amp;nbsp; &lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;Boosting&lt;/EM&gt; (Freund and Schapire, 1996) ,also supported through start&amp;nbsp; &amp;gt;&amp;nbsp; tree &amp;gt; end , goes one step further and weights observations that are misclassified in the previous models more heavily for inclusion into subsequent samples.&amp;nbsp; The successive samples are adjusted to accommodate previously computed inaccuracies. &lt;EM&gt;Gradient boosting&lt;/EM&gt; (Friedman 2001) resamples the training data several times to generate results that form a weighted average of the resampled data set.&amp;nbsp; Each tree in the series is fit to the residual of the prediction from the earlier trees in the series. The residual is defined in terms of the derivative of a loss function. For squared error loss and an interval target, the residual is simply the target value minus the predicted value.&amp;nbsp; Because each successive sample is weighted according to the classification accuracy of previous models, this approach is sometimes called &lt;EM&gt;stochastic gradient boosting&lt;/EM&gt;. &lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Random forests is my favorite data mining algorithm especially, when I have little subject knowledge of the application.&amp;nbsp;&amp;nbsp; You grow many large decision trees at random and vote over all trees in the forest. The algorithm works as follows:&lt;/P&gt;&lt;OL style="list-style-type: decimal;"&gt;&lt;LI&gt;You develop random samples of the data and grow &lt;EM&gt;k&lt;/EM&gt; decision trees.&amp;nbsp; The size of &lt;EM&gt;k&lt;/EM&gt; is large, usually greater than or equal to 100.&amp;nbsp;&amp;nbsp; A typical sample size is about two-thirds of the training data. &lt;/LI&gt;&lt;LI&gt;At each split point for each tree you evaluate a random subset of candidate inputs (predictors) are evaluated. You hold the size of the subset constant across all trees. &lt;/LI&gt;&lt;LI&gt;You grow each tree is as large as possible without pruning. &lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;In a random forest this case you are perturbing not only the data but also the variables that are used to construct each tree. The error rate is measured on the remaining holdout data not used for training.&amp;nbsp;&amp;nbsp; This remaining one-third of the data is called the out-of-bag sample.&amp;nbsp; Variable importance can also be inferred based on how often an input was used in the construction of the trees.&amp;nbsp;&amp;nbsp;&amp;nbsp; &lt;/P&gt;&lt;/BODY&gt;&lt;/HTML&gt;</description>
      <pubDate>Thu, 09 Jan 2014 16:17:49 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/Difference-between-boosting-through-start-groups-and-gradient/m-p/189933#M2363</guid>
      <dc:creator>WayneThompson</dc:creator>
      <dc:date>2014-01-09T16:17:49Z</dc:date>
    </item>
    <item>
      <title>Re: Clustering binary data with Enterprise Miner</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/Clustering-binary-data-with-Enterprise-Miner/m-p/115008#M963</link>
      <description>&lt;HTML&gt;&lt;HEAD&gt;&lt;/HEAD&gt;&lt;BODY&gt;&lt;P&gt;HI Sonnfan,&amp;nbsp; &lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;It appears you want to cluster variables and not observations.&amp;nbsp; In that case, you can use variable clustering node, or&lt;BR /&gt;factor analysis (see PROC FACTOR) or princicpal components.&amp;nbsp;&amp;nbsp; If you want to cluster rows, the f&lt;SPAN style="background: white;"&gt;or binary data, the&lt;BR /&gt;Euclidean distance measure used by K-Means is equivalent to counting the number of variables on which two cases disagree. However, you can try one of the&lt;BR /&gt;following approaches:&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;OL style="list-style-type: decimal;"&gt;&lt;LI&gt;&lt;SPAN style="background: white;"&gt;Run proc distance by a selecting the distance type&lt;BR /&gt;that you want and apply clustering using that distance matrix.&lt;/SPAN&gt;&lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN style="background: white;"&gt;OR&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;OL style="list-style-type: decimal;"&gt;&lt;LI&gt;You can project the binary variables and do&lt;BR /&gt;clustering as follows:&lt;/LI&gt;&lt;LI&gt;run Factor Analysis or PCA on the binary&lt;BR /&gt;variables&lt;/LI&gt;&lt;LI&gt; save the factor or component scores as new&lt;BR /&gt;variables &lt;/LI&gt;&lt;LI&gt;cluster on the basis of those scores (In that&lt;BR /&gt;case, the data will no longer be binary)&lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Good luck&lt;/P&gt;&lt;/BODY&gt;&lt;/HTML&gt;</description>
      <pubDate>Tue, 20 Aug 2013 20:10:48 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/Clustering-binary-data-with-Enterprise-Miner/m-p/115008#M963</guid>
      <dc:creator>WayneThompson</dc:creator>
      <dc:date>2013-08-20T20:10:48Z</dc:date>
    </item>
    <item>
      <title>RPM vs. Enterprise Miner</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/RPM-vs-Enterprise-Miner/m-p/22728#M115</link>
      <description>&lt;HTML&gt;&lt;HEAD&gt;&lt;/HEAD&gt;&lt;BODY&gt;&lt;P&gt; Thanks DLing.&amp;nbsp; Here is a older SGF paper that discuss the RPM task within SAS EG and SAS AMO for Excel and how it is integrated and inturn can be extended in EM. &lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;A href="http://support.sas.com/resources/papers/proceedings10/113-2010.pdf"&gt;http://support.sas.com/resources/papers/proceedings10/113-2010.pdf&lt;/A&gt;&lt;/P&gt;&lt;/BODY&gt;&lt;/HTML&gt;</description>
      <pubDate>Mon, 03 Oct 2011 21:36:54 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/RPM-vs-Enterprise-Miner/m-p/22728#M115</guid>
      <dc:creator>WayneThompson</dc:creator>
      <dc:date>2011-10-03T21:36:54Z</dc:date>
    </item>
    <item>
      <title>Decision Tree</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/Decision-Tree/m-p/63473#M386</link>
      <description>&lt;HTML&gt;&lt;HEAD&gt;&lt;/HEAD&gt;&lt;BODY&gt;&lt;P&gt; Prediction=0 in the table refers to the prediction for the selected node. Assuming you are strating with the root node and&amp;nbsp; and&amp;nbsp; you have more 0s than 1s then the prediction classification for the root node is =0.&amp;nbsp;&amp;nbsp; When you split on the root node and continue growing the tree interactivelhy hopefully you are resulting in some nodes (leaves) with prediciton = 1.&amp;nbsp; &lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;In some cases when you have a rare target event (1s) and little if any signal in the data the null root node can result in being the final classification. In this case trying using the inverse priors options under "Decisions" for the Input Data Source node.&amp;nbsp;&amp;nbsp;&amp;nbsp; Or obtain add some additonal predictors if you avaialble.&amp;nbsp; Anyway I could still be of base with your question. Hope this helps. &lt;/P&gt;&lt;/BODY&gt;&lt;/HTML&gt;</description>
      <pubDate>Wed, 14 Sep 2011 14:45:17 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/Decision-Tree/m-p/63473#M386</guid>
      <dc:creator>WayneThompson</dc:creator>
      <dc:date>2011-09-14T14:45:17Z</dc:date>
    </item>
    <item>
      <title>Decision Tree</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/Decision-Tree/m-p/63470#M383</link>
      <description>&lt;HTML&gt;&lt;HEAD&gt;&lt;/HEAD&gt;&lt;BODY&gt;&lt;P&gt; Hi Nikhil,&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;My fault but I probably and not understanding the question well.&lt;/P&gt;&lt;P&gt;When you run similar startup code for the project:&lt;/P&gt;&lt;P&gt;%let EM_INTERACTIVE_TREE_MAXOBS=100000;&lt;/P&gt;&lt;P&gt;%let EM_INTERACTIVE_TREE_SAMPLEMETHOD=RANDOM;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;And you are modeling a binary response, do you have 1's and 0's&amp;nbsp; distributed in the root node?&lt;/P&gt;&lt;P&gt;Is your target variable indeed a binary target and set to the binary variable role in the input node or is it set as interval?&amp;nbsp; &lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;&lt;/BODY&gt;&lt;/HTML&gt;</description>
      <pubDate>Tue, 13 Sep 2011 18:36:11 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/Decision-Tree/m-p/63470#M383</guid>
      <dc:creator>WayneThompson</dc:creator>
      <dc:date>2011-09-13T18:36:11Z</dc:date>
    </item>
    <item>
      <title>Decision Tree</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/Decision-Tree/m-p/63467#M380</link>
      <description>&lt;P&gt;Some users may wish to override default Enterprise Miner interactive decision tree sampling strategies. Enterprise Miner provides two macros that you can issue with your project startup code that will modify interactive decision tree input data sampling behaviors:&lt;/P&gt;
&lt;P&gt;&lt;A name="p19pw9kub55y3bn19wc13g2txyr4" target="_blank"&gt;&lt;/A&gt;&lt;/P&gt;
&lt;PRE&gt;%let EM_INTERACTIVE_TREE_MAXOBS= &amp;lt;max-number-of-observations-in-sample&amp;gt;;&lt;BR /&gt;%let EM_INTERACTIVE_TREE_SAMPLEMETHOD=&amp;lt;RANDOM | FIRSTN&amp;gt;;&lt;/PRE&gt;
&lt;P&gt;The first macro specifies the maximum number of observations that can exist in an Interactive Decision Tree node sample. You use this macro if you want to manually control the sample size. Otherwise, Enterprise Miner will use its own algorithms to perform sampling for your interactive decision tree.&lt;/P&gt;
&lt;P&gt;The second macro specifies the sampling methodology that will be used to create an Interactive Decision Tree node sample. You can use this macro if you want to manually control the methodology Enterprise Miner uses to create interactive decision tree samples. By default, Enterprise Miner uses random sampling for interactive decision trees. You can use the macro to choose between RANDOM and FIRSTN sample creation. You use the EM_INTERACTIVE_TREE_MAXOBS macro to specify the number of observations for both RANDOM and FIRSTN sampling strategies.&lt;/P&gt;</description>
      <pubDate>Fri, 07 Jul 2017 17:55:13 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/Decision-Tree/m-p/63467#M380</guid>
      <dc:creator>WayneThompson</dc:creator>
      <dc:date>2017-07-07T17:55:13Z</dc:date>
    </item>
    <item>
      <title>Survival node in Enterprise Miner</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/Survival-node-in-Enterprise-Miner/m-p/62402#M371</link>
      <description>&lt;HTML&gt;&lt;HEAD&gt;&lt;/HEAD&gt;&lt;BODY&gt;&lt;P&gt;By default the last 1/4 of the interval is used for validation.&amp;nbsp; The help below defines further how to specify a user defined interval for validatoin. &lt;/P&gt;&lt;H3&gt;Survival Node Train Properties: Survival Validation &lt;/H3&gt;&lt;P&gt;&lt;A name="p0fxtt3hqptm6jn1df7pzih9u2sx"&gt;&lt;/A&gt;The Survival Validation section of the Survival node Properties panel provides the following settings for configuring the validation of your Survival model: &lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;&lt;STRONG&gt;Survival Validation Method&lt;/STRONG&gt; – Use the&lt;STRONG&gt; Survival Validation Method&lt;/STRONG&gt; property to specify how the validation holdout sample is generated. By default, the training, validation and test data sets that are passed to the Survival node are split into two time intervals. The first 75% of the interval is set aside as training data, for model creation. The remaining portion of the interval is set aside for survival model validation, or scoring. When the default survival validation method is selected, the entire scoring interval is used for validation calculations. However, you can choose to create a user-specified time interval if you desire. When you select user-specified time intervals, you can specify a specific hypothetical scoring date as well as the interval length. &lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Scoring Time&lt;/STRONG&gt; – When you select User Specified as the value for your &lt;STRONG&gt;Survival Validation Method&lt;/STRONG&gt; setting, you can use the &lt;STRONG&gt;Scoring Time&lt;/STRONG&gt; property to specify a hypothetical scoring date. The hypothetical scoring date will divide the data into the two subsets of data. Select the button to the right of the &lt;STRONG&gt;Scoring Time&lt;/STRONG&gt; property to open a table that you use to specify the date value for scoring. The &lt;STRONG&gt;Scoring Time&lt;/STRONG&gt; date value that you specify must fall between your defined start date and censor date. &lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Interval Length&lt;/STRONG&gt; – When you select &lt;EM&gt;User Specified&lt;/EM&gt; as the value for your &lt;STRONG&gt;Survival Validation Method&lt;/STRONG&gt;, the &lt;STRONG&gt;Interval Length&lt;/STRONG&gt; property specifies the length of the time interval used to perform survival validation. The interval will begin with the hypothetical scoring date that you specified in the &lt;STRONG&gt;Scoring Time &lt;/STRONG&gt;property setting, and then extend the specified number of time intervals forward. If the &lt;STRONG&gt;Time Unit&lt;/STRONG&gt; property is set to &lt;EM&gt;Day&lt;/EM&gt; and the &lt;STRONG&gt;Interval Length&lt;/STRONG&gt; property is set to &lt;EM&gt;15&lt;/EM&gt;, your scoring interval is the 15 days that follow the hypothetical scoring date. &lt;/LI&gt;&lt;/UL&gt;&lt;/BODY&gt;&lt;/HTML&gt;</description>
      <pubDate>Fri, 09 Sep 2011 18:50:26 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/Survival-node-in-Enterprise-Miner/m-p/62402#M371</guid>
      <dc:creator>WayneThompson</dc:creator>
      <dc:date>2011-09-09T18:50:26Z</dc:date>
    </item>
    <item>
      <title>Enterprise Miner 7.1 SAS9.3 Released</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/Enterprise-Miner-7-1-SAS9-3-Released/m-p/50584#M291</link>
      <description>&lt;HTML&gt;&lt;HEAD&gt;&lt;/HEAD&gt;&lt;BODY&gt;&lt;P&gt;Some you may already know but EM 7.1 SAS9.3 was released July 12th.&amp;nbsp; The pubs team prepared a nice summary here on support.sas.com&lt;/P&gt;&lt;P&gt;&lt;A href="http://support.sas.com/documentation/cdl/en/whatsnew/64209/HTML/default/viewer.htm#emdocwhatsnew71.htm"&gt;http://support.sas.com/documentation/cdl/en/whatsnew/64209/HTML/default/viewer.htm#emdocwhatsnew71.htm&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Good luck with your data mining projects. &lt;/P&gt;&lt;/BODY&gt;&lt;/HTML&gt;</description>
      <pubDate>Thu, 18 Aug 2011 14:33:08 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/Enterprise-Miner-7-1-SAS9-3-Released/m-p/50584#M291</guid>
      <dc:creator>WayneThompson</dc:creator>
      <dc:date>2011-08-18T14:33:08Z</dc:date>
    </item>
    <item>
      <title>What happens when WARN is M or MU</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/What-happens-when-WARN-is-M-or-MU/m-p/46156#M272</link>
      <description>&lt;HTML&gt;&lt;HEAD&gt;&lt;/HEAD&gt;&lt;BODY&gt;&lt;P&gt; This means missing inputs and unknown inputs.&amp;nbsp; Please try to check your scoring table input vars. &lt;/P&gt;&lt;/BODY&gt;&lt;/HTML&gt;</description>
      <pubDate>Thu, 18 Aug 2011 14:28:03 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/What-happens-when-WARN-is-M-or-MU/m-p/46156#M272</guid>
      <dc:creator>WayneThompson</dc:creator>
      <dc:date>2011-08-18T14:28:03Z</dc:date>
    </item>
    <item>
      <title>SAS Rapid Predictive Modeler Documentation</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/SAS-Rapid-Predictive-Modeler-Documentation/m-p/41301#M234</link>
      <description>&lt;HTML&gt;&lt;HEAD&gt;&lt;/HEAD&gt;&lt;BODY&gt;&lt;P&gt;When you open the Rapid Predictive Modeler task select the help button on the task.&amp;nbsp; Totally agree that we should add the help into the main EG online help.&amp;nbsp;&amp;nbsp; Here is an SGF paper &lt;A href="http://support.sas.com/resources/papers/proceedings10/113-2010.pdf"&gt;http://support.sas.com/resources/papers/proceedings10/113-2010.pdf&lt;/A&gt;&lt;/P&gt;&lt;P&gt;The best way to learn how the models are developed is to run them from EG (AMO ) and then review them in EM.&amp;nbsp; Good luck. &lt;/P&gt;&lt;/BODY&gt;&lt;/HTML&gt;</description>
      <pubDate>Thu, 18 Aug 2011 14:25:04 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/SAS-Rapid-Predictive-Modeler-Documentation/m-p/41301#M234</guid>
      <dc:creator>WayneThompson</dc:creator>
      <dc:date>2011-08-18T14:25:04Z</dc:date>
    </item>
    <item>
      <title>Re: renaming variables in E Miner</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/renaming-variables-in-E-Miner/m-p/14819#M62</link>
      <description>Please also see this tech support note &lt;A href="http://support.sas.com/kb/38/578.html" target="_blank"&gt;http://support.sas.com/kb/38/578.html&lt;/A&gt;&lt;BR /&gt;
&lt;BR /&gt;
As you indicate I would use the SAS Code node to rename the variables  using the macros varaible names for the data sources you want to import and export. &lt;BR /&gt;
&lt;BR /&gt;
SAS has a 32 character limit.&lt;BR /&gt;
Teradata has a 30 character limit.&lt;BR /&gt;
Long variable names is something we look into.</description>
      <pubDate>Fri, 17 Jun 2011 16:22:00 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/renaming-variables-in-E-Miner/m-p/14819#M62</guid>
      <dc:creator>WayneThompson</dc:creator>
      <dc:date>2011-06-17T16:22:00Z</dc:date>
    </item>
    <item>
      <title>Re: Exterprise Miner Error</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/Exterprise-Miner-Error/m-p/59113#M344</link>
      <description>Thanks for using SAS EM.  Please see this usage note about max levels exceed 512.  &lt;A href="http://support.sas.com/kb/20/054.html" target="_blank"&gt;http://support.sas.com/kb/20/054.html&lt;/A&gt;&lt;BR /&gt;
&lt;BR /&gt;
&lt;BR /&gt;
Also as suggested I would recommend being careful about using a variable like zip code or other nominal variables that have a really large number of discrete nominal values as predictors (inputs).  You can reject this variable or using a binning technique.</description>
      <pubDate>Fri, 29 Apr 2011 14:38:15 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/Exterprise-Miner-Error/m-p/59113#M344</guid>
      <dc:creator>WayneThompson</dc:creator>
      <dc:date>2011-04-29T14:38:15Z</dc:date>
    </item>
    <item>
      <title>Rexler Data Mining Survey</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/Rexler-Data-Mining-Survey/m-p/53957#M301</link>
      <description>Karl Rexler has asked me to let folks from the SAS DM community  know about his Data Miner Survey which cover trends in data mining, applications and vendor tool usage. &lt;BR /&gt;
 &lt;BR /&gt;
If you are interested in the survey it should take approximately 20 minutes to complete.   Survey Link:&lt;BR /&gt;
 &lt;A href="http://www.RexerAnalytics.com/Data-Miner-Survey-2011-Intro2.html" target="_blank"&gt;www.RexerAnalytics.com/Data-Miner-Survey-2011-Intro2.html&lt;/A&gt;&lt;BR /&gt;
Access Code:  KS37P&lt;BR /&gt;
&lt;BR /&gt;
thanks much</description>
      <pubDate>Wed, 20 Apr 2011 19:54:06 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/Rexler-Data-Mining-Survey/m-p/53957#M301</guid>
      <dc:creator>WayneThompson</dc:creator>
      <dc:date>2011-04-20T19:54:06Z</dc:date>
    </item>
    <item>
      <title>Re: Neural Network Graphics</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/Neural-Network-Graphics/m-p/44148#M256</link>
      <description>An interactive neural network buidler to define and visualize the network is a feature we would like to add to EM.   thanks</description>
      <pubDate>Fri, 03 Dec 2010 22:07:59 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/Neural-Network-Graphics/m-p/44148#M256</guid>
      <dc:creator>WayneThompson</dc:creator>
      <dc:date>2010-12-03T22:07:59Z</dc:date>
    </item>
    <item>
      <title>Re: Kernals in Support Vector Machines</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/Kernals-in-Support-Vector-Machines/m-p/31963#M170</link>
      <description>An experimenta linear and nonlinear kernel SVM node is planned for EM 7.1 SAS9.3 mid next year. Thanks for you use of the software. There are quite a few other classification and prediction tools you will want to try.  Some users have reported good Gradient Boosting results. Will keep you and the forum up to date.</description>
      <pubDate>Wed, 10 Nov 2010 22:31:04 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/Kernals-in-Support-Vector-Machines/m-p/31963#M170</guid>
      <dc:creator>WayneThompson</dc:creator>
      <dc:date>2010-11-10T22:31:04Z</dc:date>
    </item>
    <item>
      <title>Re: Data Mining without using EM</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/Data-Mining-without-using-EM/m-p/70581#M438</link>
      <description>Hello Aha,&lt;BR /&gt;
&lt;BR /&gt;
Please contact me at wayne.thompson@sas.com to discuss EM pricing relative to your budget and application needs. I am here to listen to your concerns at a minimum.  &lt;BR /&gt;
&lt;BR /&gt;
STAT provides a broad range of predictive/classifciation methods along with clustering and many other techniques.  EM provide some algorithms like decision trees, gradient boosting, neural networks, memory based reasoning , etc not found in STAT along with the GUI for  developing repeatable self documenting analysis very quickly with an emphasis on complete score code generation for deployment.  Others  from the list may have more practical reasons -- I work at SAS.   Anway we hope you are using SAS and STAT is a great product --- EM requires it and includes the SAS code node for embeding STAT procedures into your EM analysis.  &lt;BR /&gt;
&lt;BR /&gt;
Wayne</description>
      <pubDate>Tue, 07 Sep 2010 18:38:07 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/Data-Mining-without-using-EM/m-p/70581#M438</guid>
      <dc:creator>WayneThompson</dc:creator>
      <dc:date>2010-09-07T18:38:07Z</dc:date>
    </item>
    <item>
      <title>Enterprise Miner 6.2 SAS 9.2 released August 17th</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/Enterprise-Miner-6-2-SAS-9-2-released-August-17th/m-p/66572#M398</link>
      <description>SAS Enterprise Miner 6.2 was released last week which delivers major new features targeted at specific business applications and complex information technology deployments. Rapid Predictive Modeer is a new feature for general business users that need to develop reliable models for predicting customer response and retention. The Interactive Decision Tree is enhanced to display more information about the nodes and leaves and to show more plots based on validation data. The credit scoring functions have been improved to provide more customization of bins and to provide more control over scorecard points (Credit Scoring for EM 6.2). For more details please see the What's new document at &lt;BR /&gt;
&lt;BR /&gt;
&lt;A href="http://support.sas.com/documentation/cdl/en/whatsnew/62580/HTML/default/viewer.htm#/documentation/cdl/en/whatsnew/62580/HTML/default/emwhatsnew62.htm" target="_blank"&gt;http://support.sas.com/documentation/cdl/en/whatsnew/62580/HTML/default/viewer.htm#/documentation/cdl/en/whatsnew/62580/HTML/default/emwhatsnew62.htm&lt;/A&gt;&lt;BR /&gt;
&lt;BR /&gt;
We hope you find these features useful and fun to use.  Wayne Thompson SAS Product Manager Analytics.</description>
      <pubDate>Mon, 23 Aug 2010 20:44:13 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/Enterprise-Miner-6-2-SAS-9-2-released-August-17th/m-p/66572#M398</guid>
      <dc:creator>WayneThompson</dc:creator>
      <dc:date>2010-08-23T20:44:13Z</dc:date>
    </item>
    <item>
      <title>Re: Score or points allocations after building logistic regression</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/Score-or-points-allocations-after-building-logistic-regression/m-p/15014#M67</link>
      <description>&lt;P&gt;Use the Credit Scoring for EM product which provides variable selection, interactive or batch classing (binning) including weights of evidence calculations, scorecard construction with scaling and diagnostics, and reject inference.&lt;/P&gt;</description>
      <pubDate>Fri, 07 Jul 2017 19:12:24 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/Score-or-points-allocations-after-building-logistic-regression/m-p/15014#M67</guid>
      <dc:creator>WayneThompson</dc:creator>
      <dc:date>2017-07-07T19:12:24Z</dc:date>
    </item>
  </channel>
</rss>

