<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Interval Targets from Gradient Boosting in SAS Data Science</title>
    <link>https://communities.sas.com/t5/SAS-Data-Science/Interval-Targets-from-Gradient-Boosting/m-p/353884#M5255</link>
    <description>&lt;P&gt;I don't believe the model is predicting the values (negatives) poorly as the observations' actual values are usually 0 or near 0.&lt;BR /&gt;&lt;BR /&gt;The negative predictions seem to be the result of the gradient boosting node's underlying algorithm. Since the boosting node constructs an additive regression model by sequentially fitting a base-learner to current pseudo residuals at each iteration, the final model is linear. This explains why the negatives are occurring despite there being no actual negative values.&lt;BR /&gt;&lt;BR /&gt;I believe that performing a log transformation prior to modeling (like I tried) is not possible being that the the final linear model is based on the pseudo-residuals of the base-learning trees. I guess I was looking for further explanation on this, and if another transformation may work to force positive predictions (I don't think so).&lt;BR /&gt;&lt;BR /&gt;Last resort, would be to truncate negatives to 0, but that seems like it may be the only option.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thank you.&lt;/P&gt;</description>
    <pubDate>Wed, 26 Apr 2017 20:42:24 GMT</pubDate>
    <dc:creator>JFlyers00</dc:creator>
    <dc:date>2017-04-26T20:42:24Z</dc:date>
    <item>
      <title>Interval Targets from Gradient Boosting</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/Interval-Targets-from-Gradient-Boosting/m-p/352333#M5232</link>
      <description>&lt;P&gt;Hi, &amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I was wondering if there is a way to force the gradient boosting node to always produce a positive predictions for an interval target. &amp;nbsp;The target is not negative for the training data, and it could never be negative realisitcally. &amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I understand why the model predicts negative values, and I have tried doing a log transformation and exponentiating the predictions. &amp;nbsp;That did not work and the resulting model was poor. &amp;nbsp;I can elaborate more if needed. &amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thank you,&lt;/P&gt;&lt;P&gt;James&lt;/P&gt;</description>
      <pubDate>Fri, 21 Apr 2017 20:46:18 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/Interval-Targets-from-Gradient-Boosting/m-p/352333#M5232</guid>
      <dc:creator>JFlyers00</dc:creator>
      <dc:date>2017-04-21T20:46:18Z</dc:date>
    </item>
    <item>
      <title>Re: Interval Targets from Gradient Boosting</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/Interval-Targets-from-Gradient-Boosting/m-p/353746#M5253</link>
      <description>&lt;P&gt;In general, boosting produces predictions that are out of range because the model predicts those observations poorly (was that too obvious?). &amp;nbsp;Better predictions sometimes result from increasing the number of trees while decreasing the learning rate (the SHRINKAGE= parameter). &amp;nbsp;Other than that, no, there is no option within the boosting algorithm. &amp;nbsp;The predictions would have to be post-processed, perhaps simply by truncating negative predictions to 0.&lt;/P&gt;
&lt;P&gt;-Padraic&lt;/P&gt;</description>
      <pubDate>Wed, 26 Apr 2017 15:05:52 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/Interval-Targets-from-Gradient-Boosting/m-p/353746#M5253</guid>
      <dc:creator>PadraicGNeville</dc:creator>
      <dc:date>2017-04-26T15:05:52Z</dc:date>
    </item>
    <item>
      <title>Re: Interval Targets from Gradient Boosting</title>
      <link>https://communities.sas.com/t5/SAS-Data-Science/Interval-Targets-from-Gradient-Boosting/m-p/353884#M5255</link>
      <description>&lt;P&gt;I don't believe the model is predicting the values (negatives) poorly as the observations' actual values are usually 0 or near 0.&lt;BR /&gt;&lt;BR /&gt;The negative predictions seem to be the result of the gradient boosting node's underlying algorithm. Since the boosting node constructs an additive regression model by sequentially fitting a base-learner to current pseudo residuals at each iteration, the final model is linear. This explains why the negatives are occurring despite there being no actual negative values.&lt;BR /&gt;&lt;BR /&gt;I believe that performing a log transformation prior to modeling (like I tried) is not possible being that the the final linear model is based on the pseudo-residuals of the base-learning trees. I guess I was looking for further explanation on this, and if another transformation may work to force positive predictions (I don't think so).&lt;BR /&gt;&lt;BR /&gt;Last resort, would be to truncate negatives to 0, but that seems like it may be the only option.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thank you.&lt;/P&gt;</description>
      <pubDate>Wed, 26 Apr 2017 20:42:24 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Data-Science/Interval-Targets-from-Gradient-Boosting/m-p/353884#M5255</guid>
      <dc:creator>JFlyers00</dc:creator>
      <dc:date>2017-04-26T20:42:24Z</dc:date>
    </item>
  </channel>
</rss>

