<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Proc Logistics gone CRAZY! in Statistical Procedures</title>
    <link>https://communities.sas.com/t5/Statistical-Procedures/Proc-Logistics-gone-CRAZY/m-p/146721#M7714</link>
    <description>&lt;HTML&gt;&lt;HEAD&gt;&lt;/HEAD&gt;&lt;BODY&gt;&lt;P&gt;Thank you very much Steve.&lt;/P&gt;&lt;/BODY&gt;&lt;/HTML&gt;</description>
    <pubDate>Thu, 19 Jun 2014 16:16:17 GMT</pubDate>
    <dc:creator>niam</dc:creator>
    <dc:date>2014-06-19T16:16:17Z</dc:date>
    <item>
      <title>Proc Logistics gone CRAZY!</title>
      <link>https://communities.sas.com/t5/Statistical-Procedures/Proc-Logistics-gone-CRAZY/m-p/146717#M7710</link>
      <description>&lt;HTML&gt;&lt;HEAD&gt;&lt;/HEAD&gt;&lt;BODY&gt;&lt;P&gt;I have a small data set with only 20 observations. &lt;/P&gt;&lt;P&gt;N= number of trials&lt;/P&gt;&lt;P&gt;Y= number of successes&lt;/P&gt;&lt;P&gt;Here is the data-set:&lt;/P&gt;&lt;P&gt;data temp;&lt;/P&gt;&lt;P&gt;input Y N;&lt;/P&gt;&lt;P&gt;cards;&lt;/P&gt;&lt;P&gt;97870&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 270000&lt;/P&gt;&lt;P&gt;12890&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 55000&lt;/P&gt;&lt;P&gt;120071&amp;nbsp;&amp;nbsp;&amp;nbsp; 313000&lt;/P&gt;&lt;P&gt;43446&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 150000&lt;/P&gt;&lt;P&gt;1405102&amp;nbsp;&amp;nbsp; 1903000&lt;/P&gt;&lt;P&gt;125402&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 254000&lt;/P&gt;&lt;P&gt;79192&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 109000&lt;/P&gt;&lt;P&gt;14087&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 29000&lt;/P&gt;&lt;P&gt;10714&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 9000&lt;/P&gt;&lt;P&gt;983775&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 1587000&lt;/P&gt;&lt;P&gt;316543&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 654000&lt;/P&gt;&lt;P&gt;8592&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 29000&lt;/P&gt;&lt;P&gt;76061&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 130000&lt;/P&gt;&lt;P&gt;217492&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 501000&lt;/P&gt;&lt;P&gt;132423&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 354000&lt;/P&gt;&lt;P&gt;29163&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 127000&lt;/P&gt;&lt;P&gt;57013&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 161000&lt;/P&gt;&lt;P&gt;82747&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 192000&lt;/P&gt;&lt;P&gt;101778&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 344000&lt;/P&gt;&lt;P&gt;44258&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 77000&lt;/P&gt;&lt;P&gt;;&lt;/P&gt;&lt;P&gt;run;&lt;/P&gt;&lt;P&gt;If I create a bunch of totally irrelevant random variable such that:&lt;/P&gt;&lt;P&gt;proc temp;&lt;/P&gt;&lt;P&gt;set temp;&lt;/P&gt;&lt;P&gt;x1=rand('BETA',3,0.1);&lt;/P&gt;&lt;P&gt;x2=rand('CAUCHY');&lt;/P&gt;&lt;P&gt;x3=rand('CHISQ',22);&lt;/P&gt;&lt;P&gt;x4=rand('ERLANG', 7);&lt;/P&gt;&lt;P&gt;x5=rand('EXPO');&lt;/P&gt;&lt;P&gt;x6=rand('F',12,322);&lt;/P&gt;&lt;P&gt;run;&lt;/P&gt;&lt;P&gt;And then run a LOGISTIC regression such as:&lt;/P&gt;&lt;P&gt;Proc Logistic data=temp;&lt;/P&gt;&lt;P&gt;model y/n=x1 x2 x3 x4 x5 x6;&lt;/P&gt;&lt;P&gt;run;&lt;/P&gt;&lt;P&gt;all of the independent variables are statistically significant at p&amp;lt;0.0001 level! despite the fact that they are all random and none of them should logically be significant! I tried it with so many other variables, but it is almost impossible to get an INSIGNIFICANT results from the proc logistic with events/trials option!&lt;/P&gt;&lt;P&gt;Do you know why this is happening? &lt;/P&gt;&lt;/BODY&gt;&lt;/HTML&gt;</description>
      <pubDate>Wed, 18 Jun 2014 18:09:03 GMT</pubDate>
      <guid>https://communities.sas.com/t5/Statistical-Procedures/Proc-Logistics-gone-CRAZY/m-p/146717#M7710</guid>
      <dc:creator>niam</dc:creator>
      <dc:date>2014-06-18T18:09:03Z</dc:date>
    </item>
    <item>
      <title>Re: Proc Logistics gone CRAZY!</title>
      <link>https://communities.sas.com/t5/Statistical-Procedures/Proc-Logistics-gone-CRAZY/m-p/146718#M7711</link>
      <description>&lt;HTML&gt;&lt;HEAD&gt;&lt;/HEAD&gt;&lt;BODY&gt;&lt;P&gt;You have an overpowered analysis.&amp;nbsp; With over 7 million trials, the Wald standard error is going to be very small, and as a result, the Wald chi square very large..&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Try the following:&lt;/P&gt;&lt;P&gt;data temp;&lt;/P&gt;&lt;P&gt;call streaminit(123);&lt;/P&gt;&lt;P&gt;set temp;&lt;/P&gt;&lt;P&gt;x1=1-rand('BETA',3,0.1);&lt;/P&gt;&lt;P&gt;x2=rand('CAUCHY');&lt;/P&gt;&lt;P&gt;x3=rand('CHISQ',22);&lt;/P&gt;&lt;P&gt;x4=rand('ERLANG', 7);&lt;/P&gt;&lt;P&gt;x5=rand('EXPO');&lt;/P&gt;&lt;P&gt;x6=rand('F',12,322);&lt;/P&gt;&lt;P&gt;trial=_n_;&lt;/P&gt;&lt;P&gt;run;&lt;/P&gt;&lt;P&gt;proc glimmix data=temp method=laplace;&lt;/P&gt;&lt;P&gt;class trial;&lt;/P&gt;&lt;P&gt;model y/n=x1 x2 x3 x4 x5/solution chisquare;&lt;/P&gt;&lt;P&gt;random intercept/subject=trial;&lt;/P&gt;&lt;P&gt;run;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;You will see that none of the type III tests, using the chi-squared value are significant.&amp;nbsp; Be careful about just summing across trials. &lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Steve Denham&lt;/P&gt;&lt;/BODY&gt;&lt;/HTML&gt;</description>
      <pubDate>Wed, 18 Jun 2014 18:44:07 GMT</pubDate>
      <guid>https://communities.sas.com/t5/Statistical-Procedures/Proc-Logistics-gone-CRAZY/m-p/146718#M7711</guid>
      <dc:creator>SteveDenham</dc:creator>
      <dc:date>2014-06-18T18:44:07Z</dc:date>
    </item>
    <item>
      <title>Re: Proc Logistics gone CRAZY!</title>
      <link>https://communities.sas.com/t5/Statistical-Procedures/Proc-Logistics-gone-CRAZY/m-p/146719#M7712</link>
      <description>&lt;HTML&gt;&lt;HEAD&gt;&lt;/HEAD&gt;&lt;BODY&gt;&lt;P&gt;Thank you very much for the great help;&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;would you please elaborate on the random intercept option of the glimmix procedure?&amp;nbsp; are you assuming that each of the observations have unobserved characteristics and would like to adjust for this by including a random intercept?&lt;/LI&gt;&lt;LI&gt;&lt;SPAN style="font-size: 10pt; line-height: 1.5em;"&gt;How about dividing both the number of trials and evens by say 100,000. This will keep the proportion the same, but reduces the total number of trials. Will it solve the problem with Proc logistic?&lt;/SPAN&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;Thanks&lt;/P&gt;&lt;/BODY&gt;&lt;/HTML&gt;</description>
      <pubDate>Wed, 18 Jun 2014 19:35:37 GMT</pubDate>
      <guid>https://communities.sas.com/t5/Statistical-Procedures/Proc-Logistics-gone-CRAZY/m-p/146719#M7712</guid>
      <dc:creator>niam</dc:creator>
      <dc:date>2014-06-18T19:35:37Z</dc:date>
    </item>
    <item>
      <title>Re: Proc Logistics gone CRAZY!</title>
      <link>https://communities.sas.com/t5/Statistical-Procedures/Proc-Logistics-gone-CRAZY/m-p/146720#M7713</link>
      <description>&lt;HTML&gt;&lt;HEAD&gt;&lt;/HEAD&gt;&lt;BODY&gt;&lt;P&gt;Q1: I am assuming that each observation is a random sample from a universe of possible observations.&amp;nbsp; Ordinarily, I would just state:&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;random trial;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;but in order to use method=laplace, a subject must be specified, which in this case means rewriting as&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;random intercept/subject=trial;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;In GLIMMIX, these are equivalent, and no implication on intercept needs to be made.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Q2: Dividing by 100,000 would reduce the total to about 72, so instead I divided by 10,000 and used the floor function to give integer values.&amp;nbsp; The p values I obtained were x1 (0.0689), x2 (0.0233), x3 (0.0372), x4 (0.4644) and x5 (0.1134).&amp;nbsp; This strongly implies that the very small p values obtained in the first analysis are due to overpowering the analysis, as LOGISTIC "consolidates" all values when calculating standard errors.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Steve Denham&lt;/P&gt;&lt;/BODY&gt;&lt;/HTML&gt;</description>
      <pubDate>Thu, 19 Jun 2014 13:44:22 GMT</pubDate>
      <guid>https://communities.sas.com/t5/Statistical-Procedures/Proc-Logistics-gone-CRAZY/m-p/146720#M7713</guid>
      <dc:creator>SteveDenham</dc:creator>
      <dc:date>2014-06-19T13:44:22Z</dc:date>
    </item>
    <item>
      <title>Re: Proc Logistics gone CRAZY!</title>
      <link>https://communities.sas.com/t5/Statistical-Procedures/Proc-Logistics-gone-CRAZY/m-p/146721#M7714</link>
      <description>&lt;HTML&gt;&lt;HEAD&gt;&lt;/HEAD&gt;&lt;BODY&gt;&lt;P&gt;Thank you very much Steve.&lt;/P&gt;&lt;/BODY&gt;&lt;/HTML&gt;</description>
      <pubDate>Thu, 19 Jun 2014 16:16:17 GMT</pubDate>
      <guid>https://communities.sas.com/t5/Statistical-Procedures/Proc-Logistics-gone-CRAZY/m-p/146721#M7714</guid>
      <dc:creator>niam</dc:creator>
      <dc:date>2014-06-19T16:16:17Z</dc:date>
    </item>
  </channel>
</rss>

