<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic reading data from huge flatfile in SAS Programming</title>
    <link>https://communities.sas.com/t5/SAS-Programming/reading-data-from-huge-flatfile/m-p/141302#M261538</link>
    <description>&lt;HTML&gt;&lt;HEAD&gt;&lt;/HEAD&gt;&lt;BODY&gt;&lt;P&gt;I have flatfile of size 247 gb .. with help of this flatfile I am creating 8 different datasets which happens to read the sequentially into datasets one after the other .. so overall time thats taking is around 23 hrs to complete my job. As this file exists in unix enviromnent .. intially I thought of using split command and to divide the file&amp;nbsp; but issue here is its going to take lot of space on the server .. I would really appreciate if anyone could suggest better approch to reduce the time .. I am thinking for a solution where I can reduce time by reading data parallelism type ...&lt;/P&gt;&lt;P&gt;Thanks in advance&lt;/P&gt;&lt;P&gt;Mvk&amp;nbsp; &lt;/P&gt;&lt;/BODY&gt;&lt;/HTML&gt;</description>
    <pubDate>Wed, 12 Mar 2014 17:56:36 GMT</pubDate>
    <dc:creator>MVK</dc:creator>
    <dc:date>2014-03-12T17:56:36Z</dc:date>
    <item>
      <title>reading data from huge flatfile</title>
      <link>https://communities.sas.com/t5/SAS-Programming/reading-data-from-huge-flatfile/m-p/141302#M261538</link>
      <description>&lt;HTML&gt;&lt;HEAD&gt;&lt;/HEAD&gt;&lt;BODY&gt;&lt;P&gt;I have flatfile of size 247 gb .. with help of this flatfile I am creating 8 different datasets which happens to read the sequentially into datasets one after the other .. so overall time thats taking is around 23 hrs to complete my job. As this file exists in unix enviromnent .. intially I thought of using split command and to divide the file&amp;nbsp; but issue here is its going to take lot of space on the server .. I would really appreciate if anyone could suggest better approch to reduce the time .. I am thinking for a solution where I can reduce time by reading data parallelism type ...&lt;/P&gt;&lt;P&gt;Thanks in advance&lt;/P&gt;&lt;P&gt;Mvk&amp;nbsp; &lt;/P&gt;&lt;/BODY&gt;&lt;/HTML&gt;</description>
      <pubDate>Wed, 12 Mar 2014 17:56:36 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Programming/reading-data-from-huge-flatfile/m-p/141302#M261538</guid>
      <dc:creator>MVK</dc:creator>
      <dc:date>2014-03-12T17:56:36Z</dc:date>
    </item>
    <item>
      <title>Re: reading data from huge flatfile</title>
      <link>https://communities.sas.com/t5/SAS-Programming/reading-data-from-huge-flatfile/m-p/141303#M261539</link>
      <description>&lt;HTML&gt;&lt;HEAD&gt;&lt;/HEAD&gt;&lt;BODY&gt;&lt;P&gt;Check the admin of your system, they probably will be able to give more practical advice.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Reading that large of a file will take a long time, but writing is most likely the bottleneck. Probably the biggest improvement will be in using as many different physical disks as possible. &lt;SPAN style="font-size: 10pt; line-height: 1.5em;"&gt;So replicate your program 8 times with each one creating one of the output files and run it in parallel , but make sure to write to different output disks.&amp;nbsp; &lt;/SPAN&gt;&lt;/P&gt;&lt;/BODY&gt;&lt;/HTML&gt;</description>
      <pubDate>Wed, 12 Mar 2014 18:18:44 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Programming/reading-data-from-huge-flatfile/m-p/141303#M261539</guid>
      <dc:creator>Tom</dc:creator>
      <dc:date>2014-03-12T18:18:44Z</dc:date>
    </item>
  </channel>
</rss>

