<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: running multiple SAS programs in parallel using SLURM Batch script in SAS Programming</title>
    <link>https://communities.sas.com/t5/SAS-Programming/running-multiple-SAS-programs-in-parallel-using-SLURM-Batch/m-p/497198#M131702</link>
    <description>&lt;P&gt;Thanks a lot for your quick response. Ignoring issue 1 (which I wish to take care of by splitting the original data file into yearly files) is this a decent method of parallel processing sas programs?&lt;/P&gt;</description>
    <pubDate>Wed, 19 Sep 2018 23:14:52 GMT</pubDate>
    <dc:creator>raulroy</dc:creator>
    <dc:date>2018-09-19T23:14:52Z</dc:date>
    <item>
      <title>running multiple SAS programs in parallel using SLURM Batch script</title>
      <link>https://communities.sas.com/t5/SAS-Programming/running-multiple-SAS-programs-in-parallel-using-SLURM-Batch/m-p/497171#M131683</link>
      <description>&lt;P&gt;Is the following an efficient way of running several sas programs in parallel?&amp;nbsp;&lt;/P&gt;&lt;P&gt;Each program uses the same data and filters it by year 2006,2007...ect. and runs independent regression like anslysis on them. The outputs are eported as CSV files with year indexed names.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;PRE&gt;&lt;CODE class=" language-sas"&gt;#!/bin/bash

#SBATCH -t 4:00:00
#SBATCH --job-name=SAS01
#SBATCH -N 1
#SBATCH -n 16
#SBATCH --partition=Bigg


. /etc/profile.d/modules.sh
echo "Job running on SLURM NODELIST: $SLURM_NODELIST "


# Modules needed for this SAS job

module purge

module load SAS

#SAS Program execution command

sas /home/user1/SASprog/prog1.sas -sysparm '2006' -log /home/user1//log/proglog2006.log &amp;amp;
sas /home/user1/SASprog/prog1.sas -sysparm '2007' -log /home/user1//log/proglog2007.log &amp;amp;
sas /home/user1/SASprog/prog1.sas -sysparm '2008' -log /home/user1//log/proglog2008.log &amp;amp;
sas /home/user1/SASprog/prog1.sas -sysparm '2009' -log /home/user1//log/proglog2009.log &amp;amp;&lt;BR /&gt;&lt;BR /&gt;wait&lt;/CODE&gt;&lt;/PRE&gt;&lt;P&gt;&lt;SPAN&gt;1) Does accessing the same data in parallel create issues?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;2) My understanding is each program will invoke a separate&amp;nbsp;sas session. This should not create conflict in terms of the Work library right?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;3) Is there a way to explicitly&amp;nbsp;purge the work libraries at the end of the program? Could there be dumps of earlier work library stashed somewhere I am not seeing, which in turn might be impacting subsequent memory use?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;After a few trials, I am unable to successfully run the program. Getting the following error:&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;PRE&gt;&lt;CODE class=" language-sas"&gt;ERROR: Insufficient space in file WORK.REG_DEC.DATA.
ERROR: File WORK.REG_DEC.DATA is damaged. I/O processing did not complete.
WARNING: The data set WORK.REG_DEC may be incomplete.  When this step was stopped there were 949663 observations and 178 variables.
WARNING: Data set WORK.REG_DEC was not replaced because this step was stopped.&lt;/CODE&gt;&lt;/PRE&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Getting similar error message in all except one log.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Should I just&amp;nbsp;reach out to the admin of this cluster regarding memory? Or am I doing something&amp;nbsp;fundamentally&amp;nbsp;wrong. Suggestions are appreciated.&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 19 Sep 2018 22:13:25 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Programming/running-multiple-SAS-programs-in-parallel-using-SLURM-Batch/m-p/497171#M131683</guid>
      <dc:creator>raulroy</dc:creator>
      <dc:date>2018-09-19T22:13:25Z</dc:date>
    </item>
    <item>
      <title>Re: running multiple SAS programs in parallel using SLURM Batch script</title>
      <link>https://communities.sas.com/t5/SAS-Programming/running-multiple-SAS-programs-in-parallel-using-SLURM-Batch/m-p/497180#M131689</link>
      <description>&lt;P&gt;&lt;EM&gt;1) Does accessing the same data in parallel create issues?&lt;/EM&gt;&lt;/P&gt;
&lt;P&gt;It makes the disk reads more random, therefore it slows down the program if you only have one SAS table in one location.&lt;/P&gt;
&lt;P&gt;It may be faster to run the programs sequentially.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;EM&gt;2) My understanding is each program will invoke a separate&amp;nbsp;sas session. This should not create conflict in terms of the Work library right?&lt;/EM&gt;&lt;/P&gt;
&lt;P&gt;Correct&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;EM&gt;3) Is there a way to explicitly&amp;nbsp;purge the work libraries at the end of the program? Could there be dumps of earlier work library stashed somewhere I am not seeing, which in turn might be impacting subsequent memory use?&lt;/EM&gt;&lt;/P&gt;
&lt;P&gt;SAS deletes its work libraries&amp;nbsp;when it ends.&lt;/P&gt;
&lt;P&gt;You can do intermediate purges run running&lt;FONT face="courier new,courier"&gt; proc datasets noprint kill;&amp;nbsp;&lt;/FONT&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;EM&gt;After a few trials, I am unable to successfully run the program. Getting the following error.&lt;/EM&gt;&lt;/P&gt;
&lt;P&gt;You need more disk space for the WORK library to avoid this error.&lt;/P&gt;
&lt;P&gt;The comment on 1) is true here too: the work library will be used with more random accesses when you have more processes using it.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;I would run one program, then run 2 concurrently, then 3, to see how run time is impacted, and what volume triggers a full disk space error.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 19 Sep 2018 22:38:40 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Programming/running-multiple-SAS-programs-in-parallel-using-SLURM-Batch/m-p/497180#M131689</guid>
      <dc:creator>ChrisNZ</dc:creator>
      <dc:date>2018-09-19T22:38:40Z</dc:date>
    </item>
    <item>
      <title>Re: running multiple SAS programs in parallel using SLURM Batch script</title>
      <link>https://communities.sas.com/t5/SAS-Programming/running-multiple-SAS-programs-in-parallel-using-SLURM-Batch/m-p/497198#M131702</link>
      <description>&lt;P&gt;Thanks a lot for your quick response. Ignoring issue 1 (which I wish to take care of by splitting the original data file into yearly files) is this a decent method of parallel processing sas programs?&lt;/P&gt;</description>
      <pubDate>Wed, 19 Sep 2018 23:14:52 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Programming/running-multiple-SAS-programs-in-parallel-using-SLURM-Batch/m-p/497198#M131702</guid>
      <dc:creator>raulroy</dc:creator>
      <dc:date>2018-09-19T23:14:52Z</dc:date>
    </item>
    <item>
      <title>Re: running multiple SAS programs in parallel using SLURM Batch script</title>
      <link>https://communities.sas.com/t5/SAS-Programming/running-multiple-SAS-programs-in-parallel-using-SLURM-Batch/m-p/497199#M131703</link>
      <description>&lt;P&gt;1- You haven't solved any disk access issue if all the "split" tables are in the same location.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;2- It's been years since I used Unix, but&amp;nbsp;&lt;a href="https://communities.sas.com/t5/user/viewprofilepage/user-id/11562"&gt;@Kurt_Bremser&lt;/a&gt;&amp;nbsp;would be able to&amp;nbsp; comment on the script.&lt;/P&gt;
&lt;P&gt;Another method is to manage everything from within a SAS session using &lt;A href="https://support.sas.com/rnd/scalability/connect/mp.html" target="_self"&gt;MP Connect&lt;/A&gt;.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 19 Sep 2018 23:23:44 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Programming/running-multiple-SAS-programs-in-parallel-using-SLURM-Batch/m-p/497199#M131703</guid>
      <dc:creator>ChrisNZ</dc:creator>
      <dc:date>2018-09-19T23:23:44Z</dc:date>
    </item>
    <item>
      <title>Re: running multiple SAS programs in parallel using SLURM Batch script</title>
      <link>https://communities.sas.com/t5/SAS-Programming/running-multiple-SAS-programs-in-parallel-using-SLURM-Batch/m-p/497205#M131705</link>
      <description>&lt;P&gt;Sometimes running out of WORK space in Unix is a side effect of multiple users running jobs at the same time and your SAS Admin may not be able to expand the WORK library for you.&amp;nbsp; If you have access to other disk space, you can have each program point to its own work library.&amp;nbsp; If you can put the different WORK libraries on different spindles, you will have better performance that way too.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;There are also some coding practices that lead to creating lots of WORK.data files that consume space.&amp;nbsp; This Google search will get you lots of help there&lt;/P&gt;
&lt;P&gt;efficient disk usage site:sas.com&lt;/P&gt;</description>
      <pubDate>Thu, 20 Sep 2018 01:01:26 GMT</pubDate>
      <guid>https://communities.sas.com/t5/SAS-Programming/running-multiple-SAS-programs-in-parallel-using-SLURM-Batch/m-p/497205#M131705</guid>
      <dc:creator>Doc_Duke</dc:creator>
      <dc:date>2018-09-20T01:01:26Z</dc:date>
    </item>
  </channel>
</rss>

