Hi @evgenys
First things, first... This is frustrating situation to find yourself in. Everyone who has participated in this thread has experienced it. We feel your pain.
TL;DR
Best advice: Get the official file format and data details from the producer/creator of the file.
Now, for a nonsensical philosophical discussion.
When I have to do things like this I ask myself a series of questions. The first two questions are the most important. The answer to them will dictate whether I ask the other questions.
What's the cost of being wrong?
If this data is really important then figuring out the format on your own is a bad idea because being wrong is expensive. In other words: the creator/producer of the data has responsibility for ACCURATELY describing it.
If the data isn't critical you can use this as a learning exercise but it isn't a really great learning exericise.
Is this a one time exercise or is it going to be repeated? If it needs to be a repeatable process how often will it run?
If this is a one-time exercise then running guessrows on the entire file may not be a huge deal. Start the job before you go to lunch or leave for the day.
If this is for a production job, you want a real, stable, description of the file layout.
Pro Tip: if it takes your machine a long time to run this job, use this as justification for a new, awesome, computer.
Where does the data originate?
If the data comes from a database a DBA can point you to the SQL (Data Definition Language (DDL) that was used to create the table). Using this, I know the longest possible length of the character strings.
If the data is something someone threw together it may be a good idea to ask them how they are doing it. Perhaps they can provide the length of the strings.
Others have pointed-out that we should always ask the source of the data (person or machine) for the file layout and column length information. It is solid advice. In fact, I believe this is the best advice you could be given. As you can tell from the comments, it is very frustrating to be asked to read in a file that you know nothing about.
How big is this CSV file (KB, MB, GB)?
4 million rows is not a lot these days. For example, when I am experimenting with database loads I will typically use 10 to 20 million rows. My desktop PC handles this with no problem even when I am loading data into cloud databases. That being said, if the file is huge (many GB) it could take some time to process which leads to the next question.
Can I subset the file in so that my chances of finding the longest text fields increases?
If may be possible to sort the file so the longest records appear at the top. This could help you figure-out the max length of the character fields.
Does anyone know something that can help me?
This is a variant of the first question. You may have a co-worker who has read the file before and can help you.
I know this isn't what you are after but I think it is the only real answer. Anyone who has been through this can identify with it. It is frustrating.
Best wishes,
Jeff
Very simple. Write a data step according to the file description.
To get help in this, you best post that description here.
Then DEMAND the description from whoever gave you that file.
Your only other option is to run proc import with a large enough guessingrows value, and hope for the best.
Unless you want to inspect all rows with the good old eyeballs Mk 1.
There isn't much of a choice then. I suggest the standard process of using proc import to generate code. Copy the code from the log and customize according to your errors until they're gone.
I would recommend setting GUESSINGROWS to 1 million for initial proc import.
You can limit total amount read using OBS option.
Option obs=1500000;
Theb reset after:
option obs=Max;
If 1 million doesn't get it that would be very surprising.
I suggested 1.5,million.
@evgenys wrote:
I tried GUESSINGROWS 4 million ...it doesn't end.😃
Oh yeah, 4 million rows * 150 columns takes a lot of time to inspect. It is what happens when you try something stupid.
Get back to the data source and demand the description, and tell them that the data is useless without it.
This isn't an easy problem!
Do any of your fields have internal commas? If not, then you could use (untested):
data _null_;
retain MaxCols 0;
infile x end=LastRec;
input;
ColCount = count(_infile_, ",") + 1;
if ColCount > MaxCols then MaxCols = ColCount;
if LastRec then call symput("ColCount", put(MaxCols, best8.));
run;
This should give you the highest number of columns in your file.
Then maybe something along the lines of:
%macro GetLen;
data _null_;
retain MaxLen 0;
infile x end=LastRec;
input;
%do &i = 1 to &ColCount;
ColLen = length(scan(_infile_, &i))
if ColLen > MaxLen then MaxLen = ColLen;
%end;
if LastRec then call symput("ColLen", put(MaxLen, best8.));
run;
%mend;
%GetLen;
And finally:
%macro GetData;
data Want;
length Col1-Col&ColCount. $&ColLen.;
infile x;
input;
%do &i = 1 to &ColCount;
Col&i. scan(_infile_, &i);
%end;
run;
%mend;
%GetData;
Good luck!
Tom
You don't need to know what is in a CSV file to read it in as character strings. If you don't even know how many columns there are just use a larger number (150 in the example below) than you expect. If the last column is not empty then increase the number and read the file again.
data temp (compress=yes);
infile 'myfile.csv' dsd truncover firstobs=2 ;
length x1-x150 $200 ;
input x1-x150;
run;
You can then analyze the character strings yourself and make your own decision on what is in it.
If the first line has variable names then you could read that line in separately and use it to rename the variables.
Here is a quick way to take a 1% random sample of the records using a line pointer control in the input statement
data sample;
length str $200;
infile "&sasforum\datasets\frame.csv" truncover line=lineNo;
linePt = ceil(2 * 100 * rand("uniform"));
input #linePt str&;
line + lineNo;
keep line str;
run;
My rough stab would be to run proc import with guessingows in the 32000 range. The save and inspect the generated datastep code.
I would likely increase the lengths of character variables in case the generated 73 or such doesn't quite work for all of them, likely around 10 percent added.
But if you don't have a document that says what any of the columns are then what are you going to do with this data file? Unless you have column headers long enough to be explanatory it may be hard to tell what anything else is.
Join us for SAS Innovate 2025, our biggest and most exciting global event of the year, in Orlando, FL, from May 6-9. Sign up by March 14 for just $795.
Need to connect to databases in SAS Viya? SAS’ David Ghan shows you two methods – via SAS/ACCESS LIBNAME and SAS Data Connector SASLIBS – in this video.
Find more tutorials on the SAS Users YouTube channel.