BookmarkSubscribeRSS Feed
racrown
Calcite | Level 5

We have Visual Analytics 7.3 ( installed in a distributed environment -4 worker nodes with 128GB RAM each.(no co-located data store)  I am loading data from Hive(separate Hadoop cluster w/o SAS compoents) into LASR.  My users want to search raw data sets which are aproximately 200GB+.  I have no problem loading smaller sets which are as large as 80GB into LASR, but the 200GB+ fail to load.  When they apply the filters to get to the data they need, it should be less than 1% of the 200GB.  Pulling it all into the LASR memory structures seem to be very cumbersome for this use case.   I am thinking a different SAS tool would have better served their needs.  I am thinking Data Miner, Web Reports or SAS/ACCESS Interface to Hadoop might map well.  Advice is appreciated.

1 REPLY 1
SASKiwi
PROC Star

I would suggest that running SQL queries against the data in Hadoop would be a far better option. LASR servers should be for  reporting and exploring using the VA front-end only.

SAS Innovate 2025: Call for Content

Are you ready for the spotlight? We're accepting content ideas for SAS Innovate 2025 to be held May 6-9 in Orlando, FL. The call is open until September 25. Read more here about why you should contribute and what is in it for you!

Submit your idea!

Tips for filtering data sources in SAS Visual Analytics

See how to use one filter for multiple data sources by mapping your data from SAS’ Alexandria McCall.

Find more tutorials on the SAS Users YouTube channel.

Discussion stats
  • 1 reply
  • 700 views
  • 1 like
  • 2 in conversation