Help using Base SAS procedures

sashadoop connectivity

Accepted Solution Solved
Reply
New Contributor
Posts: 3
Accepted Solution

sashadoop connectivity

Dear All,

 

In need of your advise,

 

we are working for a customer where they have sas system and hadoop system,now the challenge is we need to establish connectivity between sas and hadoop,more over customer wants us to develop a connecctor instead of buying licence for hadoop connecctivity,

is there any way where I can bridge in sas and hadoop,and I have seen some options as below

 

So far we have obsurved 3 approaches to connect 

 

Option 1 : LIBNAME statements can be used to make Hive tables look like SAS data sets on top of which SAS Procedures and SAS DATA steps can interact.

                 

 Option 2 : PROC SQL commands provide the ability to execute direct Hive SQL commands on Hadoop.

Option 3 : PROC HADOOP provides the ability to directly submit MapReduce, Apache Pig, and HDFS commands from the SAS execution environment to your CDH cluster.

if not these above options do we have any way to connect

Thanks,

Suneel


Accepted Solutions
Solution
‎09-21-2016 03:40 AM
New Contributor
Posts: 3

Re: sashadoop connectivity

Dear Gergely,

 

Thanks a Ton for your prompt reply,we are following instructions given by you ,here is the link I found to configure hadoop with sas,

https://support.sas.com/resources/thirdpartysupport/v94/hadoop/hadoopbacg.pdf

 

Thanks,

Suneel.

 

View solution in original post


All Replies
SAS Employee
Posts: 340

Re: sashadoop connectivity

Option1 and Option2 are basically the same: if you established a connection to Hive or Impala then you can use proc sql, data step or any other proc to access data in hadoop.
Yes: proc sql with hive or impala libname engine is able to pass through (aka. push down) the query to hadoop (as with almost all database engines). Access to Hadoop or Access to Impala license is needed for Option1/Option2.

Option3: With only Base SAS license you can use proc hadoop to run MR,Pig and HDFS commands. Also you can read and write HDFS files by creating a fireref with the HADOOP Access Method.

Option4: You could start external programs from SAS with the x command. Those external programs could upload/download files, datasets to/from Hadoop.
Similar to this: you can run java and groovy programs from SAS.
Depending on the applications on Hadoop, if they support some kind of API (REST for example), you can connect to it with custom developed SAS programs.

Downside of the later solutions: performance. You usually use Hadoop, because "data is big", and you don't want to wait. Making intermediate copies on the client or on server side slows down execution.
Solution
‎09-21-2016 03:40 AM
New Contributor
Posts: 3

Re: sashadoop connectivity

Dear Gergely,

 

Thanks a Ton for your prompt reply,we are following instructions given by you ,here is the link I found to configure hadoop with sas,

https://support.sas.com/resources/thirdpartysupport/v94/hadoop/hadoopbacg.pdf

 

Thanks,

Suneel.

 

☑ This topic is solved.

Need further help from the community? Please ask a new question.

Discussion stats
  • 2 replies
  • 250 views
  • 0 likes
  • 2 in conversation