Your requirement isn't very specific. MapR is an Hadoop distribution coming with many storage tools and file or data electronic formats. On the other hand, an Oracle table is a proprietary data format - as far as I now - not compatible with any Hadoop native storage format. You've been asked to migrate data from an Oracle database to Hadoop MapR using SAS as a kind of handy bridge/migration tool at your disposal. Dumping data from Oracle with any SQL client that generates flat files (e g CSV) would work all the same if I understand correctly your question.
There are even tools in Mapr (Hadoop) like Sqoop + Oozie or Sqoop2 which can transfer directly - if your network rules allows this - betwen Oracle and Hadoop/HDFS (flat-files stored in distributed chunks inside Hadoop) using JDBC connection. An Oracle JDBC driver installed on the Hadoop corresponding node (executing Sqoop or Sqoop2) together with a network flow opening (if feasible) can launch SQL requests on the Oracle DBMS and feed the Hadoop File System, generating CSV files for instance or Hive Tables sometimes.
In which format are you expected to deliver your data on the target system - namely the Hadoop MapR cluster ? flat-files (CSV), binary files (PARQUET, AVRO etc.). Hive tables etc. ?
Unfortunately, your 9.3 SAS system won't be able to communicate directly with MapR since the connector SAS/ACCESS to Hadoop is supported with MapR distribution only with SAS 9.4 latest release M3.
HTH ...