Hi All I am reading data from a HADOOP table on a simple DI job. One of the columns in the source object has some values that are non English character set, e.g BANČNA PODRUŽNICA. The job fails with the following message: ERROR: Data from column 'fincl_acnt_num' in row 273210 of the result set was not presented in Hadoop UTF-8 format. The length of this data is 22 bytes, and the first 7 characters are '2025088'. Adding -JREOPTIONS (-Dfile.encoding=UTF-8) to the SAS invocation may circumvent the issue. Otherwise the data should be corrected to UTF-8 format. I have gone through NLS forums but none of the solutions apply or I wasn't able to apply for my case on DI Studio. Thanks! Samir
... View more