Hi,
I'm downloading mainframe files using FTP in binary mode on to local file
system. These files are now seen as EBCDIC. The information about these
files are
(a) fixed in length ( each
field in record has fixed length).
(b)each record is of some KB
( This KB is fixed for each record).
Now here I'm able to convert this EBCDIC files to ASCII files with in unix
file system using the following command.
dd if <INPUTFILENAME> of <OUTPUTFILENAME> conv=ascii,unblock cbs=150
150 being the record size.
So, here I want this conversion to be done in Hadoop to leverage the use of
parallel processing. I was wondering is there any record reader available
for this kind of files and also about how to convert Packed Decimals COMP(3)
files to ASCII.
Any suggestions on how this can be done.
Srinivas