Yes, we have files which are only EBCDIC and files containing  EBCDIC +
PACKED DECIMALS. As a first step started working on only EBCDIC files. dd
command working fine but the intention is to do this conversion with in HDFS
to leverage parallel processing.

On Sat, Oct 15, 2011 at 10:58 AM, Michel Segel <[email protected]>wrote:

> You may not want to do this...
> Does the data contain and packed or zoned decimals?
> If so, the dd conversion will corrupt your data.
>
>
> Sent from a remote device. Please excuse any typos...
>
> Mike Segel
>
> On Oct 15, 2011, at 3:51 AM, SRINIVAS SURASANI <[email protected]> wrote:
>
> > Hi,
> >
> > I'm downloading mainframe files using FTP in binary mode on to local file
> > system. These files are now seen as EBCDIC. The information about these
> > files are
> >                                                (a) fixed in length ( each
> > field in record has fixed length).
> >                                                (b)each record is of some
> KB
> > ( This KB is fixed for each record).
> >
> > Now here I'm able to convert this EBCDIC files to ASCII files with in
> unix
> > file system using the following command.
> > dd if <INPUTFILENAME>  of <OUTPUTFILENAME> conv=ascii,unblock cbs=150
> > 150 being  the record size.
> >
> > So, here I want this conversion to be done in Hadoop to leverage the use
> of
> > parallel processing. I was wondering is there any record reader available
> > for this kind of files and also about how to convert Packed Decimals
> COMP(3)
> > files to ASCII.
> >
> > Any suggestions on how this can be done.
> >
> >
> > Srinivas
>

Reply via email to