Hi all,
I've contributed a couple of classes to support fixed length/width
records in input files. The JIRA issue and attachments are located here:
https://issues.apache.org/jira/browse/MAPREDUCE-1176
thanks, and I hope this helps others out.
On Oct 28, 2009, at 1:58 PM, Aaron Kimball wrote:
I think these would be good to add to mapreduce in the
{{org.apache.hadoop.mapreduce.lib.input}} package. Please file a
JIRA and
apply a patch!
- Aaron
On Wed, Oct 28, 2009 at 11:15 AM, yz5od2 <woods5242-
[email protected]>wrote:
Hi all,
I am working on writing a FixedLengthInputFormat class and a
corresponding
FixedLengthRecordReader.
Would the Hadoop commons project have interest in these? Basically
these
are for reading inputs of textual record data, where each record is
a fixed
length, (no carriage returns or separators etc)
thanks
On Oct 20, 2009, at 11:00 PM, Aaron Kimball wrote:
You'll need to write your own, I'm afraid. You should subclass
FileInputFormat and go from there. You may want to look at
TextInputFormat
/
LineRecordReader for an example of how an IF/RR gets put together,
but
there
isn't an existing fixed-len record reader.
- Aaron
On Tue, Oct 20, 2009 at 12:59 PM, yz5od2 <[email protected]
wrote:
Hi,
I have input files, that contain NO carriage returns/line feeds.
Each
record is a fixed length (i.e. 202 bytes).
Which FileInputFormat should I be using? so that each call to my
Mapper
receives one K,V pair, where the KEY is null or something (I
don't care)
and
the VALUE is the 202 byte record?
thanks!