Is it possible to split compressed input from a single job to multiple map
tasks? My current configuration has several task trackers but the job I
kick off results in a single map task. I'm launching these jobs in
sequence via a shell script, so they end up going through a pipeline of 1
concurrent map which is kinda suboptimal.
When I run this task via a full local hadoop stack it does seem to split
the file into multiple small task chunks.
--
Jimmy
- Splitting compressed input from a single job to multiple map... Jimmy Wan
-