Is it possible to split compressed input from a single job to multiple map tasks? My current configuration has several task trackers but the job I kick off results in a single map task. I'm launching these jobs in sequence via a shell script, so they end up going through a pipeline of 1 concurrent map which is kinda suboptimal.

When I run this task via a full local hadoop stack it does seem to split the file into multiple small task chunks.

--
Jimmy

Reply via email to