Subir, No, not unless you use a specialized streaming library (pydoop, dumbo, etc. for python, for example).
On Tue, Feb 28, 2012 at 2:19 PM, Subir S <subir.sasiku...@gmail.com> wrote: > Can multiple output be used with Hadoop Streaming? > > On Tue, Feb 28, 2012 at 2:07 PM, madhu phatak <phatak....@gmail.com> wrote: > >> Hi Mohit , >> A and B refers to two different output files (multipart name). The file >> names will be seq-A* and seq-B*. Its similar to "r" in part-r-00000 >> >> On Tue, Feb 28, 2012 at 11:37 AM, Mohit Anchlia <mohitanch...@gmail.com >> >wrote: >> >> > Thanks that's helpful. In that example what is "A" and "B" referring to? >> Is >> > that the output file name? >> > >> > mos.getCollector("seq", "A", reporter).collect(key, new Text("Bye")); >> > mos.getCollector("seq", "B", reporter).collect(key, new Text("Chau")); >> > >> > >> > On Mon, Feb 27, 2012 at 9:53 PM, Harsh J <ha...@cloudera.com> wrote: >> > >> > > Mohit, >> > > >> > > Use the MultipleOutputs API: >> > > >> > > >> > >> http://hadoop.apache.org/common/docs/current/api/org/apache/hadoop/mapred/lib/MultipleOutputs.html >> > > to have a named output of bad records. There is an example of use >> > > detailed on the link. >> > > >> > > On Tue, Feb 28, 2012 at 3:48 AM, Mohit Anchlia <mohitanch...@gmail.com >> > >> > > wrote: >> > > > What's the best way to write records to a different file? I am doing >> > xml >> > > > processing and during processing I might come accross invalid xml >> > format. >> > > > Current I have it under try catch block and writing to log4j. But I >> > think >> > > > it would be better to just write it to an output file that just >> > contains >> > > > errors. >> > > >> > > >> > > >> > > -- >> > > Harsh J >> > > >> > >> >> >> >> -- >> Join me at http://hadoopworkshop.eventbrite.com/ >> -- Harsh J