Re: How do I implement a Writable into another Writable?

2008-10-19 Thread Joman Chu
(out). Have a good one, Joman Chu http://www.notatypewriter.com/ AIM: ARcanUSNUMquam On Sun, Oct 19, 2008 at 3:30 AM, Yih Sun Khoo [EMAIL PROTECTED] wrote: I don't quite know how to write the read and write functions, but I want to write my own writable, which should have a DoubleWritable

Re: How do I implement a Writable into another Writable?

2008-10-19 Thread Joman Chu
Since the ArrayListWritable extends ArrayList, you have access to all the ArrayList methods as well. Once you read data in from the DataInput stream, you should be able to use ArrayListWritable just like a regular ArrayList. Joman Chu http://www.notatypewriter.com/ AIM: ARcanUSNUMquam On Sun

Re: How do I implement a Writable into another Writable?

2008-10-19 Thread Joman Chu
{ MyWritable w = new MyWritable(); w.readFields(in); return w; } EDIT: I was able to sort of replicate your error. In my constructor, i had my instance variables assigned to null. Make sure you assign them to new instances of whatever Writable you are using. Joman Chu http

Re: A question about Mapper

2008-10-04 Thread Joman Chu
before a mapper exits. Thanks Have a good one, -- Joman Chu Carnegie Mellon University School of Computer Science 2011 AIM: ARcanUSNUMquam -- Joman Chu Carnegie Mellon University School of Computer Science 2011 AIM: ARcanUSNUMquam

Re: A question about Mapper

2008-10-03 Thread Joman Chu
Have a good one, -- Joman Chu Carnegie Mellon University School of Computer Science 2011 AIM: ARcanUSNUMquam

Re: Text search on a PDF file using hadoop

2008-07-23 Thread Joman Chu
I've been investigating this recently, and I came across Apache PDFBox (http://incubator.apache.org/projects/pdfbox.html), which may accomplish this in native Java. Try it out and get back to us on how well it works, I'd be curious to know. Joman Chu AIM: ARcanUSNUMquam IRC: irc.liquid-silver.net

Re: How to chain multiple hadoop jobs?

2008-07-16 Thread Joman Chu
{ int res = ToolRunner.run(new Configuration(), new MyMapReduceTool(), new String[]{some, arguments}); System.exit(res); } } Joman Chu AIM: ARcanUSNUMquam IRC: irc.liquid-silver.net On Mon, Jul 14, 2008 at 5:46 PM, Joman Chu [EMAIL PROTECTED] wrote: Hi, I

Re: multiple Output Collectors ?

2008-07-14 Thread Joman Chu
One cheap hack that comes to mind is to extend the GenericWritable and ArrayWritable classes and write a second and third MapReduce job that will both parse over your first job's output, and each will select for the Key-Value pair it wants. Joman Chu AIM: ARcanUSNUMquam IRC: irc.liquid-silver.net

Re: How to chain multiple hadoop jobs?

2008-07-14 Thread Joman Chu
//do all sorts of configuration here //ie, set your Map, Combine, Reduce class //look at the Configuration class API } } The main think to know is that the ToolRunner.run() will call your class's run() method. Joman Chu AIM: ARcanUSNUMquam IRC

Re: File permissions issue

2008-07-09 Thread Joman Chu
, July 9, 2008 2:05 am, heyongqiang said: because in your permission set, the other role can not write the temp directory. and user3 is not in the same group with user2. heyongqiang 2008-07-09 ·¢¼þÈË£º Joman Chu ·¢ËÍʱ¼ä£º 2008-07-09 13:06:51 ÊÕ¼þÈË£º core-user@hadoop.apache.org

File permissions issue

2008-07-08 Thread Joman Chu
measure is to run a job as the user that started Hadoop. That is, in our example, after user1 starts Hadoop, user1 runs a job. Everything seems to work fine then. Thanks, Joman Chu