To get such huge files just google on "how to create large files."
Using cat command is one such option.

On Tue, Apr 19, 2011 at 11:03 AM, Mehmet Tepedelenlioglu <
mehmets...@gmail.com> wrote:

> For such small input, the only way you would see speed gains would be if
> your job was dominated
> by cpu time, and not i/o. Since word-count is mostly an i/o problem and
> your
> input size is quite small, you are seeing similar run times. 3 computers is
> better than 1
> only if you need them.
>
> On Apr 18, 2011, at 10:06 PM, praveenesh kumar wrote:
>
> > The input were  3  plain text files..
> >
> > 1 file was around 665 KB and other 2 files were around 1.5 MB each..
> >
> > Thanks,
> > Praveeenesh
> >
> >
> >
> > On Tue, Apr 19, 2011 at 10:27 AM, real great.. <
> greatness.hardn...@gmail.com
> >> wrote:
> >
> >> Whats your input size?
> >>
> >> On Tue, Apr 19, 2011 at 10:21 AM, praveenesh kumar <
> praveen...@gmail.com
> >>> wrote:
> >>
> >>> Hello everyone,
> >>>
> >>> I am new to hadoop...
> >>> I set up a  hadoop cluster of 4 ubuntu systems. ( Hadoop 0.20.2)
> >>> and I am running the well known word count (gutenberg) example to test
> >> how
> >>> fast my hadoop is working..
> >>>
> >>> But whenever I am running wordcount example..I am not able to see any
> >> much
> >>> processing time difference..
> >>> On single node the wordcount is taking the same time.. and on cluster
> of
> >> 4
> >>> systems also it is taking almost the same time..
> >>>
> >>> Am I  doing anything wrong here ??
> >>> Can anyone explain me why its happening.. and how can I make maximum
> use
> >> of
> >>> my cluster ??
> >>>
> >>> Thanks.
> >>> Praveenesh
> >>>
> >>
> >>
> >>
> >> --
> >> Regards,
> >> R.V.
> >>
>
>


-- 
Regards,
R.V.

Reply via email to