To get such huge files just google on "how to create large files." Using cat command is one such option.
On Tue, Apr 19, 2011 at 11:03 AM, Mehmet Tepedelenlioglu < mehmets...@gmail.com> wrote: > For such small input, the only way you would see speed gains would be if > your job was dominated > by cpu time, and not i/o. Since word-count is mostly an i/o problem and > your > input size is quite small, you are seeing similar run times. 3 computers is > better than 1 > only if you need them. > > On Apr 18, 2011, at 10:06 PM, praveenesh kumar wrote: > > > The input were 3 plain text files.. > > > > 1 file was around 665 KB and other 2 files were around 1.5 MB each.. > > > > Thanks, > > Praveeenesh > > > > > > > > On Tue, Apr 19, 2011 at 10:27 AM, real great.. < > greatness.hardn...@gmail.com > >> wrote: > > > >> Whats your input size? > >> > >> On Tue, Apr 19, 2011 at 10:21 AM, praveenesh kumar < > praveen...@gmail.com > >>> wrote: > >> > >>> Hello everyone, > >>> > >>> I am new to hadoop... > >>> I set up a hadoop cluster of 4 ubuntu systems. ( Hadoop 0.20.2) > >>> and I am running the well known word count (gutenberg) example to test > >> how > >>> fast my hadoop is working.. > >>> > >>> But whenever I am running wordcount example..I am not able to see any > >> much > >>> processing time difference.. > >>> On single node the wordcount is taking the same time.. and on cluster > of > >> 4 > >>> systems also it is taking almost the same time.. > >>> > >>> Am I doing anything wrong here ?? > >>> Can anyone explain me why its happening.. and how can I make maximum > use > >> of > >>> my cluster ?? > >>> > >>> Thanks. > >>> Praveenesh > >>> > >> > >> > >> > >> -- > >> Regards, > >> R.V. > >> > > -- Regards, R.V.