Regarding the performance problem:

The schema of my application is:

1. I execute perl script which performs a search in a public database. It
gets total results in *several pages*. Pressing "Next Page" button (with
perl script) i get a list of all the links related to my query (70.000 more
or less) I write down all these links in a unique text file.

2. From the Java i read each of the 70.000 links and i create a new file
containing the current i'm reading. Then i call a perl script which uses
this link as input parameter. It browses it and get website content saving
it in a local html file.

I'm having performance problems ,,,,  i've tried to don't create a single
file containing url for each of the 70.000 links and pass it automatically
to perl script as input parameter but it fails...

I've heard about LWP module? do you recomend me to use it??
Have you ever done something similar to this? can you give me some advice?
Thanks

T.





On 1/15/07, Rob Dixon <[EMAIL PROTECTED]> wrote:

Tatiana Lloret Iglesias wrote:
> Hi all,
> from my java applicaation i invoke a perl script which downloads a huge
> quantity of files from an external database using WWW-mechanize library
and
> my problem is that I have big CPU performance problems ... can you give
me
> any advice to avoid this?

Hi Tatiana

Do you really mean "CPU performance problems"? If you're downloading a lot
from
the Internet then your problem is more likely to be limited by the speed
of the
network, in which case you need to look at what you're downloading and see
if
you can get the information you require without moving as much extraneous
data.
Can you anticipate the URLs you need to access instead of following links
from
other pages, for instance?

If your process is indeed CPU-bound then it is most liley to be an error
in your
coding, which we would need to see before we could help.

HTH,

Rob

Reply via email to