Question of speed of delivery.
Scenario:
I have a remote server A that calls remote server B through the firewall to retrieve a
pipe-delimited string of real-time market quotes. At present the webserver makes a
Perl CGI call to server A to get the data, formats it into an HTML table, and serves
it up on the homepage (www.troweprice.com). So every HTTP GET request to the
webserver spawns a separate process to fetch and process the quote data.
New Design Options (forget CORBA for the moment):
1. Have a cron run the Perl script to write the quote data to a flatfile every 10
seconds; have a Java servlet read that file every five seconds, holding the data in
memory, and delivering the formatted HTML to the clients per request by spawning
multiple threads.
2. Have a cron run a C version of the script to get the data every 10 seconds and
renew an otherwise static HTML page that will be served by the webserver per every
HTTP GET request.
Which solution do you think would be the faster? Are there others I am neglecting?
Thanks for the input (pun intended)!
-mark
___________________________________________________________________________
To unsubscribe, send email to [EMAIL PROTECTED] and include in the body
of the message "signoff SERVLET-INTEREST".
Archives: http://archives.java.sun.com/archives/servlet-interest.html
Resources: http://java.sun.com/products/servlet/external-resources.html
LISTSERV Help: http://www.lsoft.com/manuals/user/user.html