I had similar issues, I achieved it by setting up parameters on the URL and batching the fetchs. Something like this myapp.appspot.com/ urltofetch?index=20. Setup multiple Cron jobs to fetch them as needed, play around with your URLs to workout how quickly they can return and therefore how many you can batch at one time.
Another way of doing it (and I don't know the code but it's in here) is that there is an error triggered when your deadline is exceeded. You can catch this and then execute the next batch of URLs, write an index of where you are each time you get a successful fetch so your new process knows where to start from. Cheers, Tim On Apr 28, 12:44 am, kukaracha <[email protected]> wrote: > Hi all, > > I have a phyton script (no multithread) wich request 200 urls to read > data . > > I want to include in cron but i have a doubt. I will have Timeout > problems ? > > What is the best way to do this ? > > Regards. --~--~---------~--~----~------------~-------~--~----~ You received this message because you are subscribed to the Google Groups "Google App Engine" group. To post to this group, send email to [email protected] To unsubscribe from this group, send email to [email protected] For more options, visit this group at http://groups.google.com/group/google-appengine?hl=en -~----------~----~----~----~------~----~------~--~---
