If your count keeps increasing you will always run into some sort of time limit. Why not consider doing this processing in a task (they can run for 10mins, or multiple tasks. ) I am assuming your trying to summarise etc....
T On Friday, August 24, 2012 3:26:00 AM UTC+8, Phil wrote: > > In some initialization work my app needs to run through the all of the > datastore entities of a given kind. I have a lot of these entities (80k > currently) and it's increasing rapidly. I'm currently trying to read these > in using a single datastore query, but running up against the default > datastore timeout of 30 seconds. > > Is there a good practice for sharding this or otherwise breaking this up > so that I won't hit these deadlines? I was thinking I would do a keyOnly > query and then break up the keys into a number of reasonably sized > sub-queries, but perhaps there is a better approach out there? > > Thanks, > Phil > -- You received this message because you are subscribed to the Google Groups "Google App Engine" group. To view this discussion on the web visit https://groups.google.com/d/msg/google-appengine/-/5x0WQc9ODVwJ. To post to this group, send email to [email protected]. To unsubscribe from this group, send email to [email protected]. For more options, visit this group at http://groups.google.com/group/google-appengine?hl=en.
