There is not any error in the logs. I think it is because of protection. The website use https connection. Is there anyway to pass over it?
On 18 Şubat, 19:45, Ernesto Karim Oltra <[email protected]> wrote: > Perhaps the page has been protected so you need UserAgent headers, > etc. Anyway, check the Logs section in Dashboard to check for errors. > > On 17 feb, 23:18, theone <[email protected]> wrote: > > > > > > > > > Actually for my case using google docs api very complicated and tiring > > because I just want to get the content to make search in files. I > > won't make any modification on the document. I think that using an > > external service > > likehttps://secure.convert-doc.com/converters/doc-to-txt.html > > might be appropriate solution. > > > I tried using urlfetch > > > file = self.request.get("file") > > form_fields = { > > "input_file": file, > > "output_type": "txt", > > "output_method": "1", > > ".cgifields": > > "output_method" > > } > > form_data = urllib.urlencode(form_fields) > > result = > > urlfetch.fetch(url="https://secure.convert-doc.com/convert- > > file", > > payload=form_fields, > > method=urlfetch.POST, > > headers={'Content-Type': 'multipart/form- > > data'}) > > print result.content > > > but it gave Internal Server Error. I don't know what is wrong with it. > > The system seems working normally but I could not run it with > > urlfetch. > > > On 17 Şubat, 17:12, Ernesto Karim Oltra <[email protected]> > > wrote: > > > > Taskqueues, that's what I'm using now. To update a word document of > > > about ten pages, it takes more or less 1500ms average (at least for me > > > =) ). And this have a big advantage, you can take control of retries > > > if docs is temporaly unavailable. > > > > Another useful option would be to store a cache in datastore with some > > > common used data. For example, i have a downloads section, containing > > > some documents stored in google docs. Hourly (it may be five minutes > > > or 3 days, whatever you want) cron gets the titles, the URL of the > > > document, etc. and save in datastore. Then I haven't got to deal with > > > docs when building responses to a user, only use the data cached > > > locally. > > > > On 17 feb, 02:41, Calvin <[email protected]> wrote: > > > > > I think he means that importing and retrieving the converted document > > > > using > > > > the gdata api may not always be possible within the 30 second limit of a > > > > user-facing app engine request. > > > > > If that's the case it would be a good idea to do the conversion using a > > > > task > > > > queue, which has a much higher limit. > > > > > Another thing to keep in mind is that Google Docs has a size limit on > > > > word > > > > documents that it will import. -- You received this message because you are subscribed to the Google Groups "Google App Engine" group. To post to this group, send email to [email protected]. To unsubscribe from this group, send email to [email protected]. For more options, visit this group at http://groups.google.com/group/google-appengine?hl=en.
