I see that for the web2py Wiki, Content Search is a high priority
Action Item, and I imagine for T3 as well.
But since most of the content is in the database, not in easy to crawl
static pages,
I wonder what would be a good technique for exposing the content from
the db to the crawler?

Myself I would suppose one could create a private app-specific
extracter process over the content db in a Cron task which would
create some artifcial 'crawler-friendly' static pages into a special
directory, and then it would generate a 'Google XML Sitemap' (see
http://www.gcwweb.com/search-engine-optimisation/google-sitemaps-search-engine-website-submission.htm)
to point to the special crawlable pages.  These pages would have to
contain some type of index of links to the actual content, perhaps
driven by tags which hopefully the content creators have been creating
for each page.

Does anyone out there have more ideas?


--~--~---------~--~----~------------~-------~--~----~
You received this message because you are subscribed to the Google Groups 
"web2py Web Framework" group.
To post to this group, send email to [email protected]
To unsubscribe from this group, send email to 
[email protected]
For more options, visit this group at 
http://groups.google.com/group/web2py?hl=en
-~----------~----~----~----~------~----~------~--~---

Reply via email to