I see.

Online you can find a list of ip addresses used by those bots.
Assuming you have the list of IPS of allowed bots in BOTS you can
replace

@auth.requires_login()

with

@auth.requires(auth.user or request.client in BOTS)

On Sep 2, 9:57 pm, Narendran <[email protected]> wrote:
> Thanks for the reply Massimo.
>
> But my requirement is that the bots and Facebook scrapers should be
> able to access the page, without authentication. I've seen a few
> newspaper sites, and online tech journal sites have this
> functionality, where in Google will have the content indexed, but as a
> user, I need to be a registered user to access the content. I'm
> looking for such a functionality on my site.
>
> On Sep 2, 6:32 pm, mdipierro <[email protected]> wrote:
>
> > In your model you can do something like this:
>
> > if not "Mozilla" in request.env.http_user_agent: raise HTTP(500)
>
> > On Sep 2, 4:19 am, Narendran <[email protected]> wrote:
>
> > > Hello all,
> > > I have a few controller methods (and corres. web pages) that are
> > > authorized to be viewed only by registered users. But I don't want to
> > > authenticate search engine bots, Facebook scrapers, and other such
> > > services. Is there a recommended way of doing this? (like setting auth
> > > requirements based on user agents, etc.)

Reply via email to