Google announces support for robots.txt to control (some) aspects of their AI

Back in February, in:

Giving Creators and Websites Control Over Generative AI
https://lauren.vortex.com/2023/02/14/giving-creators-and-websites-control-over-generative-ai

I suggested use of the Robots Exclusion Protocol (robots.txt) to
control how site contents are accessed and used by AI systems. OpenAI
announced this support for future versions of ChatGPT back in August,
and now Google has done the same for future versions of some of their
AI products:

https://www.engadget.com/google-will-let-publishers-hide-their-content-from-its-insatiable-ai-202015557.html

There are some details about this that I do not fully understand at
this time. In particular, the note that robots.txt cannot be used to
control how Google Search accesses information from sites for its
"Search Generative Experience" -- SGE. The implication is that the
only way to block use of data by SGE is by telling Google not to index
the data for Search at all -- even in its traditional "non-Generative
AI" form. If this is actually the case it is highly problematic for
obvious reasons -- but again the details on this are not entirely
clear to me at the moment.  -L

- - -
--Lauren--
Lauren Weinstein lau...@vortex.com (https://www.vortex.com/lauren)
Lauren's Blog: https://lauren.vortex.com
Mastodon: https://mastodon.laurenweinstein.org/@lauren
Pebble (formerly T2): https://pebble.is/laurenweinstein
Twitter: https://twitter.com/laurenweinstein
Founder: Network Neutrality Squad: https://www.nnsquad.org
        PRIVACY Forum: https://www.vortex.com/privacy-info
Co-Founder: People For Internet Responsibility
Tel: +1 (818) 225-2800
_______________________________________________
nnsquad mailing list
https://lists.nnsquad.org/mailman/listinfo/nnsquad

Reply via email to