For three decades, a tiny text file has kept the Internet from chaos. This text file has no particular legal or technical authority, and it's not even particularly complicated. It represents a handshake deal between some of the earliest pioneers of the Internet to respect each other's wishes and build the internet in a way that benefitted everybody. It's a mini constitution for the Internet, written in code.
It's called robots.txt and is usually located at yourwebsite.com/robots.txt. That file allows anyone who runs a website — big or small, cooking blog or multinational corporation — to tell the Web who's allowed in and who isn't. Which search engines can index your site? What archival projects can grab a version of your page and save it? Can competitors keep tabs on your pages for their own files? You get to decide and declare that to the Web.
It's not a perfect system, but it works. Used to, anyway. For decades, the main focus of robots.txt was on search engines; you'd let them scrape your site and in exchange they'd promise to send people back to you. Now AI has changed the equation: companies around the Web are using your site and its data to build massive sets of training data, in order to build models and products that may not acknowledge your existence at all.
From The Verge
No entries found