The Internet Archive has often been a valuable resource for journalists, from it’s finding records of deleted tweets or providing academic texts for background research. However, the advent of AI has created a new tension between the parties. A few major publications have begun blocking the nonprofit digital library’s access to their content based on concerns that AI companies’ bots are using the Internet Archive’s collections to indirectly scrape their articles.

“A lot of these AI businesses are looking for readily available, structured databases of content,” Robert Hahn, head of business affairs and licensing for The Guardian, told Nieman Lab. “The Internet Archive’s API would have been an obvious place to plug their own machines into and suck out the IP.”

The New York Times took a similar step. “We are blocking the Internet Archive’s bot from accessing the Times because the Wayback Machine provides unfettered access to Times content — including by AI companies — without authorization,” a representative from the newspaper confirmed to Nieman Lab. Subscription-focused publication the Financial Times and social forum Reddit have also made moves to selectively block how the Internet Archive catalogs their material.

Many publishers have attempted to sue AI businesses for how they access content used to train large language models. To name a few just from the realm of journalism:

Other media outlets have sought financial deals before offering up their libraries as training material, although those arrangements seem to provide compensation to the publishing companies rather than the writers. And that’s not even delving into the copyright and piracy issues also being fought against AI tools by other creative fields, from fiction writers to visual artists to musicians. The whole Nieman Lab story is well worth a read for anyone who has been following any of these creative industries’ responses to artificial intelligence.

Share.
Leave A Reply

Exit mobile version