As we’ve discussed before, the rise of large artificial intelligence (AI) models has fundamentally disrupted the social contract governing machine use of web content. Today, machines don’t just access the web to make it more searchable or to help unlock new insights; they feed algorithms that fundamentally change (and threaten) the web we know. What once functioned as a mostly reciprocal ecosystem now risks becoming extractive by default.
In response, new approaches are emerging to support creators, publishers, and stewards of content to reclaim agency over how their works are used.
Pay-to-crawl is one approach beginning to come into focus. Pay-to-crawl refers to emerging technical systems used by websites to automate compensation for when their digital content—such as text, images, and structured data—is accessed by machines. We’ve recently published our interpretation and observations of pay-to-crawl systems in this dedicated issue brief.

Implemented responsibly, pay-to-crawl could represent a way for websites to sustain the creation and sharing of their content, and manage substitutive uses, keeping content publicly accessible where it might otherwise not be shared or would disappear behind even more restrictive paywalls.
However, we do have significant reservations.
Pay-to-crawl may represent an appropriate strategy for independent websites seeking to prevent AI crawlers from knocking them offline or to generate supplementary revenue. But elsewhere, pay-to-crawl systems could be cynically exploited by rightsholders to generate excessive profits, at the expense of human access and without necessarily benefiting the original creators.
Pay-to-crawl systems themselves could become new concentrations of power, with the ability to dictate how we experience the web. They could seek to watch and control how content is used in ways that resemble the worst of Digital Rights Management (DRM), turning the web from a medium of sharing and remixing into a tightly monitored content delivery channel.
We’re also concerned that indiscriminate use of pay-to-crawl systems could block off access to content for researchers, nonprofits, cultural heritage institutions, educators, and other actors working in the public interest. Legal rights to access content afforded by exceptions and limitations to copyright law, such as noncommercial research (in the EU) or fair use exemptions (in the US), as well as provisions for translation and accessibility tools, have been carefully negotiated and adjusted over time. These rights could be impeded by the introduction of blunt, poorly designed pay-to-crawl systems.
Pay-to-crawl systems are not neutral infrastructure. It’s vital that these systems are built and used in ways that serve the interests of creators and the commons, rather than simply create barriers to the sharing of knowledge and creativity, and benefit the few.
We’re proposing the following set of principles as a way to guide the development of pay-to-crawl systems in alignment with this vision:
We believe now is the moment to engage, to influence, and to infuse pay-to-crawl systems with values that prioritize reciprocity, openness, and the commons.
We welcome feedback and dialogue on the principles outlined here. Your input will help guide our engagement with pay-to-crawl systems and related initiatives moving forward, as well as inform the wider CC community’s understanding of them.
Thank you to Jack Hardinges for his contributions to this post.
Back to Posts