Can Licensing Save the Content Economy?

Two views on licensing and the creator economy.

Can Licensing Save the Content Economy?
Photo by TheRegisti / Unsplash

This week, Sonia Hendy and Nick Sullivan offer different perspectives on the role licensing should play in restoring the financial incentives for web content creators.

Hendy, a creator herself and founder of an AI licensing infrastructure platform, is the more bullish of the two, arguing that workable infrastructure already exists and that creators cannot afford to wait for perfect policy.

Sullivan, an applied cryptographer and internet standards veteran who sits on the Internet Architecture Board, is more bearish, warning that licensing alone cannot fix what is fundamentally a governance problem.

Both agree it is not enough.

License-first? Mind the Gap!

By Sonia Hendy

After watching independent creators lose control of their work to extractive AI scraping, I realized that licensing infrastructure, not another government consultation, could change that. Even without perfect policy.

Like many people in the creative industries in the UK, I waited for last Wednesday's government report on copyright and AI with both excitement and anxiety. I’d tried to ignore the rumblings in the press about the likelihood of a deferral rather than a workable outcome; I was still poring over the House of Lords report, and glowing from the European Parliament’s Voss vote that validated the licensing-first argument at a major institutional level the week before. Could we expect something extraordinary? Would the economic power of the creative industries triumph against the AI-first dominance of the past two years? Would it deliver something with the gusto and confidence of the UK government’s quantum investment just 24 hours before? 

Spoiler alert – it did not. It did dismiss its most contentious positioning (the opt-out text and data mining exception), but it struggled to offer anything more – no legislative commitment, no enforcement timelines, no clear signal of what AI companies compliance looks like. The 100+ pages merely promised to gather more evidence, review and monitor.

The report observed that the ‘licensing market continues to grow to the benefit of right holders and AI developers, though limited public information on licensing deals makes it difficult to fully assess their impact’, and didn’t consider there to be ‘sufficient evidence to justify government intervention’. In other words, the spoils continue to go to organizations with the catalog depth, scale and legal resources to negotiate directly. This sits among a gallery of the usual suspects to avoid mandating a license-first approach; a lack of agreed technical standards; the argument that there’s sufficient legislation in place, whilst promising further consultations to solve others, notably digital replicas; and the damning acknowledgement of a ‘more limited prospect of a direct licensing market developing for individual right holders and SME right holders’.

It's almost exactly the same stagnancy the creative industries have been floating in since the consultation’s inception, during which creators have continued to lose ground unless they happen to be part of the aforementioned closed door deals…So how do we proceed? Firstly, the absence of a perfect universal technical standard is not the same as the absence of workable ones; ISCC fingerprinting, C2PA Content Credentials and RSL (mentioned in the report) already exist and function now. Infrastructure that waits for legislative and technical perfection will still be waiting when the EU AI Act enforcement deadline arrives in August, and far, far beyond. 

Secondly, the gravy train of publicly available training data will run dry sometime between now and 2032; training AI on AI-generated content leads to model collapse – human content is the foundation the technology depends on – and deals are being made.

Thirdly, existing contracts and legislation enforcement across the creative industries varies wildly; and as Creative Industries Policy and Evidence Centre asserted, copyright holders may 'retain little control over their terms and conditions, with the potential for legal rights being overwritten via private contracting'. These analogue approaches don’t enable the transparency to license (or prohibit use of) creative outputs in the AI age. This is the gap independently identified by three landmark reports; the House of Lords Communications and Digital Committee called for a licensing-first ecosystem; the European Parliament adopted the Voss report by a landslide, explicitly recognizing collective management organizations as central to any functioning framework; and now Westminster’s report indicates the same direction whilst omitting a timeline (or vehicle) to get there. Some may argue that no policy is better than a bad one – but waiting for the caprices of the government’s vague timelines, the potential volatility caused by cabinet reshuffles, the uncertainty of legal outcomes (especially the Getty vs. Stability AI appeal) to be resolved means it could take years for smaller and independent creators to gain any control over their art, consent for its use, and compensation for their craft.

The relentless and passionate campaigns of UK rights organizations undoubtedly shifted the landscape of this report – but campaigns alone do not get creators paid, or protect their work from further unfettered use – they need infrastructure. 

As a creative, I grew tired of these circular arguments and built Magpie Standard in this uncertain landscape as a boot-strapped founder who wanted to offer a solution that intersects legislation, IP and creator rights. Our infrastructure is research-based (embodying the Access, Control, Consent, Compensation and Transparency (ACCCT) framework – cited multiple times in the House of Lords report); interoperable at its core; can be integrated into existing rights organizations; offers AI companies a route to access (compliant) machine-readable licensed human creativity; and most importantly it offers creators control now – not in three years’ time. The waitlist for our creator direct platform is already open…

The question remains – are we brave enough to ‘mind the gap’ without waiting for the ‘perfect’ policy? Or will we simply watch the gap widen and let the creative economy be consumed by it?

Sonia Hendy is the founder of The Hendy Collective and creator of Magpie Standard, an AI licensing infrastructure platform for rights organizations. A published poet, songwriter, and fiction writer with bylines in The Guardian and HuffPost UK, she brings two decades across higher education leadership, transformation consultancy, and creative practice to the question of how creators license their work for AI training. 

Beyond Licensing: Why the Open Web Needs Governance, Not Just Deals

By Nick Sullivan

Licensing deals between AI companies and content publishers can't fix the web's AI governance problem because much of what AI actually trains on can't be licensed in the first place.

Licensing is a rational response to a real problem. Professionally produced content is being copied at scale for model training, often in ways that reduce referral traffic and revenue. Because of this, large publishers and AI firms have already cut deals: Reuters has reported on Reddit's reported $60 million-a-year agreement with Google and the Wikimedia Foundation's paid enterprise-access deals with Microsoft, Meta, Amazon, Perplexity, and Mistral. Those arrangements make sense for institutions with bargaining power, recognizable inventories, and ongoing demand. Still, they also reveal the model's limits: a governance system built around licensing will privilege actors already organized to sell access.

There is also a deeper structural problem with a licensing-only approach. As WIPO puts it, "copyright laws are territorial." The Internet is not. For example, in the EU, text and data mining of lawfully accessible works is permitted unless rightsholders have expressly reserved their rights. In the United States, courts are still drawing lines around fair use and market harm in AI training cases. Licensing-first solutions often assume a stable, global legal baseline that does not exist. A meaningful fraction of what AI models actually train on sits in categories where copyright's reach is limited or unclear. The US Copyright Office is direct on this: copyright "does not protect facts, ideas, systems, or methods of operation." User-generated content, community-maintained reference works, and factual material cannot be licensed away because, in many cases, there is nothing to license.

Licensing also risks centralizing the wrong things. A licensing regime turns the Internet from a system that rewards "who gets heard" into one that rewards "who can invoice." Australia's parliamentary review of its News Media Bargaining Code found that the bulk of funding from commercial agreements went to legacy outlets, deepening disadvantages for smaller publishers. Pew found that the most frequently cited sources in Google AI summaries were Wikipedia, YouTube, and Reddit, and only 5% linked to news sites. The biggest publishers and platforms will have the clearest leverage. This compounds with the fact that there is a mismatch between what content is important for AI and who can monetize through licensing. Major foundation model training pipelines rely on web-scale crawls. Much of the material AI systems actually value is collaborative, user-generated, or community-maintained, and it rarely comes packaged in the clean rights bundles collective licensing assumes. Common Crawl alone spans over 300 billion pages across millions of distinct sites. The vast majority of those sites will never be party to a licensing deal, and a system that routes value only through bilateral agreements will deepen the centralization it claims to remedy.

The answer has to be broader than compensation. We need open technical standards that let sites, creators, and users express preferences about automated consumption in a machine-readable way, and we need model providers to respect them. The emerging crawler controls show what a norms-based Internet could look like. OpenAI lets sites allow OAI-SearchBot for search while disallowing GPTBot for training. Apple lets publishers direct it not to use their site content for model training, and lets individuals object to the crawling of URLs that contain their personal data. Google is less clear-cut: AI is now part of Search, but they still allow sites to differentiate between crawler use of data via Google-Extended.

The EU framework already points in that direction, and the W3C's TDM Reservation Protocol is designed to express the reservation of mining rights and to ease the discovery of associated policies. The IETF's AIPREF work builds on the current robots.txt standard and is developing a standard vocabulary for AI usage preferences that extends well beyond copyright. Even the UK Parliament committee arguing for a licensing-first approach says government should back "open, globally aligned standards" for rights reservation, provenance, and labelling.

There is still a long way to go to establish these rules and standards universally. Researchers auditing consent signals across 14,000 domains underlying major training corpora describe the current situation as "symptoms of ineffective web protocols," with widespread inconsistencies between what sites intend and what robots.txt and other signals actually convey. Being open to being read, cited, or indexed is not the same as consenting to model training. Licensing should remain one tool in the stack, but a stable equilibrium for the AI web requires transparency, provenance, privacy, and meaningful consent for both people and publishers.

Nick Sullivan is an applied cryptographer, security researcher, and entrepreneur. He founded Cloudflare Research, co-authored several TLS extensions, is a member of the Internet Architecture Board, and holds several roles at the IETF where he works on the standards that shape how the Internet operates.

Support the Internet Exchange

If you find our emails useful, consider becoming a paid subscriber! You'll get access to our members-only Signal community where we share ideas, discuss upcoming topics, and exchange links. Paid subscribers can also leave comments on posts and enjoy a warm, fuzzy feeling.

Not ready for a long-term commitment? You can always leave us a tip.

Become A Paid Subscriber

🚨
Stop press! Do you enjoy our links? From next week, the links roundup will be available to paid Internet Exchange subscribers only. For a limited time, we’re offering 50% off an annual subscription. Normally $50 per year: right now you can subscribe for $25. Become a paid subscriber today.

Open Social Web

Internet Governance

Digital Rights

Technology for Society

Privacy and Security

Upcoming Events

Careers and Funding Opportunities

Opportunities to Get Involved

What did we miss? Please send us a reply or write to editor@exchangepoint.tech.

💡
Want to see some of our week's links in advance? Follow us on Mastodon, Bluesky or LinkedIn, and don't forget to forward and share!